[llvm] [X86] shouldReduceLoadWidth - don't split loads if ANY uses are a extract+store or a full width legal binop (PR #129695)

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Wed Mar 5 01:18:07 PST 2025


================
@@ -11,7 +11,7 @@ define <8 x i16> @pr25080(<8 x i32> %a) nounwind {
 ; AVX1-LABEL: pr25080:
 ; AVX1:       # %bb.0: # %entry
 ; AVX1-NEXT:    vextractf128 $1, %ymm0, %xmm0
-; AVX1-NEXT:    vpand {{\.?LCPI[0-9]+_[0-9]+}}(%rip), %xmm0, %xmm0
+; AVX1-NEXT:    vpand {{\.?LCPI[0-9]+_[0-9]+}}+16(%rip), %xmm0, %xmm0
----------------
RKSimon wrote:

In this case, whats happened is that the extract has now loaded from the upper half of a single 256-bit constant pool (the other fold is loading the lower half). In the old codegen, we have 2 separate 128-bit constant pool entries.

https://github.com/llvm/llvm-project/pull/129695


More information about the llvm-commits mailing list