[PATCH] D63517: [x86] avoid vector load narrowing with extracted store uses (PR42305)

Sanjay Patel via Phabricator via llvm-commits llvm-commits at lists.llvm.org
Wed Jun 19 10:52:32 PDT 2019


spatel marked an inline comment as done.
spatel added inline comments.


================
Comment at: llvm/test/CodeGen/X86/sandybridge-loads.ll:44
   ret void
 }
 
----------------
RKSimon wrote:
> Not (very) related - but do we need a variant of this with that has an unaligned load:
> ```
>   %v0 = load <8 x float>, <8 x float>* %a, align 32 ; <--- aligned
>   %v1 = load <8 x float>, <8 x float>* %b, align 16 ; <--- unaligned
>   store <8 x float> %v0, <8 x float>* %b, align 32 ; <--- aligned
>   store <8 x float> %v1, <8 x float>* %a, align 16 ; <--- unaligned
>   ret void
> ```
I think for the purpose of this patch, that's covered by the next test diff (in widen_load-3.ll). 

And for the purpose of SandyBridge-specific (actual CPU feature is -slow-unaligned-mem-32), the test just above here covers it. But I can add that suggestion for completeness.


CHANGES SINCE LAST ACTION
  https://reviews.llvm.org/D63517/new/

https://reviews.llvm.org/D63517





More information about the llvm-commits mailing list