[LLVMdev] x86-64 backend generates aligned ADDPS with unaligned address
Reid Kleckner
rnk at google.com
Wed Jul 29 13:54:48 PDT 2015
This load instruction assumes the default ABI alignment for the <4 x float>
type, which is 16:
%15 = load <4 x float>* %14
You can set the alignment of loads to something lower than 16 in your
frontend, and this will make LLVM use movups instructions:
%15 = load <4 x float>* %14, align 4
If some LLVM mid-level pass is introducing this load without proving that
the vector is 16-byte aligned, then that's a bug
On Wed, Jul 29, 2015 at 1:02 PM, Frank Winter <fwinter at jlab.org> wrote:
> When I compile attached IR with LLVM 3.6
>
> llc -march=x86-64 -o f.S f.ll
>
> it generates an aligned ADDPS with unaligned address. See attached f.S,
> here an extract:
>
> addq $12, %r9 # $12 is not a multiple of 4, thus for
> xmm0 this is unaligned
> xorl %esi, %esi
> .align 16, 0x90
> .LBB0_1: # %loop2
> # =>This Inner Loop Header: Depth=1
> movq offset_array3(,%rsi,8), %rdi
> movq offset_array2(,%rsi,8), %r10
> movss -28(%rax), %xmm0
> movss -8(%rax), %xmm1
> movss -4(%rax), %xmm2
> unpcklps %xmm0, %xmm2 # xmm2 =
> xmm2[0],xmm0[0],xmm2[1],xmm0[1]
> movss (%rax), %xmm0
> unpcklps %xmm0, %xmm1 # xmm1 =
> xmm1[0],xmm0[0],xmm1[1],xmm0[1]
> unpcklps %xmm2, %xmm1 # xmm1 =
> xmm1[0],xmm2[0],xmm1[1],xmm2[1]
> addps (%r9), %xmm1 # here, it gets used, causes a
> segfault
>
>
> Frank
>
>
> _______________________________________________
> LLVM Developers mailing list
> LLVMdev at cs.uiuc.edu http://llvm.cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvmdev
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-dev/attachments/20150729/fe2b1842/attachment.html>
More information about the llvm-dev
mailing list