[LLVMbugs] [Bug 21710] New: [X86][AVX] suboptimal expansion of 256 bit vector loads.
bugzilla-daemon at llvm.org
bugzilla-daemon at llvm.org
Tue Dec 2 08:43:06 PST 2014
http://llvm.org/bugs/show_bug.cgi?id=21710
Bug ID: 21710
Summary: [X86][AVX] suboptimal expansion of 256 bit vector
loads.
Product: libraries
Version: trunk
Hardware: PC
OS: Linux
Status: NEW
Severity: normal
Priority: P
Component: Backend: X86
Assignee: unassignedbugs at nondot.org
Reporter: andrea.dibiagio at gmail.com
CC: llvmbugs at cs.uiuc.edu
Classification: Unclassified
Per-element loads for 128-bit vectors works fine (uses unaligned load
instructions):
$ cat test.c
///
#include <x86intrin.h>
__m128 test1(const float* ptr) {
return (__m128){ ptr[0], ptr[1], ptr[2], ptr[3] };
}
__m128d test2(const double* ptr) {
return (__m128d){ ptr[0], ptr[1] };
}
///
$ clang test.c -march=btver2 -O2 -S -emit-llvm -o test.ll
$ cat test.ll
define <4 x float> @test1(float* nocapture readonly %ptr) {
entry:
%0 = load float* %ptr, align 4, !tbaa !1
%vecinit = insertelement <4 x float> undef, float %0, i32 0
%arrayidx1 = getelementptr inbounds float* %ptr, i64 1
%1 = load float* %arrayidx1, align 4, !tbaa !1
%vecinit2 = insertelement <4 x float> %vecinit, float %1, i32 1
%arrayidx3 = getelementptr inbounds float* %ptr, i64 2
%2 = load float* %arrayidx3, align 4, !tbaa !1
%vecinit4 = insertelement <4 x float> %vecinit2, float %2, i32 2
%arrayidx5 = getelementptr inbounds float* %ptr, i64 3
%3 = load float* %arrayidx5, align 4, !tbaa !1
%vecinit6 = insertelement <4 x float> %vecinit4, float %3, i32 3
ret <4 x float> %vecinit6
}
define <2 x double> @test2(double* nocapture readonly %ptr) {
entry:
%0 = load double* %ptr, align 8, !tbaa !5
%vecinit = insertelement <2 x double> undef, double %0, i32 0
%arrayidx1 = getelementptr inbounds double* %ptr, i64 1
%1 = load double* %arrayidx1, align 8, !tbaa !5
%vecinit2 = insertelement <2 x double> %vecinit, double %1, i32 1
ret <2 x double> %vecinit2
}
$ llc -mcpu=btver2 test.ll -o -
test1:
vmovups (%rdi), %xmm0
retq
test2:
vmovups (%rdi), %xmm0
retq
---
However, 256-bit vectors loads are only partially vectorized.
/////
#include <x86intrin.h>
__m256 test1(const float* ptr) {
return (__m256){ ptr[0], ptr[1], ptr[2], ptr[3], ptr[4], ptr[5], ptr[6],
ptr[7] };
}
__m256d test2(const double* ptr) {
return (__m256d){ ptr[0], ptr[1], ptr[2], ptr[3] };
}
/////
$ clang test2.c -march=btver2 -O2 -S -emit-llvm -o test2.ll
$ cat test2.ll
define <8 x float> @test1(float* nocapture readonly %ptr) {
entry:
%0 = load float* %ptr, align 4, !tbaa !1
%vecinit = insertelement <8 x float> undef, float %0, i32 0
%arrayidx1 = getelementptr inbounds float* %ptr, i64 1
%1 = load float* %arrayidx1, align 4, !tbaa !1
%vecinit2 = insertelement <8 x float> %vecinit, float %1, i32 1
%arrayidx3 = getelementptr inbounds float* %ptr, i64 2
%2 = load float* %arrayidx3, align 4, !tbaa !1
%vecinit4 = insertelement <8 x float> %vecinit2, float %2, i32 2
%arrayidx5 = getelementptr inbounds float* %ptr, i64 3
%3 = load float* %arrayidx5, align 4, !tbaa !1
%vecinit6 = insertelement <8 x float> %vecinit4, float %3, i32 3
%arrayidx7 = getelementptr inbounds float* %ptr, i64 4
%4 = load float* %arrayidx7, align 4, !tbaa !1
%vecinit8 = insertelement <8 x float> %vecinit6, float %4, i32 4
%arrayidx9 = getelementptr inbounds float* %ptr, i64 5
%5 = load float* %arrayidx9, align 4, !tbaa !1
%vecinit10 = insertelement <8 x float> %vecinit8, float %5, i32 5
%arrayidx11 = getelementptr inbounds float* %ptr, i64 6
%6 = load float* %arrayidx11, align 4, !tbaa !1
%vecinit12 = insertelement <8 x float> %vecinit10, float %6, i32 6
%arrayidx13 = getelementptr inbounds float* %ptr, i64 7
%7 = load float* %arrayidx13, align 4, !tbaa !1
%vecinit14 = insertelement <8 x float> %vecinit12, float %7, i32 7
ret <8 x float> %vecinit14
}
define <4 x double> @test2(double* nocapture readonly %ptr) {
entry:
%0 = load double* %ptr, align 8, !tbaa !5
%vecinit = insertelement <4 x double> undef, double %0, i32 0
%arrayidx1 = getelementptr inbounds double* %ptr, i64 1
%1 = load double* %arrayidx1, align 8, !tbaa !5
%vecinit2 = insertelement <4 x double> %vecinit, double %1, i32 1
%arrayidx3 = getelementptr inbounds double* %ptr, i64 2
%2 = load double* %arrayidx3, align 8, !tbaa !5
%vecinit4 = insertelement <4 x double> %vecinit2, double %2, i32 2
%arrayidx5 = getelementptr inbounds double* %ptr, i64 3
%3 = load double* %arrayidx5, align 8, !tbaa !5
%vecinit6 = insertelement <4 x double> %vecinit4, double %3, i32 3
ret <4 x double> %vecinit6
}
$ llc -mcpu=btver2 test.ll -o -
test1:
vmovss 16(%rdi), %xmm1
vmovups (%rdi), %xmm0
vinsertps $16, 20(%rdi), %xmm1, %xmm1
vinsertps $32, 24(%rdi), %xmm1, %xmm1
vinsertps $48, 28(%rdi), %xmm1, %xmm1
vinsertf128 $1, %xmm1, %ymm0, %ymm0
retq
test2:
vmovsd 16(%rdi), %xmm1
vmovupd (%rdi), %xmm0
vmovhpd 24(%rdi), %xmm1, %xmm1
vinsertf128 %1, %xmm1, %ymm0, %ymm0
retq
Ideally, on AVX targets with feature FastUAMem and !SlowUAMem32, we should get:
test1:
vmovups (%rdi), %ymm0
test2:
vmovups (%rdi), %ymm0
The problem seems to be caused by a sub-optimal choice done during BUILD_VECTOR
legalization.
The backend knows how to expand a 128-bit BUILD_VECTOR into a single 128-bit
vector load. However, if a BUILD_VECTOR is performing a 256-bit vector load,
the backend expands it into a 128-bit load plus an insert_subvector and a chain
of insert_vector_elt nodes. That is the reason why for example 'test1'
generates a sequence of insertps followed by a vinsertf128.
--
You are receiving this mail because:
You are on the CC list for the bug.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-bugs/attachments/20141202/4dcda3d1/attachment.html>
More information about the llvm-bugs
mailing list