[llvm] r358170 - [X86] SimplifyDemandedVectorElts - add X86ISD::VPERMILPV mask support
Simon Pilgrim via llvm-commits
llvm-commits at lists.llvm.org
Thu Apr 11 07:15:01 PDT 2019
Author: rksimon
Date: Thu Apr 11 07:15:01 2019
New Revision: 358170
URL: http://llvm.org/viewvc/llvm-project?rev=358170&view=rev
Log:
[X86] SimplifyDemandedVectorElts - add X86ISD::VPERMILPV mask support
Modified:
llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
llvm/trunk/test/CodeGen/X86/vector-shuffle-combining-avx.ll
Modified: llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86ISelLowering.cpp?rev=358170&r1=358169&r2=358170&view=diff
==============================================================================
--- llvm/trunk/lib/Target/X86/X86ISelLowering.cpp (original)
+++ llvm/trunk/lib/Target/X86/X86ISelLowering.cpp Thu Apr 11 07:15:01 2019
@@ -33225,7 +33225,8 @@ bool X86TargetLowering::SimplifyDemanded
}
break;
}
- case X86ISD::PSHUFB: {
+ case X86ISD::PSHUFB:
+ case X86ISD::VPERMILPV: {
// TODO - simplify other variable shuffle masks.
SDValue Mask = Op.getOperand(1);
APInt MaskUndef, MaskZero;
Modified: llvm/trunk/test/CodeGen/X86/vector-shuffle-combining-avx.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/vector-shuffle-combining-avx.ll?rev=358170&r1=358169&r2=358170&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/vector-shuffle-combining-avx.ll (original)
+++ llvm/trunk/test/CodeGen/X86/vector-shuffle-combining-avx.ll Thu Apr 11 07:15:01 2019
@@ -188,41 +188,10 @@ define <8 x float> @combine_vpermilvar_8
ret <8 x float> %1
}
define <8 x float> @demandedelts_vpermilvar_8f32_movshdup(<8 x float> %a0, i32 %a1) {
-; X86-LABEL: demandedelts_vpermilvar_8f32_movshdup:
-; X86: # %bb.0:
-; X86-NEXT: vbroadcastss {{[0-9]+}}(%esp), %ymm1
-; X86-NEXT: vblendps {{.*#+}} ymm1 = mem[0,1,2,3,4,5,6],ymm1[7]
-; X86-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X86-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5,6,6]
-; X86-NEXT: retl
-;
-; X64-AVX1-LABEL: demandedelts_vpermilvar_8f32_movshdup:
-; X64-AVX1: # %bb.0:
-; X64-AVX1-NEXT: vmovd %edi, %xmm1
-; X64-AVX1-NEXT: vpshufd {{.*#+}} xmm1 = xmm1[0,1,2,0]
-; X64-AVX1-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm1
-; X64-AVX1-NEXT: vblendps {{.*#+}} ymm1 = mem[0,1,2,3,4,5,6],ymm1[7]
-; X64-AVX1-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX1-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5,6,6]
-; X64-AVX1-NEXT: retq
-;
-; X64-AVX2-LABEL: demandedelts_vpermilvar_8f32_movshdup:
-; X64-AVX2: # %bb.0:
-; X64-AVX2-NEXT: vmovd %edi, %xmm1
-; X64-AVX2-NEXT: vpbroadcastd %xmm1, %ymm1
-; X64-AVX2-NEXT: vpblendd {{.*#+}} ymm1 = mem[0,1,2,3,4,5,6],ymm1[7]
-; X64-AVX2-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX2-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5,6,6]
-; X64-AVX2-NEXT: retq
-;
-; X64-AVX512-LABEL: demandedelts_vpermilvar_8f32_movshdup:
-; X64-AVX512: # %bb.0:
-; X64-AVX512-NEXT: vmovd %edi, %xmm1
-; X64-AVX512-NEXT: vpbroadcastd %xmm1, %ymm1
-; X64-AVX512-NEXT: vpblendd {{.*#+}} ymm1 = mem[0,1,2,3,4,5,6],ymm1[7]
-; X64-AVX512-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX512-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5,6,6]
-; X64-AVX512-NEXT: retq
+; CHECK-LABEL: demandedelts_vpermilvar_8f32_movshdup:
+; CHECK: # %bb.0:
+; CHECK-NEXT: vmovshdup {{.*#+}} ymm0 = ymm0[1,1,3,3,5,5,7,7]
+; CHECK-NEXT: ret{{[l|q]}}
%1 = insertelement <8 x i32> <i32 1, i32 1, i32 3, i32 3, i32 undef, i32 5, i32 7, i32 7>, i32 %a1, i32 7
%2 = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %a0, <8 x i32> %1)
%3 = shufflevector <8 x float> %2, <8 x float> undef, <8 x i32> <i32 0, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 6>
@@ -238,55 +207,23 @@ define <8 x float> @combine_vpermilvar_8
ret <8 x float> %1
}
define <8 x float> @demandedelts_vpermilvar_8f32_movsldup(<8 x float> %a0, i32 %a1) {
-; X86-AVX1-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X86-AVX1: # %bb.0:
-; X86-AVX1-NEXT: vmovdqa {{.*#+}} xmm1 = <u,0,2,2,4,4,6,6>
-; X86-AVX1-NEXT: vpinsrd $0, {{[0-9]+}}(%esp), %xmm1, %xmm1
-; X86-AVX1-NEXT: vblendps {{.*#+}} ymm1 = ymm1[0,1,2,3],mem[4,5,6,7]
-; X86-AVX1-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X86-AVX1-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X86-AVX1-NEXT: retl
-;
-; X86-AVX2-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X86-AVX2: # %bb.0:
-; X86-AVX2-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
-; X86-AVX2-NEXT: vblendps {{.*#+}} ymm1 = ymm1[0],mem[1,2,3,4,5,6,7]
-; X86-AVX2-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X86-AVX2-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X86-AVX2-NEXT: retl
-;
-; X86-AVX512-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X86-AVX512: # %bb.0:
-; X86-AVX512-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero
-; X86-AVX512-NEXT: vblendps {{.*#+}} ymm1 = ymm1[0],mem[1,2,3,4,5,6,7]
-; X86-AVX512-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X86-AVX512-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X86-AVX512-NEXT: retl
-;
-; X64-AVX1-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X64-AVX1: # %bb.0:
-; X64-AVX1-NEXT: vmovdqa {{.*#+}} xmm1 = <u,0,2,2,4,4,6,6>
-; X64-AVX1-NEXT: vpinsrd $0, %edi, %xmm1, %xmm1
-; X64-AVX1-NEXT: vblendps {{.*#+}} ymm1 = ymm1[0,1,2,3],mem[4,5,6,7]
-; X64-AVX1-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX1-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X64-AVX1-NEXT: retq
-;
-; X64-AVX2-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X64-AVX2: # %bb.0:
-; X64-AVX2-NEXT: vmovd %edi, %xmm1
-; X64-AVX2-NEXT: vpblendd {{.*#+}} ymm1 = ymm1[0],mem[1,2,3,4,5,6,7]
-; X64-AVX2-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX2-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X64-AVX2-NEXT: retq
-;
-; X64-AVX512-LABEL: demandedelts_vpermilvar_8f32_movsldup:
-; X64-AVX512: # %bb.0:
-; X64-AVX512-NEXT: vmovd %edi, %xmm1
-; X64-AVX512-NEXT: vpblendd {{.*#+}} ymm1 = ymm1[0],mem[1,2,3,4,5,6,7]
-; X64-AVX512-NEXT: vpermilps %ymm1, %ymm0, %ymm0
-; X64-AVX512-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
-; X64-AVX512-NEXT: retq
+; AVX1-LABEL: demandedelts_vpermilvar_8f32_movsldup:
+; AVX1: # %bb.0:
+; AVX1-NEXT: vmovaps {{.*#+}} xmm1 = <u,0,2,2,4,4,6,6>
+; AVX1-NEXT: vblendps {{.*#+}} ymm1 = ymm1[0,1,2,3],mem[4,5,6,7]
+; AVX1-NEXT: vpermilps %ymm1, %ymm0, %ymm0
+; AVX1-NEXT: vpermilps {{.*#+}} ymm0 = ymm0[1,1,2,3,4,5,6,7]
+; AVX1-NEXT: ret{{[l|q]}}
+;
+; AVX2-LABEL: demandedelts_vpermilvar_8f32_movsldup:
+; AVX2: # %bb.0:
+; AVX2-NEXT: vmovsldup {{.*#+}} ymm0 = ymm0[0,0,2,2,4,4,6,6]
+; AVX2-NEXT: ret{{[l|q]}}
+;
+; AVX512-LABEL: demandedelts_vpermilvar_8f32_movsldup:
+; AVX512: # %bb.0:
+; AVX512-NEXT: vmovsldup {{.*#+}} ymm0 = ymm0[0,0,2,2,4,4,6,6]
+; AVX512-NEXT: ret{{[l|q]}}
%1 = insertelement <8 x i32> <i32 0, i32 0, i32 2, i32 2, i32 4, i32 4, i32 6, i32 6>, i32 %a1, i32 0
%2 = tail call <8 x float> @llvm.x86.avx.vpermilvar.ps.256(<8 x float> %a0, <8 x i32> %1)
%3 = shufflevector <8 x float> %2, <8 x float> undef, <8 x i32> <i32 1, i32 1, i32 2, i32 3, i32 4, i32 5, i32 6, i32 7>
More information about the llvm-commits
mailing list