[llvm] r341227 - [X86] Add support for turning vXi1 shuffles into KSHIFTL/KSHIFTR.
Craig Topper via llvm-commits
llvm-commits at lists.llvm.org
Fri Aug 31 10:17:21 PDT 2018
Author: ctopper
Date: Fri Aug 31 10:17:21 2018
New Revision: 341227
URL: http://llvm.org/viewvc/llvm-project?rev=341227&view=rev
Log:
[X86] Add support for turning vXi1 shuffles into KSHIFTL/KSHIFTR.
This patch recognizes shuffles that shift elements and fill with zeros. I've copied and modified the shift matching code we use for normal vector registers to do this. I'm not sure if there's a good way to share more of this code without making the existing function more complex than it already is.
This will be used to enable kshift intrinsics in clang.
Differential Revision: https://reviews.llvm.org/D51401
Modified:
llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
llvm/trunk/test/CodeGen/X86/avx512-skx-insert-subvec.ll
llvm/trunk/test/CodeGen/X86/kshift.ll
Modified: llvm/trunk/lib/Target/X86/X86ISelLowering.cpp
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Target/X86/X86ISelLowering.cpp?rev=341227&r1=341226&r2=341227&view=diff
==============================================================================
--- llvm/trunk/lib/Target/X86/X86ISelLowering.cpp (original)
+++ llvm/trunk/lib/Target/X86/X86ISelLowering.cpp Fri Aug 31 10:17:21 2018
@@ -15053,6 +15053,39 @@ static SDValue lower512BitVectorShuffle(
}
}
+// Determine if this shuffle can be implemented with a KSHIFT instruction.
+// Returns the shift amount if possible or -1 if not. This is a simplified
+// version of matchVectorShuffleAsShift.
+static int match1BitShuffleAsKSHIFT(unsigned &Opcode, ArrayRef<int> Mask,
+ int MaskOffset, const APInt &Zeroable) {
+ int Size = Mask.size();
+
+ auto CheckZeros = [&](int Shift, bool Left) {
+ for (int j = 0; j < Shift; ++j)
+ if (!Zeroable[j + (Left ? 0 : (Size - Shift))])
+ return false;
+
+ return true;
+ };
+
+ auto MatchShift = [&](int Shift, bool Left) {
+ unsigned Pos = Left ? Shift : 0;
+ unsigned Low = Left ? 0 : Shift;
+ unsigned Len = Size - Shift;
+ return isSequentialOrUndefInRange(Mask, Pos, Len, Low + MaskOffset);
+ };
+
+ for (int Shift = 1; Shift != Size; ++Shift)
+ for (bool Left : {true, false})
+ if (CheckZeros(Shift, Left) && MatchShift(Shift, Left)) {
+ Opcode = Left ? X86ISD::KSHIFTL : X86ISD::KSHIFTR;
+ return Shift;
+ }
+
+ return -1;
+}
+
+
// Lower vXi1 vector shuffles.
// There is no a dedicated instruction on AVX-512 that shuffles the masks.
// The only way to shuffle bits is to sign-extend the mask vector to SIMD
@@ -15062,6 +15095,9 @@ static SDValue lower1BitVectorShuffle(co
const APInt &Zeroable,
const X86Subtarget &Subtarget,
SelectionDAG &DAG) {
+ assert(Subtarget.hasAVX512() &&
+ "Cannot lower 512-bit vectors w/o basic ISA!");
+
unsigned NumElts = Mask.size();
// Try to recognize shuffles that are just padding a subvector with zeros.
@@ -15088,9 +15124,21 @@ static SDValue lower1BitVectorShuffle(co
Extract, DAG.getIntPtrConstant(0, DL));
}
+ // Try to match KSHIFTs.
+ // TODO: Support narrower than legal shifts by widening and extracting.
+ if (NumElts >= 16 || (Subtarget.hasDQI() && NumElts == 8)) {
+ unsigned Offset = 0;
+ for (SDValue V : { V1, V2 }) {
+ unsigned Opcode;
+ int ShiftAmt = match1BitShuffleAsKSHIFT(Opcode, Mask, Offset, Zeroable);
+ if (ShiftAmt >= 0)
+ return DAG.getNode(Opcode, DL, VT, V,
+ DAG.getConstant(ShiftAmt, DL, MVT::i8));
+ Offset += NumElts; // Increment for next iteration.
+ }
+ }
+
- assert(Subtarget.hasAVX512() &&
- "Cannot lower 512-bit vectors w/o basic ISA!");
MVT ExtVT;
switch (VT.SimpleTy) {
default:
Modified: llvm/trunk/test/CodeGen/X86/avx512-skx-insert-subvec.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/avx512-skx-insert-subvec.ll?rev=341227&r1=341226&r2=341227&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/avx512-skx-insert-subvec.ll (original)
+++ llvm/trunk/test/CodeGen/X86/avx512-skx-insert-subvec.ll Fri Aug 31 10:17:21 2018
@@ -30,11 +30,8 @@ define <8 x i1> @test2(<2 x i1> %a) {
; CHECK: # %bb.0:
; CHECK-NEXT: vpsllq $63, %xmm0, %xmm0
; CHECK-NEXT: vpmovq2m %xmm0, %k0
-; CHECK-NEXT: vpmovm2d %k0, %ymm0
-; CHECK-NEXT: vperm2i128 {{.*#+}} ymm0 = zero,zero,ymm0[0,1]
-; CHECK-NEXT: vpmovd2m %ymm0, %k0
+; CHECK-NEXT: kshiftlb $4, %k0, %k0
; CHECK-NEXT: vpmovm2w %k0, %xmm0
-; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%res = shufflevector <2 x i1> %a, <2 x i1> zeroinitializer, <8 x i32> <i32 3, i32 3, i32 undef, i32 undef, i32 0, i32 1, i32 undef, i32 undef>
ret <8 x i1> %res
Modified: llvm/trunk/test/CodeGen/X86/kshift.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/kshift.ll?rev=341227&r1=341226&r2=341227&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/kshift.ll (original)
+++ llvm/trunk/test/CodeGen/X86/kshift.ll Fri Aug 31 10:17:21 2018
@@ -20,11 +20,7 @@ define i8 @kshiftl_v8i1_1(<8 x i64> %x,
; SKX-LABEL: kshiftl_v8i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: movb $-2, %al
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vpexpandd %ymm0, %ymm0 {%k1} {z}
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftlb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -41,12 +37,8 @@ define i8 @kshiftl_v8i1_1(<8 x i64> %x,
define i16 @kshiftl_v16i1_1(<16 x i32> %x, <16 x i32> %y) {
; KNL-LABEL: kshiftl_v16i1_1:
; KNL: # %bb.0:
-; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: movw $-2, %ax
-; KNL-NEXT: kmovw %eax, %k1
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
+; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftlw $1, %k0, %k1
; KNL-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %eax
; KNL-NEXT: # kill: def $ax killed $ax killed $eax
@@ -56,11 +48,7 @@ define i16 @kshiftl_v16i1_1(<16 x i32> %
; SKX-LABEL: kshiftl_v16i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmd %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %zmm0
-; SKX-NEXT: movw $-2, %ax
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; SKX-NEXT: vpmovd2m %zmm0, %k1
+; SKX-NEXT: kshiftlw $1, %k0, %k1
; SKX-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $ax killed $ax killed $eax
@@ -86,12 +74,9 @@ define i32 @kshiftl_v32i1_1(<32 x i16> %
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k2
; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k2} {z}
; KNL-NEXT: vpternlogd $255, %zmm1, %zmm1, %zmm1 {%k1} {z}
-; KNL-NEXT: valignd {{.*#+}} zmm1 = zmm0[15],zmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
-; KNL-NEXT: vptestmd %zmm1, %zmm1, %k1
-; KNL-NEXT: movw $-2, %ax
-; KNL-NEXT: kmovw %eax, %k2
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k2} {z}
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k2
+; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm0[15],zmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
+; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
+; KNL-NEXT: kshiftlw $1, %k2, %k2
; KNL-NEXT: vpcmpeqw %ymm4, %ymm3, %ymm0
; KNL-NEXT: vpmovsxwd %ymm0, %zmm0
; KNL-NEXT: vpcmpeqw %ymm4, %ymm2, %ymm1
@@ -108,11 +93,7 @@ define i32 @kshiftl_v32i1_1(<32 x i16> %
; SKX-LABEL: kshiftl_v32i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmw %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2w %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [32,0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
-; SKX-NEXT: vpermi2w %zmm0, %zmm2, %zmm3
-; SKX-NEXT: vpmovw2m %zmm3, %k1
+; SKX-NEXT: kshiftld $1, %k0, %k1
; SKX-NEXT: vptestnmw %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: vzeroupper
@@ -149,12 +130,9 @@ define i64 @kshiftl_v64i1_1(<64 x i8> %x
; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm1[15],zmm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k2
; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k3} {z}
-; KNL-NEXT: valignd {{.*#+}} zmm1 = zmm0[15],zmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
-; KNL-NEXT: vptestmd %zmm1, %zmm1, %k3
-; KNL-NEXT: movw $-2, %ax
-; KNL-NEXT: kmovw %eax, %k4
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k4} {z}
+; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm0[15],zmm1[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14]
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k4
+; KNL-NEXT: kshiftlw $1, %k3, %k3
; KNL-NEXT: vpcmpeqb %ymm4, %ymm3, %ymm0
; KNL-NEXT: vextracti128 $1, %ymm0, %xmm1
; KNL-NEXT: vpmovsxbd %xmm1, %zmm1
@@ -163,9 +141,9 @@ define i64 @kshiftl_v64i1_1(<64 x i8> %x
; KNL-NEXT: vextracti128 $1, %ymm2, %xmm3
; KNL-NEXT: vpmovsxbd %xmm3, %zmm3
; KNL-NEXT: vpmovsxbd %xmm2, %zmm2
-; KNL-NEXT: vptestmd %zmm2, %zmm2, %k0 {%k4}
+; KNL-NEXT: vptestmd %zmm2, %zmm2, %k0 {%k3}
; KNL-NEXT: kmovw %k0, %eax
-; KNL-NEXT: vptestmd %zmm3, %zmm3, %k0 {%k3}
+; KNL-NEXT: vptestmd %zmm3, %zmm3, %k0 {%k4}
; KNL-NEXT: kmovw %k0, %ecx
; KNL-NEXT: shll $16, %ecx
; KNL-NEXT: orl %eax, %ecx
@@ -183,17 +161,7 @@ define i64 @kshiftl_v64i1_1(<64 x i8> %x
; SKX-LABEL: kshiftl_v64i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmb %zmm0, %zmm0, %k0
-; SKX-NEXT: movl $1, %eax
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: knotd %k1, %k1
-; SKX-NEXT: vpmovm2b %k0, %zmm0
-; SKX-NEXT: vpermq {{.*#+}} ymm2 = ymm0[2,3,0,1]
-; SKX-NEXT: vpalignr {{.*#+}} ymm2 {%k1} {z} = ymm2[15],ymm0[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm2[31],ymm0[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
-; SKX-NEXT: vextracti64x4 $1, %zmm0, %ymm3
-; SKX-NEXT: vperm2i128 {{.*#+}} ymm0 = ymm0[2,3],ymm3[0,1]
-; SKX-NEXT: vpalignr {{.*#+}} ymm0 = ymm0[15],ymm3[0,1,2,3,4,5,6,7,8,9,10,11,12,13,14],ymm0[31],ymm3[16,17,18,19,20,21,22,23,24,25,26,27,28,29,30]
-; SKX-NEXT: vinserti64x4 $1, %ymm0, %zmm2, %zmm0
-; SKX-NEXT: vpmovb2m %zmm0, %k1
+; SKX-NEXT: kshiftlq $1, %k0, %k1
; SKX-NEXT: vptestnmb %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovq %k0, %rax
; SKX-NEXT: vzeroupper
@@ -224,11 +192,7 @@ define i8 @kshiftl_v8i1_7(<8 x i64> %x,
; SKX-LABEL: kshiftl_v8i1_7:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: movb $-128, %al
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vpexpandd %ymm0, %ymm0 {%k1} {z}
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftlb $7, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -245,12 +209,8 @@ define i8 @kshiftl_v8i1_7(<8 x i64> %x,
define i16 @kshiftl_v16i1_15(<16 x i32> %x, <16 x i32> %y) {
; KNL-LABEL: kshiftl_v16i1_15:
; KNL: # %bb.0:
-; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: movw $-32768, %ax # imm = 0x8000
-; KNL-NEXT: kmovw %eax, %k1
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
+; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftlw $15, %k0, %k1
; KNL-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %eax
; KNL-NEXT: # kill: def $ax killed $ax killed $eax
@@ -260,11 +220,7 @@ define i16 @kshiftl_v16i1_15(<16 x i32>
; SKX-LABEL: kshiftl_v16i1_15:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmd %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %zmm0
-; SKX-NEXT: movw $-32768, %ax # imm = 0x8000
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; SKX-NEXT: vpmovd2m %zmm0, %k1
+; SKX-NEXT: kshiftlw $15, %k0, %k1
; SKX-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $ax killed $ax killed $eax
@@ -284,12 +240,8 @@ define i32 @kshiftl_v32i1_31(<32 x i16>
; KNL-NEXT: vpxor %xmm1, %xmm1, %xmm1
; KNL-NEXT: vpcmpeqw %ymm1, %ymm0, %ymm0
; KNL-NEXT: vpmovsxwd %ymm0, %zmm0
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: movw $-32768, %ax # imm = 0x8000
-; KNL-NEXT: kmovw %eax, %k1
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
+; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftlw $15, %k0, %k1
; KNL-NEXT: vpcmpeqw %ymm1, %ymm3, %ymm0
; KNL-NEXT: vpmovsxwd %ymm0, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k1}
@@ -301,11 +253,7 @@ define i32 @kshiftl_v32i1_31(<32 x i16>
; SKX-LABEL: kshiftl_v32i1_31:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmw %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2w %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [0,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,32]
-; SKX-NEXT: vpermi2w %zmm2, %zmm0, %zmm3
-; SKX-NEXT: vpmovw2m %zmm3, %k1
+; SKX-NEXT: kshiftld $31, %k0, %k1
; SKX-NEXT: vptestnmw %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: vzeroupper
@@ -324,12 +272,8 @@ define i64 @kshiftl_v64i1_63(<64 x i8> %
; KNL-NEXT: vpxor %xmm1, %xmm1, %xmm1
; KNL-NEXT: vpcmpeqb %ymm1, %ymm0, %ymm0
; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: movw $-32768, %ax # imm = 0x8000
-; KNL-NEXT: kmovw %eax, %k1
-; KNL-NEXT: vpexpandd %zmm0, %zmm0 {%k1} {z}
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
+; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftlw $15, %k0, %k1
; KNL-NEXT: vpcmpeqb %ymm1, %ymm3, %ymm0
; KNL-NEXT: vextracti128 $1, %ymm0, %xmm0
; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
@@ -343,15 +287,7 @@ define i64 @kshiftl_v64i1_63(<64 x i8> %
; SKX-LABEL: kshiftl_v64i1_63:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmb %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2b %k0, %zmm2
-; SKX-NEXT: vpslldq {{.*#+}} xmm2 = zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,xmm2[0]
-; SKX-NEXT: vinserti128 $1, %xmm2, %ymm0, %ymm2
-; SKX-NEXT: movl $-2147483648, %eax # imm = 0x80000000
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vmovdqu8 %ymm2, %ymm2 {%k1} {z}
-; SKX-NEXT: vinserti64x4 $1, %ymm2, %zmm0, %zmm0
-; SKX-NEXT: vpmovb2m %zmm0, %k1
+; SKX-NEXT: kshiftlq $63, %k0, %k1
; SKX-NEXT: vptestnmb %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovq %k0, %rax
; SKX-NEXT: vzeroupper
@@ -382,11 +318,7 @@ define i8 @kshiftr_v8i1_1(<8 x i64> %x,
; SKX-LABEL: kshiftr_v8i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: valignd {{.*#+}} ymm0 = ymm0[1,2,3,4,5,6,7,0]
-; SKX-NEXT: vpxor %xmm2, %xmm2, %xmm2
-; SKX-NEXT: vpblendd {{.*#+}} ymm0 = ymm0[0,1,2,3,4,5,6],ymm2[7]
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftrb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -403,12 +335,8 @@ define i8 @kshiftr_v8i1_1(<8 x i64> %x,
define i16 @kshiftr_v16i1_1(<16 x i32> %x, <16 x i32> %y) {
; KNL-LABEL: kshiftr_v16i1_1:
; KNL: # %bb.0:
-; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; KNL-NEXT: vpternlogd $255, %zmm2, %zmm2, %zmm2 {%k1} {z}
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm3 = [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,31]
-; KNL-NEXT: vpermi2d %zmm0, %zmm2, %zmm3
-; KNL-NEXT: vptestmd %zmm3, %zmm3, %k1
+; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftrw $1, %k0, %k1
; KNL-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %eax
; KNL-NEXT: # kill: def $ax killed $ax killed $eax
@@ -418,11 +346,7 @@ define i16 @kshiftr_v16i1_1(<16 x i32> %
; SKX-LABEL: kshiftr_v16i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmd %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2d %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,31]
-; SKX-NEXT: vpermi2d %zmm0, %zmm2, %zmm3
-; SKX-NEXT: vpmovd2m %zmm3, %k1
+; SKX-NEXT: kshiftrw $1, %k0, %k1
; SKX-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $ax killed $ax killed $eax
@@ -449,20 +373,17 @@ define i32 @kshiftr_v32i1_1(<32 x i16> %
; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k2} {z}
; KNL-NEXT: vpternlogd $255, %zmm1, %zmm1, %zmm1 {%k1} {z}
; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],zmm1[0]
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm0 = [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,31]
-; KNL-NEXT: vpxor %xmm5, %xmm5, %xmm5
-; KNL-NEXT: vpermi2d %zmm5, %zmm1, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k2
-; KNL-NEXT: vpcmpeqw %ymm4, %ymm3, %ymm0
+; KNL-NEXT: kshiftrw $1, %k1, %k1
+; KNL-NEXT: vpcmpeqw %ymm4, %ymm2, %ymm0
; KNL-NEXT: vpmovsxwd %ymm0, %zmm0
-; KNL-NEXT: vpcmpeqw %ymm4, %ymm2, %ymm1
+; KNL-NEXT: vpcmpeqw %ymm4, %ymm3, %ymm1
; KNL-NEXT: vpmovsxwd %ymm1, %zmm1
; KNL-NEXT: vptestmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %ecx
+; KNL-NEXT: shll $16, %ecx
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k2}
; KNL-NEXT: kmovw %k0, %eax
-; KNL-NEXT: shll $16, %eax
; KNL-NEXT: orl %ecx, %eax
; KNL-NEXT: vzeroupper
; KNL-NEXT: retq
@@ -470,11 +391,7 @@ define i32 @kshiftr_v32i1_1(<32 x i16> %
; SKX-LABEL: kshiftr_v32i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmw %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2w %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,63]
-; SKX-NEXT: vpermi2w %zmm0, %zmm2, %zmm3
-; SKX-NEXT: vpmovw2m %zmm3, %k1
+; SKX-NEXT: kshiftrd $1, %k0, %k1
; SKX-NEXT: vptestnmw %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: vzeroupper
@@ -491,53 +408,50 @@ define i64 @kshiftr_v64i1_1(<64 x i8> %x
; KNL-LABEL: kshiftr_v64i1_1:
; KNL: # %bb.0:
; KNL-NEXT: vpxor %xmm4, %xmm4, %xmm4
+; KNL-NEXT: vpcmpeqb %ymm4, %ymm1, %ymm1
+; KNL-NEXT: vextracti128 $1, %ymm1, %xmm5
+; KNL-NEXT: vpmovsxbd %xmm5, %zmm5
+; KNL-NEXT: vptestmd %zmm5, %zmm5, %k3
; KNL-NEXT: vpcmpeqb %ymm4, %ymm0, %ymm0
; KNL-NEXT: vpmovsxbd %xmm0, %zmm5
; KNL-NEXT: vptestmd %zmm5, %zmm5, %k2
-; KNL-NEXT: vextracti128 $1, %ymm0, %xmm0
-; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k3
-; KNL-NEXT: vpcmpeqb %ymm4, %ymm1, %ymm0
-; KNL-NEXT: vextracti128 $1, %ymm0, %xmm1
; KNL-NEXT: vpmovsxbd %xmm1, %zmm1
; KNL-NEXT: vptestmd %zmm1, %zmm1, %k1
+; KNL-NEXT: vextracti128 $1, %ymm0, %xmm0
; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k4
; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k4} {z}
; KNL-NEXT: vpternlogd $255, %zmm1, %zmm1, %zmm1 {%k1} {z}
; KNL-NEXT: valignd {{.*#+}} zmm5 = zmm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],zmm1[0]
; KNL-NEXT: vptestmd %zmm5, %zmm5, %k1
-; KNL-NEXT: vpternlogd $255, %zmm5, %zmm5, %zmm5 {%k3} {z}
+; KNL-NEXT: vpternlogd $255, %zmm5, %zmm5, %zmm5 {%k2} {z}
; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm5[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],zmm0[0]
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k3
-; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k2} {z}
-; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],zmm5[0]
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k4
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm0 = [1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,31]
-; KNL-NEXT: vpxor %xmm5, %xmm5, %xmm5
-; KNL-NEXT: vpermi2d %zmm5, %zmm1, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k2
-; KNL-NEXT: vpcmpeqb %ymm4, %ymm3, %ymm0
+; KNL-NEXT: vpternlogd $255, %zmm0, %zmm0, %zmm0 {%k3} {z}
+; KNL-NEXT: valignd {{.*#+}} zmm0 = zmm1[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],zmm0[0]
+; KNL-NEXT: vptestmd %zmm0, %zmm0, %k4
+; KNL-NEXT: kshiftrw $1, %k3, %k3
+; KNL-NEXT: vpcmpeqb %ymm4, %ymm2, %ymm0
; KNL-NEXT: vextracti128 $1, %ymm0, %xmm1
; KNL-NEXT: vpmovsxbd %xmm1, %zmm1
; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
-; KNL-NEXT: vpcmpeqb %ymm4, %ymm2, %ymm2
-; KNL-NEXT: vextracti128 $1, %ymm2, %xmm3
-; KNL-NEXT: vpmovsxbd %xmm3, %zmm3
+; KNL-NEXT: vpcmpeqb %ymm4, %ymm3, %ymm2
+; KNL-NEXT: vpmovsxbd %xmm2, %zmm3
+; KNL-NEXT: vextracti128 $1, %ymm2, %xmm2
; KNL-NEXT: vpmovsxbd %xmm2, %zmm2
-; KNL-NEXT: vptestmd %zmm2, %zmm2, %k0 {%k4}
+; KNL-NEXT: vptestmd %zmm2, %zmm2, %k0 {%k3}
; KNL-NEXT: kmovw %k0, %eax
-; KNL-NEXT: vptestmd %zmm3, %zmm3, %k0 {%k3}
+; KNL-NEXT: shll $16, %eax
+; KNL-NEXT: vptestmd %zmm3, %zmm3, %k0 {%k4}
; KNL-NEXT: kmovw %k0, %ecx
-; KNL-NEXT: shll $16, %ecx
; KNL-NEXT: orl %eax, %ecx
-; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k1}
+; KNL-NEXT: shlq $32, %rcx
+; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k2}
; KNL-NEXT: kmovw %k0, %edx
-; KNL-NEXT: vptestmd %zmm1, %zmm1, %k0 {%k2}
+; KNL-NEXT: vptestmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %eax
; KNL-NEXT: shll $16, %eax
; KNL-NEXT: orl %edx, %eax
-; KNL-NEXT: shlq $32, %rax
; KNL-NEXT: orq %rcx, %rax
; KNL-NEXT: vzeroupper
; KNL-NEXT: retq
@@ -545,17 +459,7 @@ define i64 @kshiftr_v64i1_1(<64 x i8> %x
; SKX-LABEL: kshiftr_v64i1_1:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmb %zmm0, %zmm0, %k0
-; SKX-NEXT: movl $-2147483648, %eax # imm = 0x80000000
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: knotd %k1, %k1
-; SKX-NEXT: vpmovm2b %k0, %zmm0
-; SKX-NEXT: vextracti64x4 $1, %zmm0, %ymm2
-; SKX-NEXT: vpermq {{.*#+}} ymm3 = ymm2[2,3,0,1]
-; SKX-NEXT: vpalignr {{.*#+}} ymm3 {%k1} {z} = ymm2[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],ymm3[0],ymm2[17,18,19,20,21,22,23,24,25,26,27,28,29,30,31],ymm3[16]
-; SKX-NEXT: vperm2i128 {{.*#+}} ymm2 = ymm0[2,3],ymm2[0,1]
-; SKX-NEXT: vpalignr {{.*#+}} ymm0 = ymm0[1,2,3,4,5,6,7,8,9,10,11,12,13,14,15],ymm2[0],ymm0[17,18,19,20,21,22,23,24,25,26,27,28,29,30,31],ymm2[16]
-; SKX-NEXT: vinserti64x4 $1, %ymm3, %zmm0, %zmm0
-; SKX-NEXT: vpmovb2m %zmm0, %k1
+; SKX-NEXT: kshiftrq $1, %k0, %k1
; SKX-NEXT: vptestnmb %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovq %k0, %rax
; SKX-NEXT: vzeroupper
@@ -586,11 +490,7 @@ define i8 @kshiftr_v8i1_7(<8 x i64> %x,
; SKX-LABEL: kshiftr_v8i1_7:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: movb $-2, %al
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vpexpandd %ymm0, %ymm0 {%k1} {z}
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftlb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -607,12 +507,8 @@ define i8 @kshiftr_v8i1_7(<8 x i64> %x,
define i16 @kshiftr_v16i1_15(<16 x i32> %x, <16 x i32> %y) {
; KNL-LABEL: kshiftr_v16i1_15:
; KNL: # %bb.0:
-; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k1
-; KNL-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; KNL-NEXT: vpternlogd $255, %zmm2, %zmm2, %zmm2 {%k1} {z}
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm3 = [31,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
-; KNL-NEXT: vpermi2d %zmm2, %zmm0, %zmm3
-; KNL-NEXT: vptestmd %zmm3, %zmm3, %k1
+; KNL-NEXT: vptestnmd %zmm0, %zmm0, %k0
+; KNL-NEXT: kshiftrw $15, %k0, %k1
; KNL-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; KNL-NEXT: kmovw %k0, %eax
; KNL-NEXT: # kill: def $ax killed $ax killed $eax
@@ -622,11 +518,7 @@ define i16 @kshiftr_v16i1_15(<16 x i32>
; SKX-LABEL: kshiftr_v16i1_15:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmd %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2d %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [31,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
-; SKX-NEXT: vpermi2d %zmm2, %zmm0, %zmm3
-; SKX-NEXT: vpmovd2m %zmm3, %k1
+; SKX-NEXT: kshiftrw $15, %k0, %k1
; SKX-NEXT: vptestnmd %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $ax killed $ax killed $eax
@@ -646,12 +538,8 @@ define i32 @kshiftr_v32i1_31(<32 x i16>
; KNL-NEXT: vpxor %xmm0, %xmm0, %xmm0
; KNL-NEXT: vpcmpeqw %ymm0, %ymm1, %ymm1
; KNL-NEXT: vpmovsxwd %ymm1, %zmm1
-; KNL-NEXT: vptestmd %zmm1, %zmm1, %k1
-; KNL-NEXT: vpternlogd $255, %zmm1, %zmm1, %zmm1 {%k1} {z}
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm3 = [31,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
-; KNL-NEXT: vpxor %xmm4, %xmm4, %xmm4
-; KNL-NEXT: vpermt2d %zmm1, %zmm3, %zmm4
-; KNL-NEXT: vptestmd %zmm4, %zmm4, %k1
+; KNL-NEXT: vptestmd %zmm1, %zmm1, %k0
+; KNL-NEXT: kshiftrw $15, %k0, %k1
; KNL-NEXT: vpcmpeqw %ymm0, %ymm2, %ymm0
; KNL-NEXT: vpmovsxwd %ymm0, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k1}
@@ -662,11 +550,7 @@ define i32 @kshiftr_v32i1_31(<32 x i16>
; SKX-LABEL: kshiftr_v32i1_31:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmw %zmm0, %zmm0, %k0
-; SKX-NEXT: vpxor %xmm0, %xmm0, %xmm0
-; SKX-NEXT: vpmovm2w %k0, %zmm2
-; SKX-NEXT: vmovdqa64 {{.*#+}} zmm3 = [63,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31]
-; SKX-NEXT: vpermi2w %zmm2, %zmm0, %zmm3
-; SKX-NEXT: vpmovw2m %zmm3, %k1
+; SKX-NEXT: kshiftrd $31, %k0, %k1
; SKX-NEXT: vptestnmw %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: vzeroupper
@@ -686,12 +570,8 @@ define i64 @kshiftr_v64i1_63(<64 x i8> %
; KNL-NEXT: vpcmpeqb %ymm0, %ymm1, %ymm1
; KNL-NEXT: vextracti128 $1, %ymm1, %xmm1
; KNL-NEXT: vpmovsxbd %xmm1, %zmm1
-; KNL-NEXT: vptestmd %zmm1, %zmm1, %k1
-; KNL-NEXT: vpternlogd $255, %zmm1, %zmm1, %zmm1 {%k1} {z}
-; KNL-NEXT: vmovdqa64 {{.*#+}} zmm3 = [31,1,2,3,4,5,6,7,8,9,10,11,12,13,14,15]
-; KNL-NEXT: vpxor %xmm4, %xmm4, %xmm4
-; KNL-NEXT: vpermt2d %zmm1, %zmm3, %zmm4
-; KNL-NEXT: vptestmd %zmm4, %zmm4, %k1
+; KNL-NEXT: vptestmd %zmm1, %zmm1, %k0
+; KNL-NEXT: kshiftrw $15, %k0, %k1
; KNL-NEXT: vpcmpeqb %ymm0, %ymm2, %ymm0
; KNL-NEXT: vpmovsxbd %xmm0, %zmm0
; KNL-NEXT: vptestmd %zmm0, %zmm0, %k0 {%k1}
@@ -703,14 +583,7 @@ define i64 @kshiftr_v64i1_63(<64 x i8> %
; SKX-LABEL: kshiftr_v64i1_63:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmb %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2b %k0, %zmm0
-; SKX-NEXT: vextracti64x4 $1, %zmm0, %ymm0
-; SKX-NEXT: vextracti128 $1, %ymm0, %xmm0
-; SKX-NEXT: vpsrldq {{.*#+}} xmm0 = xmm0[15],zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero,zero
-; SKX-NEXT: movl $1, %eax
-; SKX-NEXT: kmovd %eax, %k1
-; SKX-NEXT: vmovdqu8 %ymm0, %ymm0 {%k1} {z}
-; SKX-NEXT: vpmovb2m %zmm0, %k1
+; SKX-NEXT: kshiftrq $63, %k0, %k1
; SKX-NEXT: vptestnmb %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovq %k0, %rax
; SKX-NEXT: vzeroupper
@@ -742,11 +615,7 @@ define i8 @kshiftl_v8i1_zu123u56(<8 x i6
; SKX-LABEL: kshiftl_v8i1_zu123u56:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: valignd {{.*#+}} ymm0 = ymm0[7,0,1,2,3,4,5,6]
-; SKX-NEXT: vpxor %xmm2, %xmm2, %xmm2
-; SKX-NEXT: vpblendd {{.*#+}} ymm0 = ymm2[0,1],ymm0[2,3,4,5,6,7]
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftlb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -777,9 +646,7 @@ define i8 @kshiftl_v8i1_u0123456(<8 x i6
; SKX-LABEL: kshiftl_v8i1_u0123456:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: valignd {{.*#+}} ymm0 = ymm0[7,0,1,2,3,4,5,6]
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftlb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -812,9 +679,7 @@ define i8 @kshiftr_v8i1_1u3u567z(<8 x i6
; SKX-LABEL: kshiftr_v8i1_1u3u567z:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: vpsrldq {{.*#+}} ymm0 = ymm0[4,5,6,7,8,9,10,11,12,13,14,15],zero,zero,zero,zero,ymm0[20,21,22,23,24,25,26,27,28,29,30,31],zero,zero,zero,zero
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftrb $1, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
@@ -845,9 +710,7 @@ define i8 @kshiftr_v8i1_234567uu(<8 x i6
; SKX-LABEL: kshiftr_v8i1_234567uu:
; SKX: # %bb.0:
; SKX-NEXT: vptestnmq %zmm0, %zmm0, %k0
-; SKX-NEXT: vpmovm2d %k0, %ymm0
-; SKX-NEXT: vpermq {{.*#+}} ymm0 = ymm0[1,2,3,3]
-; SKX-NEXT: vpmovd2m %ymm0, %k1
+; SKX-NEXT: kshiftrb $2, %k0, %k1
; SKX-NEXT: vptestnmq %zmm1, %zmm1, %k0 {%k1}
; SKX-NEXT: kmovd %k0, %eax
; SKX-NEXT: # kill: def $al killed $al killed $eax
More information about the llvm-commits
mailing list