[llvm] 1274ac3 - [ARM][MVE] Sink vector shift operand
Sam Parker via llvm-commits
llvm-commits at lists.llvm.org
Thu Dec 12 06:34:19 PST 2019
Author: Sam Parker
Date: 2019-12-12T14:34:00Z
New Revision: 1274ac3dc235dd596cc1ace2145c2b1e3c970b29
URL: https://github.com/llvm/llvm-project/commit/1274ac3dc235dd596cc1ace2145c2b1e3c970b29
DIFF: https://github.com/llvm/llvm-project/commit/1274ac3dc235dd596cc1ace2145c2b1e3c970b29.diff
LOG: [ARM][MVE] Sink vector shift operand
Recommit e0b966643fc2. sub instructions were being generated for the
negated value, and for some reason they were the register only ones.
I think the problem was because I was grabbing the 'zero' from
vmovimm, which is a target constant. Now I'm just generating a new
Constant zero and so rsb instructions are now generated.
Original commit message:
The shift amount operand can be provided in a general purpose
register so sink it. Flip the vdup and negate so the existing
patterns can be used for matching.
Differential Revision: https://reviews.llvm.org/D70841
Added:
llvm/test/CodeGen/Thumb2/mve-shifts-scalar.ll
Modified:
llvm/lib/Target/ARM/ARMISelLowering.cpp
llvm/test/CodeGen/Thumb2/mve-shifts.ll
Removed:
################################################################################
diff --git a/llvm/lib/Target/ARM/ARMISelLowering.cpp b/llvm/lib/Target/ARM/ARMISelLowering.cpp
index c08f748c1297..e38ce70a8dc6 100644
--- a/llvm/lib/Target/ARM/ARMISelLowering.cpp
+++ b/llvm/lib/Target/ARM/ARMISelLowering.cpp
@@ -11814,7 +11814,8 @@ static SDValue PerformADDCombine(SDNode *N,
/// PerformSUBCombine - Target-specific dag combine xforms for ISD::SUB.
///
static SDValue PerformSUBCombine(SDNode *N,
- TargetLowering::DAGCombinerInfo &DCI) {
+ TargetLowering::DAGCombinerInfo &DCI,
+ const ARMSubtarget *Subtarget) {
SDValue N0 = N->getOperand(0);
SDValue N1 = N->getOperand(1);
@@ -11823,7 +11824,28 @@ static SDValue PerformSUBCombine(SDNode *N,
if (SDValue Result = combineSelectAndUse(N, N1, N0, DCI))
return Result;
- return SDValue();
+ if (!Subtarget->hasMVEIntegerOps() || !N->getValueType(0).isVector())
+ return SDValue();
+
+ // Fold (sub (ARMvmovImm 0), (ARMvdup x)) -> (ARMvdup (sub 0, x))
+ // so that we can readily pattern match more mve instructions which can use
+ // a scalar operand.
+ SDValue VDup = N->getOperand(1);
+ if (VDup->getOpcode() != ARMISD::VDUP)
+ return SDValue();
+
+ SDValue VMov = N->getOperand(0);
+ if (VMov->getOpcode() == ISD::BITCAST)
+ VMov = VMov->getOperand(0);
+
+ if (VMov->getOpcode() != ARMISD::VMOVIMM || !isZeroVector(VMov))
+ return SDValue();
+
+ SDLoc dl(N);
+ SDValue Negate = DCI.DAG.getNode(ISD::SUB, dl, MVT::i32,
+ DCI.DAG.getConstant(0, dl, MVT::i32),
+ VDup->getOperand(0));
+ return DCI.DAG.getNode(ARMISD::VDUP, dl, N->getValueType(0), Negate);
}
/// PerformVMULCombine
@@ -14508,7 +14530,7 @@ SDValue ARMTargetLowering::PerformDAGCombine(SDNode *N,
case ARMISD::ADDE: return PerformADDECombine(N, DCI, Subtarget);
case ARMISD::UMLAL: return PerformUMLALCombine(N, DCI.DAG, Subtarget);
case ISD::ADD: return PerformADDCombine(N, DCI, Subtarget);
- case ISD::SUB: return PerformSUBCombine(N, DCI);
+ case ISD::SUB: return PerformSUBCombine(N, DCI, Subtarget);
case ISD::MUL: return PerformMULCombine(N, DCI, Subtarget);
case ISD::OR: return PerformORCombine(N, DCI, Subtarget);
case ISD::XOR: return PerformXORCombine(N, DCI, Subtarget);
@@ -14866,6 +14888,9 @@ bool ARMTargetLowering::shouldSinkOperands(Instruction *I,
case Instruction::Mul:
return true;
case Instruction::Sub:
+ case Instruction::Shl:
+ case Instruction::LShr:
+ case Instruction::AShr:
return Operand == 1;
default:
return false;
diff --git a/llvm/test/CodeGen/Thumb2/mve-shifts-scalar.ll b/llvm/test/CodeGen/Thumb2/mve-shifts-scalar.ll
new file mode 100644
index 000000000000..d2e8896d77d9
--- /dev/null
+++ b/llvm/test/CodeGen/Thumb2/mve-shifts-scalar.ll
@@ -0,0 +1,422 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -O3 -mtriple=thumbv8.1m.main-arm-none-eabi -mattr=+mve %s --verify-machineinstrs -o - | FileCheck %s
+
+define dso_local arm_aapcs_vfpcc void @sink_shl_i32(i32* nocapture readonly %in, i32* noalias nocapture %out, i32 %shift, i32 %N) {
+; CHECK-LABEL: sink_shl_i32:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #16
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #16
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB0_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #16]!
+; CHECK-NEXT: vshl.u32 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #16]!
+; CHECK-NEXT: le lr, .LBB0_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <4 x i32> undef, i32 %shift, i32 0
+ %broadcast.splat11 = shufflevector <4 x i32> %broadcast.splatinsert10, <4 x i32> undef, <4 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i32, i32* %in, i32 %index
+ %cast.in = bitcast i32* %gep.in to <4 x i32>*
+ %wide.load = load <4 x i32>, <4 x i32>* %cast.in, align 4
+ %res = shl <4 x i32> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i32, i32* %out, i32 %index
+ %cast.out = bitcast i32* %gep.out to <4 x i32>*
+ store <4 x i32> %res, <4 x i32>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_shl_i16(i16* nocapture readonly %in, i16* noalias nocapture %out, i16 %shift, i32 %N) {
+; CHECK-LABEL: sink_shl_i16:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #8
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #8
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB1_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #8]!
+; CHECK-NEXT: vshl.u16 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #8]!
+; CHECK-NEXT: le lr, .LBB1_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <8 x i16> undef, i16 %shift, i32 0
+ %broadcast.splat11 = shufflevector <8 x i16> %broadcast.splatinsert10, <8 x i16> undef, <8 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i16, i16* %in, i32 %index
+ %cast.in = bitcast i16* %gep.in to <8 x i16>*
+ %wide.load = load <8 x i16>, <8 x i16>* %cast.in, align 4
+ %res = shl <8 x i16> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i16, i16* %out, i32 %index
+ %cast.out = bitcast i16* %gep.out to <8 x i16>*
+ store <8 x i16> %res, <8 x i16>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_shl_i8(i8* nocapture readonly %in, i8* noalias nocapture %out, i8 %shift, i32 %N) {
+; CHECK-LABEL: sink_shl_i8:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #4
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #4
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB2_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #4]!
+; CHECK-NEXT: vshl.u8 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #4]!
+; CHECK-NEXT: le lr, .LBB2_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <16 x i8> undef, i8 %shift, i32 0
+ %broadcast.splat11 = shufflevector <16 x i8> %broadcast.splatinsert10, <16 x i8> undef, <16 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i8, i8* %in, i32 %index
+ %cast.in = bitcast i8* %gep.in to <16 x i8>*
+ %wide.load = load <16 x i8>, <16 x i8>* %cast.in, align 4
+ %res = shl <16 x i8> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i8, i8* %out, i32 %index
+ %cast.out = bitcast i8* %gep.out to <16 x i8>*
+ store <16 x i8> %res, <16 x i8>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_lshr_i32(i32* nocapture readonly %in, i32* noalias nocapture %out, i32 %shift, i32 %N) {
+; CHECK-LABEL: sink_lshr_i32:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #16
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #16
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB3_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #16]!
+; CHECK-NEXT: vshl.u32 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #16]!
+; CHECK-NEXT: le lr, .LBB3_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <4 x i32> undef, i32 %shift, i32 0
+ %broadcast.splat11 = shufflevector <4 x i32> %broadcast.splatinsert10, <4 x i32> undef, <4 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i32, i32* %in, i32 %index
+ %cast.in = bitcast i32* %gep.in to <4 x i32>*
+ %wide.load = load <4 x i32>, <4 x i32>* %cast.in, align 4
+ %res = lshr <4 x i32> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i32, i32* %out, i32 %index
+ %cast.out = bitcast i32* %gep.out to <4 x i32>*
+ store <4 x i32> %res, <4 x i32>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_lshr_i16(i16* nocapture readonly %in, i16* noalias nocapture %out, i16 %shift, i32 %N) {
+; CHECK-LABEL: sink_lshr_i16:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #8
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #8
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB4_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #8]!
+; CHECK-NEXT: vshl.u16 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #8]!
+; CHECK-NEXT: le lr, .LBB4_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <8 x i16> undef, i16 %shift, i32 0
+ %broadcast.splat11 = shufflevector <8 x i16> %broadcast.splatinsert10, <8 x i16> undef, <8 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i16, i16* %in, i32 %index
+ %cast.in = bitcast i16* %gep.in to <8 x i16>*
+ %wide.load = load <8 x i16>, <8 x i16>* %cast.in, align 4
+ %res = lshr <8 x i16> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i16, i16* %out, i32 %index
+ %cast.out = bitcast i16* %gep.out to <8 x i16>*
+ store <8 x i16> %res, <8 x i16>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_lshr_i8(i8* nocapture readonly %in, i8* noalias nocapture %out, i8 %shift, i32 %N) {
+; CHECK-LABEL: sink_lshr_i8:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #4
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #4
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB5_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #4]!
+; CHECK-NEXT: vshl.u8 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #4]!
+; CHECK-NEXT: le lr, .LBB5_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <16 x i8> undef, i8 %shift, i32 0
+ %broadcast.splat11 = shufflevector <16 x i8> %broadcast.splatinsert10, <16 x i8> undef, <16 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i8, i8* %in, i32 %index
+ %cast.in = bitcast i8* %gep.in to <16 x i8>*
+ %wide.load = load <16 x i8>, <16 x i8>* %cast.in, align 4
+ %res = lshr <16 x i8> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i8, i8* %out, i32 %index
+ %cast.out = bitcast i8* %gep.out to <16 x i8>*
+ store <16 x i8> %res, <16 x i8>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_ashr_i32(i32* nocapture readonly %in, i32* noalias nocapture %out, i32 %shift, i32 %N) {
+; CHECK-LABEL: sink_ashr_i32:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #16
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #16
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB6_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #16]!
+; CHECK-NEXT: vshl.s32 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #16]!
+; CHECK-NEXT: le lr, .LBB6_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <4 x i32> undef, i32 %shift, i32 0
+ %broadcast.splat11 = shufflevector <4 x i32> %broadcast.splatinsert10, <4 x i32> undef, <4 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i32, i32* %in, i32 %index
+ %cast.in = bitcast i32* %gep.in to <4 x i32>*
+ %wide.load = load <4 x i32>, <4 x i32>* %cast.in, align 4
+ %res = ashr <4 x i32> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i32, i32* %out, i32 %index
+ %cast.out = bitcast i32* %gep.out to <4 x i32>*
+ store <4 x i32> %res, <4 x i32>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_ashr_i16(i16* nocapture readonly %in, i16* noalias nocapture %out, i16 %shift, i32 %N) {
+; CHECK-LABEL: sink_ashr_i16:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #8
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #8
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB7_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #8]!
+; CHECK-NEXT: vshl.s16 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #8]!
+; CHECK-NEXT: le lr, .LBB7_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <8 x i16> undef, i16 %shift, i32 0
+ %broadcast.splat11 = shufflevector <8 x i16> %broadcast.splatinsert10, <8 x i16> undef, <8 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i16, i16* %in, i32 %index
+ %cast.in = bitcast i16* %gep.in to <8 x i16>*
+ %wide.load = load <8 x i16>, <8 x i16>* %cast.in, align 4
+ %res = ashr <8 x i16> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i16, i16* %out, i32 %index
+ %cast.out = bitcast i16* %gep.out to <8 x i16>*
+ store <8 x i16> %res, <8 x i16>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
+
+define dso_local arm_aapcs_vfpcc void @sink_ashr_i8(i8* nocapture readonly %in, i8* noalias nocapture %out, i8 %shift, i32 %N) {
+; CHECK-LABEL: sink_ashr_i8:
+; CHECK: @ %bb.0: @ %entry
+; CHECK-NEXT: .save {r7, lr}
+; CHECK-NEXT: push {r7, lr}
+; CHECK-NEXT: bic r3, r3, #3
+; CHECK-NEXT: subs r0, #4
+; CHECK-NEXT: sub.w r12, r3, #4
+; CHECK-NEXT: movs r3, #1
+; CHECK-NEXT: subs r1, #4
+; CHECK-NEXT: rsbs r2, r2, #0
+; CHECK-NEXT: add.w lr, r3, r12, lsr #2
+; CHECK-NEXT: dls lr, lr
+; CHECK-NEXT: .LBB8_1: @ %vector.body
+; CHECK-NEXT: @ =>This Inner Loop Header: Depth=1
+; CHECK-NEXT: vldrw.u32 q0, [r0, #4]!
+; CHECK-NEXT: vshl.s8 q0, r2
+; CHECK-NEXT: vstrb.8 q0, [r1, #4]!
+; CHECK-NEXT: le lr, .LBB8_1
+; CHECK-NEXT: @ %bb.2: @ %exit
+; CHECK-NEXT: pop {r7, pc}
+entry:
+ br label %vector.ph
+
+vector.ph:
+ %n.vec = and i32 %N, -4
+ %broadcast.splatinsert10 = insertelement <16 x i8> undef, i8 %shift, i32 0
+ %broadcast.splat11 = shufflevector <16 x i8> %broadcast.splatinsert10, <16 x i8> undef, <16 x i32> zeroinitializer
+ br label %vector.body
+
+vector.body: ; preds = %vector.body, %vector.ph
+ %index = phi i32 [ 0, %vector.ph ], [ %index.next, %vector.body ]
+ %gep.in = getelementptr inbounds i8, i8* %in, i32 %index
+ %cast.in = bitcast i8* %gep.in to <16 x i8>*
+ %wide.load = load <16 x i8>, <16 x i8>* %cast.in, align 4
+ %res = ashr <16 x i8> %wide.load, %broadcast.splat11
+ %gep.out = getelementptr inbounds i8, i8* %out, i32 %index
+ %cast.out = bitcast i8* %gep.out to <16 x i8>*
+ store <16 x i8> %res, <16 x i8>* %cast.out, align 4
+ %index.next = add i32 %index, 4
+ %cmp = icmp eq i32 %index.next, %n.vec
+ br i1 %cmp, label %exit, label %vector.body
+
+exit:
+ ret void
+}
diff --git a/llvm/test/CodeGen/Thumb2/mve-shifts.ll b/llvm/test/CodeGen/Thumb2/mve-shifts.ll
index a321c2dd3833..69e91a54c9ab 100644
--- a/llvm/test/CodeGen/Thumb2/mve-shifts.ll
+++ b/llvm/test/CodeGen/Thumb2/mve-shifts.ll
@@ -383,9 +383,8 @@ entry:
define arm_aapcs_vfpcc <16 x i8> @shru_qr_int8_t(<16 x i8> %src1, i8 %src2) {
; CHECK-LABEL: shru_qr_int8_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.8 q1, r0
-; CHECK-NEXT: vneg.s8 q1, q1
-; CHECK-NEXT: vshl.u8 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.u8 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <16 x i8> undef, i8 %src2, i32 0
@@ -397,9 +396,8 @@ entry:
define arm_aapcs_vfpcc <8 x i16> @shru_qr_int16_t(<8 x i16> %src1, i16 %src2) {
; CHECK-LABEL: shru_qr_int16_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.16 q1, r0
-; CHECK-NEXT: vneg.s16 q1, q1
-; CHECK-NEXT: vshl.u16 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.u16 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <8 x i16> undef, i16 %src2, i32 0
@@ -411,9 +409,8 @@ entry:
define arm_aapcs_vfpcc <4 x i32> @shru_qr_int32_t(<4 x i32> %src1, i32 %src2) {
; CHECK-LABEL: shru_qr_int32_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.32 q1, r0
-; CHECK-NEXT: vneg.s32 q1, q1
-; CHECK-NEXT: vshl.u32 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.u32 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <4 x i32> undef, i32 %src2, i32 0
@@ -449,9 +446,8 @@ entry:
define arm_aapcs_vfpcc <16 x i8> @shrs_qr_int8_t(<16 x i8> %src1, i8 %src2) {
; CHECK-LABEL: shrs_qr_int8_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.8 q1, r0
-; CHECK-NEXT: vneg.s8 q1, q1
-; CHECK-NEXT: vshl.s8 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.s8 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <16 x i8> undef, i8 %src2, i32 0
@@ -463,9 +459,8 @@ entry:
define arm_aapcs_vfpcc <8 x i16> @shrs_qr_int16_t(<8 x i16> %src1, i16 %src2) {
; CHECK-LABEL: shrs_qr_int16_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.16 q1, r0
-; CHECK-NEXT: vneg.s16 q1, q1
-; CHECK-NEXT: vshl.s16 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.s16 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <8 x i16> undef, i16 %src2, i32 0
@@ -477,9 +472,8 @@ entry:
define arm_aapcs_vfpcc <4 x i32> @shrs_qr_int32_t(<4 x i32> %src1, i32 %src2) {
; CHECK-LABEL: shrs_qr_int32_t:
; CHECK: @ %bb.0: @ %entry
-; CHECK-NEXT: vdup.32 q1, r0
-; CHECK-NEXT: vneg.s32 q1, q1
-; CHECK-NEXT: vshl.s32 q0, q0, q1
+; CHECK-NEXT: rsbs r0, r0, #0
+; CHECK-NEXT: vshl.s32 q0, r0
; CHECK-NEXT: bx lr
entry:
%i = insertelement <4 x i32> undef, i32 %src2, i32 0
More information about the llvm-commits
mailing list