[llvm] r261804 - Detecte vector reduction operations just before instruction selection.
Justin Bogner via llvm-commits
llvm-commits at lists.llvm.org
Fri Mar 18 17:07:05 PDT 2016
Cong Hou via llvm-commits <llvm-commits at lists.llvm.org> writes:
> Author: conghou
> Date: Wed Feb 24 17:40:36 2016
> New Revision: 261804
>
> URL: http://llvm.org/viewvc/llvm-project?rev=261804&view=rev
> Log:
> Detecte vector reduction operations just before instruction selection.
>
> (This is the second attemp to commit this patch, after fixing pr26652
> & pr26653).
>
> This patch detects vector reductions before instruction selection. Vector
> reductions are vectorized reduction operations, and for such operations we have
> freedom to reorganize the elements of the result as long as the
> reduction of them
> stay unchanged. This will enable some reduction pattern recognition during
> instruction combine such as SAD/dot-product on X86. A flag is added to
> SDNodeFlags to mark those vector reduction nodes to be checked during
> instruction
> combine.
>
> To detect those vector reductions, we search def-use chains starting from the
> given instruction, and check if all uses fall into two categories:
>
> 1. Reduction with another vector.
> 2. Reduction on all elements.
>
> in which 2 is detected by recognizing the pattern that the loop vectorizer
> generates to reduce all elements in the vector outside of the loop, which
> includes several ShuffleVector and one ExtractElement instructions.
>
>
> Differential revision: http://reviews.llvm.org/D15250
>
>
>
> Added:
> llvm/trunk/test/CodeGen/Generic/pr26652.ll
> llvm/trunk/test/CodeGen/Generic/vector-redux.ll
> Modified:
> llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h
> llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAGBuilder.cpp
>
> Modified: llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h
> URL:
> http://llvm.org/viewvc/llvm-project/llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h?rev=261804&r1=261803&r2=261804&view=diff
> ==============================================================================
> --- llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h (original)
> +++ llvm/trunk/include/llvm/CodeGen/SelectionDAGNodes.h Wed Feb 24 17:40:36 2016
> @@ -328,6 +328,7 @@ private:
> bool NoInfs : 1;
> bool NoSignedZeros : 1;
> bool AllowReciprocal : 1;
> + bool VectorReduction : 1;
This brings the number of bits used here to 9, which no longer fits in
one bool and doubles the size of SDNodeFlags. That's kind of unfortunate.
>
> public:
> /// Default constructor turns off all optimization flags.
> @@ -340,6 +341,7 @@ public:
> NoInfs = false;
> NoSignedZeros = false;
> AllowReciprocal = false;
> + VectorReduction = false;
> }
>
> // These are mutators for each flag.
> @@ -351,6 +353,7 @@ public:
> void setNoInfs(bool b) { NoInfs = b; }
> void setNoSignedZeros(bool b) { NoSignedZeros = b; }
> void setAllowReciprocal(bool b) { AllowReciprocal = b; }
> + void setVectorReduction(bool b) { VectorReduction = b; }
>
> // These are accessors for each flag.
> bool hasNoUnsignedWrap() const { return NoUnsignedWrap; }
> @@ -361,6 +364,7 @@ public:
> bool hasNoInfs() const { return NoInfs; }
> bool hasNoSignedZeros() const { return NoSignedZeros; }
> bool hasAllowReciprocal() const { return AllowReciprocal; }
> + bool hasVectorReduction() const { return VectorReduction; }
>
> /// Return a raw encoding of the flags.
> /// This function should only be used to add data to the NodeID value.
>
> Modified: llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAGBuilder.cpp
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAGBuilder.cpp?rev=261804&r1=261803&r2=261804&view=diff
> ==============================================================================
> --- llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAGBuilder.cpp (original)
> +++ llvm/trunk/lib/CodeGen/SelectionDAG/SelectionDAGBuilder.cpp Wed Feb 24 17:40:36 2016
> @@ -2317,6 +2317,129 @@ void SelectionDAGBuilder::visitFSub(cons
> visitBinary(I, ISD::FSUB);
> }
>
> +/// Checks if the given instruction performs a vector reduction, in which case
> +/// we have the freedom to alter the elements in the result as long as the
> +/// reduction of them stays unchanged.
> +static bool isVectorReductionOp(const User *I) {
> + const Instruction *Inst = dyn_cast<Instruction>(I);
> + if (!Inst || !Inst->getType()->isVectorTy())
> + return false;
> +
> + auto OpCode = Inst->getOpcode();
It's best not to use auto for short, simple types. This code is clearer
if you just say "unsigned" here.
> + switch (OpCode) {
> + case Instruction::Add:
> + case Instruction::Mul:
> + case Instruction::And:
> + case Instruction::Or:
> + case Instruction::Xor:
> + break;
> + case Instruction::FAdd:
> + case Instruction::FMul:
> + if (const FPMathOperator *FPOp = dyn_cast<const FPMathOperator>(Inst))
> + if (FPOp->getFastMathFlags().unsafeAlgebra())
> + break;
> + // Fall through.
> + default:
> + return false;
> + }
> +
> + unsigned ElemNum = Inst->getType()->getVectorNumElements();
> + unsigned ElemNumToReduce = ElemNum;
> +
> + // Do DFS search on the def-use chain from the given instruction. We only
> + // allow four kinds of operations during the search until we reach the
> + // instruction that extracts the first element from the vector:
> + //
> + // 1. The reduction operation of the same opcode as the given instruction.
> + //
> + // 2. PHI node.
> + //
> + // 3. ShuffleVector instruction together with a reduction operation that
> + // does a partial reduction.
> + //
> + // 4. ExtractElement that extracts the first element from the vector, and we
> + // stop searching the def-use chain here.
> + //
> + // 3 & 4 above perform a reduction on all elements of the vector. We push defs
> + // from 1-3 to the stack to continue the DFS. The given instruction is not
> + // a reduction operation if we meet any other instructions other than those
> + // listed above.
> +
> + SmallVector<const User *, 16> UsersToVisit{Inst};
> + SmallPtrSet<const User *, 16> Visited;
> + bool ReduxExtracted = false;
> +
> + while (!UsersToVisit.empty()) {
> + auto User = UsersToVisit.back();
We generally include `const` and * or & where applicable on auto types,
so this should be `const auto *`. That said, auto doesn't save any
characters here, so `const User *` is actually clearer. You'll have to
give the variable a more specific name if you do that, of course.
> + UsersToVisit.pop_back();
> + if (!Visited.insert(User).second)
> + continue;
> +
> + for (const auto &U : User->users()) {
This does a copy. You meant to say `const auto *U`. I'd probably type
User here too.
> + auto Inst = dyn_cast<Instruction>(U);
> + if (!Inst)
> + return false;
> +
> + if (Inst->getOpcode() == OpCode || isa<PHINode>(U)) {
> + if (const FPMathOperator *FPOp = dyn_cast<const FPMathOperator>(Inst))
> + if (!isa<PHINode>(FPOp) && !FPOp->getFastMathFlags().unsafeAlgebra())
> + return false;
> + UsersToVisit.push_back(U);
> + } else if (const ShuffleVectorInst *ShufInst =
> + dyn_cast<ShuffleVectorInst>(U)) {
> + // Detect the following pattern: A ShuffleVector instruction together
> + // with a reduction that do partial reduction on the first and second
> + // ElemNumToReduce / 2 elements, and store the result in
> + // ElemNumToReduce / 2 elements in another vector.
> +
> + unsigned ResultElements = ShufInst->getType()->getVectorNumElements();
> + ElemNumToReduce = ResultElements <= ElemNumToReduce ? ResultElements
> + : ElemNumToReduce;
> + if (ElemNumToReduce == 1)
> + return false;
> + if (!isa<UndefValue>(U->getOperand(1)))
> + return false;
> + for (unsigned i = 0; i < ElemNumToReduce / 2; ++i)
> + if (ShufInst->getMaskValue(i) != int(i + ElemNumToReduce / 2))
Wouldn't it make more sense for all of these unsigned variables to be
ints? AFAICT we don't want wrapping behaviour for anything here.
> + return false;
> + for (unsigned i = ElemNumToReduce / 2; i < ElemNum; ++i)
> + if (ShufInst->getMaskValue(i) != -1)
> + return false;
> +
> + // There is only one user of this ShuffleVector instruction, which
> + // must
> + // be a reduction operation.
> + if (!U->hasOneUse())
> + return false;
> +
> + auto U2 = dyn_cast<Instruction>(*U->user_begin());
> + if (!U2 || U2->getOpcode() != OpCode)
> + return false;
> +
> + // Check operands of the reduction operation.
> + if ((U2->getOperand(0) == U->getOperand(0) && U2->getOperand(1) == U) ||
> + (U2->getOperand(1) == U->getOperand(0) && U2->getOperand(0) == U)) {
> + UsersToVisit.push_back(U2);
> + ElemNumToReduce /= 2;
> + } else
> + return false;
> + } else if (isa<ExtractElementInst>(U)) {
> + // At this moment we should have reduced all elements in the vector.
> + if (ElemNumToReduce != 1)
> + return false;
> +
> + const ConstantInt *Val = dyn_cast<ConstantInt>(U->getOperand(1));
> + if (!Val || Val->getZExtValue() != 0)
> + return false;
> +
> + ReduxExtracted = true;
> + } else
> + return false;
I find this loop with heavily nested statements returning false kind of
hard to read. Would restructuring this with some judicious use of
continue and a few less else clauses improve the readability here?
> + }
> + }
> + return ReduxExtracted;
> +}
> +
> void SelectionDAGBuilder::visitBinary(const User &I, unsigned OpCode) {
> SDValue Op1 = getValue(I.getOperand(0));
> SDValue Op2 = getValue(I.getOperand(1));
> @@ -2324,6 +2447,7 @@ void SelectionDAGBuilder::visitBinary(co
> bool nuw = false;
> bool nsw = false;
> bool exact = false;
> + bool vec_redux = false;
> FastMathFlags FMF;
>
> if (const OverflowingBinaryOperator *OFBinOp =
> @@ -2337,10 +2461,16 @@ void SelectionDAGBuilder::visitBinary(co
> if (const FPMathOperator *FPOp = dyn_cast<const FPMathOperator>(&I))
> FMF = FPOp->getFastMathFlags();
>
> + if (isVectorReductionOp(&I)) {
> + vec_redux = true;
> + DEBUG(dbgs() << "Detected a reduction operation:" << I << "\n");
> + }
> +
> SDNodeFlags Flags;
> Flags.setExact(exact);
> Flags.setNoSignedWrap(nsw);
> Flags.setNoUnsignedWrap(nuw);
> + Flags.setVectorReduction(vec_redux);
> if (EnableFMFInDAG) {
> Flags.setAllowReciprocal(FMF.allowReciprocal());
> Flags.setNoInfs(FMF.noInfs());
>
> Added: llvm/trunk/test/CodeGen/Generic/pr26652.ll
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/Generic/pr26652.ll?rev=261804&view=auto
> ==============================================================================
> --- llvm/trunk/test/CodeGen/Generic/pr26652.ll (added)
> +++ llvm/trunk/test/CodeGen/Generic/pr26652.ll Wed Feb 24 17:40:36 2016
> @@ -0,0 +1,8 @@
> +; RUN: llc < %s
> +
> +define <2 x i32> @test(<4 x i32> %a, <4 x i32> %b) {
> +entry:
> + %0 = or <4 x i32> %a, %b
> + %1 = shufflevector <4 x i32> %0, <4 x i32> undef, <2 x i32> <i32 2, i32 3>
> + ret <2 x i32> %1
> +}
>
> Added: llvm/trunk/test/CodeGen/Generic/vector-redux.ll
> URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/Generic/vector-redux.ll?rev=261804&view=auto
> ==============================================================================
> --- llvm/trunk/test/CodeGen/Generic/vector-redux.ll (added)
> +++ llvm/trunk/test/CodeGen/Generic/vector-redux.ll Wed Feb 24 17:40:36 2016
> @@ -0,0 +1,237 @@
> +; RUN: llc < %s -debug-only=isel -o /dev/null 2>&1 | FileCheck %s
> +; REQUIRES: asserts
> +
> + at a = global [1024 x i32] zeroinitializer, align 16
> +
> +define i32 @reduce_add() {
> +; CHECK-LABEL: reduce_add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +; CHECK: Detected a reduction operation: {{.*}} add
> +
> +min.iters.checked:
> + br label %vector.body
> +
> +vector.body:
> + %index = phi i64 [ 0, %min.iters.checked ], [ %index.next.4, %vector.body ]
> + %vec.phi = phi <4 x i32> [ zeroinitializer, %min.iters.checked ], [ %28, %vector.body ]
> + %vec.phi4 = phi <4 x i32> [ zeroinitializer, %min.iters.checked ], [ %29, %vector.body ]
> + %0 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index
> + %1 = bitcast i32* %0 to <4 x i32>*
> + %wide.load = load <4 x i32>, <4 x i32>* %1, align 16
> + %2 = getelementptr i32, i32* %0, i64 4
> + %3 = bitcast i32* %2 to <4 x i32>*
> + %wide.load5 = load <4 x i32>, <4 x i32>* %3, align 16
> + %4 = add nsw <4 x i32> %wide.load, %vec.phi
> + %5 = add nsw <4 x i32> %wide.load5, %vec.phi4
> + %index.next = add nuw nsw i64 %index, 8
> + %6 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next
> + %7 = bitcast i32* %6 to <4 x i32>*
> + %wide.load.1 = load <4 x i32>, <4 x i32>* %7, align 16
> + %8 = getelementptr i32, i32* %6, i64 4
> + %9 = bitcast i32* %8 to <4 x i32>*
> + %wide.load5.1 = load <4 x i32>, <4 x i32>* %9, align 16
> + %10 = add nsw <4 x i32> %wide.load.1, %4
> + %11 = add nsw <4 x i32> %wide.load5.1, %5
> + %index.next.1 = add nsw i64 %index, 16
> + %12 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.1
> + %13 = bitcast i32* %12 to <4 x i32>*
> + %wide.load.2 = load <4 x i32>, <4 x i32>* %13, align 16
> + %14 = getelementptr i32, i32* %12, i64 4
> + %15 = bitcast i32* %14 to <4 x i32>*
> + %wide.load5.2 = load <4 x i32>, <4 x i32>* %15, align 16
> + %16 = add nsw <4 x i32> %wide.load.2, %10
> + %17 = add nsw <4 x i32> %wide.load5.2, %11
> + %index.next.2 = add nsw i64 %index, 24
> + %18 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.2
> + %19 = bitcast i32* %18 to <4 x i32>*
> + %wide.load.3 = load <4 x i32>, <4 x i32>* %19, align 16
> + %20 = getelementptr i32, i32* %18, i64 4
> + %21 = bitcast i32* %20 to <4 x i32>*
> + %wide.load5.3 = load <4 x i32>, <4 x i32>* %21, align 16
> + %22 = add nsw <4 x i32> %wide.load.3, %16
> + %23 = add nsw <4 x i32> %wide.load5.3, %17
> + %index.next.3 = add nsw i64 %index, 32
> + %24 = getelementptr inbounds [1024 x i32], [1024 x i32]* @a, i64 0, i64 %index.next.3
> + %25 = bitcast i32* %24 to <4 x i32>*
> + %wide.load.4 = load <4 x i32>, <4 x i32>* %25, align 16
> + %26 = getelementptr i32, i32* %24, i64 4
> + %27 = bitcast i32* %26 to <4 x i32>*
> + %wide.load5.4 = load <4 x i32>, <4 x i32>* %27, align 16
> + %28 = add nsw <4 x i32> %wide.load.4, %22
> + %29 = add nsw <4 x i32> %wide.load5.4, %23
> + %index.next.4 = add nsw i64 %index, 40
> + %30 = icmp eq i64 %index.next.4, 1000
> + br i1 %30, label %middle.block, label %vector.body
> +
> +middle.block:
> + %.lcssa10 = phi <4 x i32> [ %29, %vector.body ]
> + %.lcssa = phi <4 x i32> [ %28, %vector.body ]
> + %bin.rdx = add <4 x i32> %.lcssa10, %.lcssa
> + %rdx.shuf = shufflevector <4 x i32> %bin.rdx, <4 x i32> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef>
> + %bin.rdx6 = add <4 x i32> %bin.rdx, %rdx.shuf
> + %rdx.shuf7 = shufflevector <4 x i32> %bin.rdx6, <4 x i32> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef>
> + %bin.rdx8 = add <4 x i32> %bin.rdx6, %rdx.shuf7
> + %31 = extractelement <4 x i32> %bin.rdx8, i32 0
> + ret i32 %31
> +}
> +
> +define i32 @reduce_and() {
> +; CHECK-LABEL: reduce_and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +; CHECK: Detected a reduction operation: {{.*}} and
> +
> +entry:
> + br label %vector.body
> +
> +vector.body:
> + %lsr.iv = phi i64 [ %lsr.iv.next, %vector.body ], [ -4096, %entry ]
> + %vec.phi = phi <4 x i32> [ <i32 -1, i32 -1, i32 -1, i32 -1>, %entry ], [ %6, %vector.body ]
> + %vec.phi9 = phi <4 x i32> [ <i32 -1, i32 -1, i32 -1, i32 -1>, %entry ], [ %7, %vector.body ]
> + %uglygep33 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep3334 = bitcast i8* %uglygep33 to <4 x i32>*
> + %scevgep35 = getelementptr <4 x i32>, <4 x i32>* %uglygep3334, i64 256
> + %wide.load = load <4 x i32>, <4 x i32>* %scevgep35, align 16
> + %scevgep36 = getelementptr <4 x i32>, <4 x i32>* %uglygep3334, i64 257
> + %wide.load10 = load <4 x i32>, <4 x i32>* %scevgep36, align 16
> + %0 = and <4 x i32> %wide.load, %vec.phi
> + %1 = and <4 x i32> %wide.load10, %vec.phi9
> + %uglygep30 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep3031 = bitcast i8* %uglygep30 to <4 x i32>*
> + %scevgep32 = getelementptr <4 x i32>, <4 x i32>* %uglygep3031, i64 258
> + %wide.load.1 = load <4 x i32>, <4 x i32>* %scevgep32, align 16
> + %uglygep27 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep2728 = bitcast i8* %uglygep27 to <4 x i32>*
> + %scevgep29 = getelementptr <4 x i32>, <4 x i32>* %uglygep2728, i64 259
> + %wide.load10.1 = load <4 x i32>, <4 x i32>* %scevgep29, align 16
> + %2 = and <4 x i32> %wide.load.1, %0
> + %3 = and <4 x i32> %wide.load10.1, %1
> + %uglygep24 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep2425 = bitcast i8* %uglygep24 to <4 x i32>*
> + %scevgep26 = getelementptr <4 x i32>, <4 x i32>* %uglygep2425, i64 260
> + %wide.load.2 = load <4 x i32>, <4 x i32>* %scevgep26, align 16
> + %uglygep21 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep2122 = bitcast i8* %uglygep21 to <4 x i32>*
> + %scevgep23 = getelementptr <4 x i32>, <4 x i32>* %uglygep2122, i64 261
> + %wide.load10.2 = load <4 x i32>, <4 x i32>* %scevgep23, align 16
> + %4 = and <4 x i32> %wide.load.2, %2
> + %5 = and <4 x i32> %wide.load10.2, %3
> + %uglygep18 = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep1819 = bitcast i8* %uglygep18 to <4 x i32>*
> + %scevgep20 = getelementptr <4 x i32>, <4 x i32>* %uglygep1819, i64 262
> + %wide.load.3 = load <4 x i32>, <4 x i32>* %scevgep20, align 16
> + %uglygep = getelementptr i8, i8* bitcast ([1024 x i32]* @a to i8*), i64 %lsr.iv
> + %uglygep17 = bitcast i8* %uglygep to <4 x i32>*
> + %scevgep = getelementptr <4 x i32>, <4 x i32>* %uglygep17, i64 263
> + %wide.load10.3 = load <4 x i32>, <4 x i32>* %scevgep, align 16
> + %6 = and <4 x i32> %wide.load.3, %4
> + %7 = and <4 x i32> %wide.load10.3, %5
> + %lsr.iv.next = add nsw i64 %lsr.iv, 128
> + %8 = icmp eq i64 %lsr.iv.next, 0
> + br i1 %8, label %middle.block, label %vector.body
> +
> +middle.block:
> + %bin.rdx = and <4 x i32> %7, %6
> + %rdx.shuf = shufflevector <4 x i32> %bin.rdx, <4 x i32> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef>
> + %bin.rdx11 = and <4 x i32> %bin.rdx, %rdx.shuf
> + %rdx.shuf12 = shufflevector <4 x i32> %bin.rdx11, <4 x i32> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef>
> + %bin.rdx13 = and <4 x i32> %bin.rdx11, %rdx.shuf12
> + %9 = extractelement <4 x i32> %bin.rdx13, i32 0
> + ret i32 %9
> +}
> +
> +define float @reduce_add_float(float* nocapture readonly %a) {
> +; CHECK-LABEL: reduce_add_float
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +; CHECK: Detected a reduction operation: {{.*}} fadd fast
> +;
> +entry:
> + br label %vector.body
> +
> +vector.body:
> + %index = phi i64 [ 0, %entry ], [ %index.next.4, %vector.body ]
> + %vec.phi = phi <4 x float> [ zeroinitializer, %entry ], [ %28, %vector.body ]
> + %vec.phi9 = phi <4 x float> [ zeroinitializer, %entry ], [ %29, %vector.body ]
> + %0 = getelementptr inbounds float, float* %a, i64 %index
> + %1 = bitcast float* %0 to <4 x float>*
> + %wide.load = load <4 x float>, <4 x float>* %1, align 4
> + %2 = getelementptr float, float* %0, i64 4
> + %3 = bitcast float* %2 to <4 x float>*
> + %wide.load10 = load <4 x float>, <4 x float>* %3, align 4
> + %4 = fadd fast <4 x float> %wide.load, %vec.phi
> + %5 = fadd fast <4 x float> %wide.load10, %vec.phi9
> + %index.next = add nuw nsw i64 %index, 8
> + %6 = getelementptr inbounds float, float* %a, i64 %index.next
> + %7 = bitcast float* %6 to <4 x float>*
> + %wide.load.1 = load <4 x float>, <4 x float>* %7, align 4
> + %8 = getelementptr float, float* %6, i64 4
> + %9 = bitcast float* %8 to <4 x float>*
> + %wide.load10.1 = load <4 x float>, <4 x float>* %9, align 4
> + %10 = fadd fast <4 x float> %wide.load.1, %4
> + %11 = fadd fast <4 x float> %wide.load10.1, %5
> + %index.next.1 = add nsw i64 %index, 16
> + %12 = getelementptr inbounds float, float* %a, i64 %index.next.1
> + %13 = bitcast float* %12 to <4 x float>*
> + %wide.load.2 = load <4 x float>, <4 x float>* %13, align 4
> + %14 = getelementptr float, float* %12, i64 4
> + %15 = bitcast float* %14 to <4 x float>*
> + %wide.load10.2 = load <4 x float>, <4 x float>* %15, align 4
> + %16 = fadd fast <4 x float> %wide.load.2, %10
> + %17 = fadd fast <4 x float> %wide.load10.2, %11
> + %index.next.2 = add nsw i64 %index, 24
> + %18 = getelementptr inbounds float, float* %a, i64 %index.next.2
> + %19 = bitcast float* %18 to <4 x float>*
> + %wide.load.3 = load <4 x float>, <4 x float>* %19, align 4
> + %20 = getelementptr float, float* %18, i64 4
> + %21 = bitcast float* %20 to <4 x float>*
> + %wide.load10.3 = load <4 x float>, <4 x float>* %21, align 4
> + %22 = fadd fast <4 x float> %wide.load.3, %16
> + %23 = fadd fast <4 x float> %wide.load10.3, %17
> + %index.next.3 = add nsw i64 %index, 32
> + %24 = getelementptr inbounds float, float* %a, i64 %index.next.3
> + %25 = bitcast float* %24 to <4 x float>*
> + %wide.load.4 = load <4 x float>, <4 x float>* %25, align 4
> + %26 = getelementptr float, float* %24, i64 4
> + %27 = bitcast float* %26 to <4 x float>*
> + %wide.load10.4 = load <4 x float>, <4 x float>* %27, align 4
> + %28 = fadd fast <4 x float> %wide.load.4, %22
> + %29 = fadd fast <4 x float> %wide.load10.4, %23
> + %index.next.4 = add nsw i64 %index, 40
> + %30 = icmp eq i64 %index.next.4, 1000
> + br i1 %30, label %middle.block, label %vector.body
> +
> +middle.block:
> + %.lcssa15 = phi <4 x float> [ %29, %vector.body ]
> + %.lcssa = phi <4 x float> [ %28, %vector.body ]
> + %bin.rdx = fadd fast <4 x float> %.lcssa15, %.lcssa
> + %rdx.shuf = shufflevector <4 x float> %bin.rdx, <4 x float> undef, <4 x i32> <i32 2, i32 3, i32 undef, i32 undef>
> + %bin.rdx11 = fadd fast <4 x float> %bin.rdx, %rdx.shuf
> + %rdx.shuf12 = shufflevector <4 x float> %bin.rdx11, <4 x float> undef, <4 x i32> <i32 1, i32 undef, i32 undef, i32 undef>
> + %bin.rdx13 = fadd fast <4 x float> %bin.rdx11, %rdx.shuf12
> + %31 = extractelement <4 x float> %bin.rdx13, i32 0
> + ret float %31
> +}
>
>
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at lists.llvm.org
> http://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-commits
More information about the llvm-commits
mailing list