[llvm] r216475 - Revert r210342 and r210343, add test case for the crasher.
Joerg Sonnenberger
joerg at bec.de
Tue Aug 26 12:06:42 PDT 2014
Author: joerg
Date: Tue Aug 26 14:06:41 2014
New Revision: 216475
URL: http://llvm.org/viewvc/llvm-project?rev=216475&view=rev
Log:
Revert r210342 and r210343, add test case for the crasher.
PR 20642.
Added:
llvm/trunk/test/Transforms/SLPVectorizer/X86/crash_gep.ll
Removed:
llvm/trunk/test/Transforms/SLPVectorizer/X86/gep.ll
Modified:
llvm/trunk/lib/Transforms/Vectorize/SLPVectorizer.cpp
Modified: llvm/trunk/lib/Transforms/Vectorize/SLPVectorizer.cpp
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/lib/Transforms/Vectorize/SLPVectorizer.cpp?rev=216475&r1=216474&r2=216475&view=diff
==============================================================================
--- llvm/trunk/lib/Transforms/Vectorize/SLPVectorizer.cpp (original)
+++ llvm/trunk/lib/Transforms/Vectorize/SLPVectorizer.cpp Tue Aug 26 14:06:41 2014
@@ -1179,54 +1179,6 @@ void BoUpSLP::buildTree_rec(ArrayRef<Val
}
return;
}
- case Instruction::GetElementPtr: {
- // We don't combine GEPs with complicated (nested) indexing.
- for (unsigned j = 0; j < VL.size(); ++j) {
- if (cast<Instruction>(VL[j])->getNumOperands() != 2) {
- DEBUG(dbgs() << "SLP: not-vectorizable GEP (nested indexes).\n");
- BS.cancelScheduling(VL);
- newTreeEntry(VL, false);
- return;
- }
- }
-
- // We can't combine several GEPs into one vector if they operate on
- // different types.
- Type *Ty0 = cast<Instruction>(VL0)->getOperand(0)->getType();
- for (unsigned j = 0; j < VL.size(); ++j) {
- Type *CurTy = cast<Instruction>(VL[j])->getOperand(0)->getType();
- if (Ty0 != CurTy) {
- DEBUG(dbgs() << "SLP: not-vectorizable GEP (different types).\n");
- BS.cancelScheduling(VL);
- newTreeEntry(VL, false);
- return;
- }
- }
-
- // We don't combine GEPs with non-constant indexes.
- for (unsigned j = 0; j < VL.size(); ++j) {
- auto Op = cast<Instruction>(VL[j])->getOperand(1);
- if (!isa<ConstantInt>(Op)) {
- DEBUG(
- dbgs() << "SLP: not-vectorizable GEP (non-constant indexes).\n");
- BS.cancelScheduling(VL);
- newTreeEntry(VL, false);
- return;
- }
- }
-
- newTreeEntry(VL, true);
- DEBUG(dbgs() << "SLP: added a vector of GEPs.\n");
- for (unsigned i = 0, e = 2; i < e; ++i) {
- ValueList Operands;
- // Prepare the operand vector.
- for (unsigned j = 0; j < VL.size(); ++j)
- Operands.push_back(cast<Instruction>(VL[j])->getOperand(i));
-
- buildTree_rec(Operands, Depth + 1);
- }
- return;
- }
case Instruction::Store: {
// Check if the stores are consecutive or of we need to swizzle them.
for (unsigned i = 0, e = VL.size() - 1; i < e; ++i)
@@ -1464,20 +1416,6 @@ int BoUpSLP::getEntryCost(TreeEntry *E)
}
return VecCost - ScalarCost;
}
- case Instruction::GetElementPtr: {
- TargetTransformInfo::OperandValueKind Op1VK =
- TargetTransformInfo::OK_AnyValue;
- TargetTransformInfo::OperandValueKind Op2VK =
- TargetTransformInfo::OK_UniformConstantValue;
-
- int ScalarCost =
- VecTy->getNumElements() *
- TTI->getArithmeticInstrCost(Instruction::Add, ScalarTy, Op1VK, Op2VK);
- int VecCost =
- TTI->getArithmeticInstrCost(Instruction::Add, VecTy, Op1VK, Op2VK);
-
- return VecCost - ScalarCost;
- }
case Instruction::Load: {
// Cost of wide load - cost of scalar loads.
int ScalarLdCost = VecTy->getNumElements() *
@@ -2044,35 +1982,6 @@ Value *BoUpSLP::vectorizeTree(TreeEntry
++NumVectorInstructions;
return propagateMetadata(S, E->Scalars);
}
- case Instruction::GetElementPtr: {
- setInsertPointAfterBundle(E->Scalars);
-
- ValueList Op0VL;
- for (int i = 0, e = E->Scalars.size(); i < e; ++i)
- Op0VL.push_back(cast<GetElementPtrInst>(E->Scalars[i])->getOperand(0));
-
- Value *Op0 = vectorizeTree(Op0VL);
-
- std::vector<Value *> OpVecs;
- for (int j = 1, e = cast<GetElementPtrInst>(VL0)->getNumOperands(); j < e;
- ++j) {
- ValueList OpVL;
- for (int i = 0, e = E->Scalars.size(); i < e; ++i)
- OpVL.push_back(cast<GetElementPtrInst>(E->Scalars[i])->getOperand(j));
-
- Value *OpVec = vectorizeTree(OpVL);
- OpVecs.push_back(OpVec);
- }
-
- Value *V = Builder.CreateGEP(Op0, OpVecs);
- E->VectorizedValue = V;
- ++NumVectorInstructions;
-
- if (Instruction *I = dyn_cast<Instruction>(V))
- return propagateMetadata(I, E->Scalars);
-
- return V;
- }
case Instruction::Call: {
CallInst *CI = cast<CallInst>(VL0);
setInsertPointAfterBundle(E->Scalars);
Added: llvm/trunk/test/Transforms/SLPVectorizer/X86/crash_gep.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Transforms/SLPVectorizer/X86/crash_gep.ll?rev=216475&view=auto
==============================================================================
--- llvm/trunk/test/Transforms/SLPVectorizer/X86/crash_gep.ll (added)
+++ llvm/trunk/test/Transforms/SLPVectorizer/X86/crash_gep.ll Tue Aug 26 14:06:41 2014
@@ -0,0 +1,19 @@
+; RUN: opt < %s -basicaa -slp-vectorizer -dce -S -mtriple=x86_64-unknown-linux-gnu
+
+target datalayout = "e-m:e-i64:64-f80:128-n8:16:32:64-S128"
+target triple = "x86_64-unknown-linux-gnu"
+
+ at a = common global i64* null, align 8
+
+; Function Attrs: nounwind uwtable
+define i32 @fn1() {
+entry:
+ %0 = load i64** @a, align 8
+ %add.ptr = getelementptr inbounds i64* %0, i64 1
+ %1 = ptrtoint i64* %add.ptr to i64
+ %arrayidx = getelementptr inbounds i64* %0, i64 2
+ store i64 %1, i64* %arrayidx, align 8
+ %2 = ptrtoint i64* %arrayidx to i64
+ store i64 %2, i64* %add.ptr, align 8
+ ret i32 undef
+}
Removed: llvm/trunk/test/Transforms/SLPVectorizer/X86/gep.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Transforms/SLPVectorizer/X86/gep.ll?rev=216474&view=auto
==============================================================================
--- llvm/trunk/test/Transforms/SLPVectorizer/X86/gep.ll (original)
+++ llvm/trunk/test/Transforms/SLPVectorizer/X86/gep.ll (removed)
@@ -1,41 +0,0 @@
-; RUN: opt < %s -basicaa -slp-vectorizer -S |FileCheck %s
-target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128"
-
-; Test if SLP can handle GEP expressions.
-; The test perform the following action:
-; x->first = y->first + 16
-; x->second = y->second + 16
-
-; CHECK-LABEL: foo1
-; CHECK: <2 x i32*>
-define void @foo1 ({ i32*, i32* }* noalias %x, { i32*, i32* }* noalias %y) {
- %1 = getelementptr inbounds { i32*, i32* }* %y, i64 0, i32 0
- %2 = load i32** %1, align 8
- %3 = getelementptr inbounds i32* %2, i64 16
- %4 = getelementptr inbounds { i32*, i32* }* %x, i64 0, i32 0
- store i32* %3, i32** %4, align 8
- %5 = getelementptr inbounds { i32*, i32* }* %y, i64 0, i32 1
- %6 = load i32** %5, align 8
- %7 = getelementptr inbounds i32* %6, i64 16
- %8 = getelementptr inbounds { i32*, i32* }* %x, i64 0, i32 1
- store i32* %7, i32** %8, align 8
- ret void
-}
-
-; Test that we don't vectorize GEP expressions if indexes are not constants.
-; We can't produce an efficient code in that case.
-; CHECK-LABEL: foo2
-; CHECK-NOT: <2 x i32*>
-define void @foo2 ({ i32*, i32* }* noalias %x, { i32*, i32* }* noalias %y, i32 %i) {
- %1 = getelementptr inbounds { i32*, i32* }* %y, i64 0, i32 0
- %2 = load i32** %1, align 8
- %3 = getelementptr inbounds i32* %2, i32 %i
- %4 = getelementptr inbounds { i32*, i32* }* %x, i64 0, i32 0
- store i32* %3, i32** %4, align 8
- %5 = getelementptr inbounds { i32*, i32* }* %y, i64 0, i32 1
- %6 = load i32** %5, align 8
- %7 = getelementptr inbounds i32* %6, i32 %i
- %8 = getelementptr inbounds { i32*, i32* }* %x, i64 0, i32 1
- store i32* %7, i32** %8, align 8
- ret void
-}
More information about the llvm-commits
mailing list