[llvm] e65ddca - Revert "[SLP] Vectorize jumbled stores."

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Thu Oct 31 02:44:17 PDT 2019


Do you have a repro example of such a crash? Cheers, Simon.

On 31/10/2019 09:23, Haojian Wu via llvm-commits wrote:
> Author: Haojian Wu
> Date: 2019-10-31T10:21:24+01:00
> New Revision: e65ddcafee547eddb72aa63b1d49f76e46acdd9a
>
> URL: https://github.com/llvm/llvm-project/commit/e65ddcafee547eddb72aa63b1d49f76e46acdd9a
> DIFF: https://github.com/llvm/llvm-project/commit/e65ddcafee547eddb72aa63b1d49f76e46acdd9a.diff
>
> LOG: Revert "[SLP] Vectorize jumbled stores."
>
> This reverts commit 21d498c9c0f32dcab5bc89ac593aa813b533b43a.
>
> This commit causes some crashes on some targets.
>
> Added:
>      
>
> Modified:
>      llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp
>      llvm/test/Transforms/SLPVectorizer/X86/store-jumbled.ll
>      llvm/test/Transforms/SLPVectorizer/X86/stores_vectorize.ll
>
> Removed:
>      
>
>
> ################################################################################
> diff  --git a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp b/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp
> index ad860ccb4ef1..bdcd5bfb9815 100644
> --- a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp
> +++ b/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp
> @@ -2666,74 +2666,24 @@ void BoUpSLP::buildTree_rec(ArrayRef<Value *> VL, unsigned Depth,
>       }
>       case Instruction::Store: {
>         // Check if the stores are consecutive or if we need to swizzle them.
> -      llvm::Type *ScalarTy = cast<StoreInst>(VL0)->getValueOperand()->getType();
> -      // Make sure all stores in the bundle are simple - we can't vectorize
> -      // atomic or volatile stores.
> -      SmallVector<Value *, 4> PointerOps(VL.size());
> -      ValueList Operands(VL.size());
> -      auto POIter = PointerOps.begin();
> -      auto OIter = Operands.begin();
> -      for (Value *V : VL) {
> -        auto *SI = cast<StoreInst>(V);
> -        if (!SI->isSimple()) {
> +      for (unsigned i = 0, e = VL.size() - 1; i < e; ++i)
> +        if (!isConsecutiveAccess(VL[i], VL[i + 1], *DL, *SE)) {
>             BS.cancelScheduling(VL, VL0);
>             newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx,
>                          ReuseShuffleIndicies);
> -          LLVM_DEBUG(dbgs() << "SLP: Gathering non-simple stores.\n");
> +          LLVM_DEBUG(dbgs() << "SLP: Non-consecutive store.\n");
>             return;
>           }
> -        *POIter = SI->getPointerOperand();
> -        *OIter = SI->getValueOperand();
> -        ++POIter;
> -        ++OIter;
> -      }
>   
> -      OrdersType CurrentOrder;
> -      // Check the order of pointer operands.
> -      if (llvm::sortPtrAccesses(PointerOps, *DL, *SE, CurrentOrder)) {
> -        Value *Ptr0;
> -        Value *PtrN;
> -        if (CurrentOrder.empty()) {
> -          Ptr0 = PointerOps.front();
> -          PtrN = PointerOps.back();
> -        } else {
> -          Ptr0 = PointerOps[CurrentOrder.front()];
> -          PtrN = PointerOps[CurrentOrder.back()];
> -        }
> -        const SCEV *Scev0 = SE->getSCEV(Ptr0);
> -        const SCEV *ScevN = SE->getSCEV(PtrN);
> -        const auto *Diff =
> -            dyn_cast<SCEVConstant>(SE->getMinusSCEV(ScevN, Scev0));
> -        uint64_t Size = DL->getTypeAllocSize(ScalarTy);
> -        // Check that the sorted pointer operands are consecutive.
> -        if (Diff && Diff->getAPInt() == (VL.size() - 1) * Size) {
> -          if (CurrentOrder.empty()) {
> -            // Original stores are consecutive and does not require reordering.
> -            ++NumOpsWantToKeepOriginalOrder;
> -            TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S,
> -                                         UserTreeIdx, ReuseShuffleIndicies);
> -            TE->setOperandsInOrder();
> -            buildTree_rec(Operands, Depth + 1, {TE, 0});
> -            LLVM_DEBUG(dbgs() << "SLP: added a vector of stores.\n");
> -          } else {
> -            // Need to reorder.
> -            auto I = NumOpsWantToKeepOrder.try_emplace(CurrentOrder).first;
> -            ++(I->getSecond());
> -            TreeEntry *TE =
> -                newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx,
> -                             ReuseShuffleIndicies, I->getFirst());
> -            TE->setOperandsInOrder();
> -            buildTree_rec(Operands, Depth + 1, {TE, 0});
> -            LLVM_DEBUG(dbgs() << "SLP: added a vector of jumbled stores.\n");
> -          }
> -          return;
> -        }
> -      }
> +      TreeEntry *TE = newTreeEntry(VL, Bundle /*vectorized*/, S, UserTreeIdx,
> +                                   ReuseShuffleIndicies);
> +      LLVM_DEBUG(dbgs() << "SLP: added a vector of stores.\n");
>   
> -      BS.cancelScheduling(VL, VL0);
> -      newTreeEntry(VL, None /*not vectorized*/, S, UserTreeIdx,
> -                   ReuseShuffleIndicies);
> -      LLVM_DEBUG(dbgs() << "SLP: Non-consecutive store.\n");
> +      ValueList Operands;
> +      for (Value *V : VL)
> +        Operands.push_back(cast<Instruction>(V)->getOperand(0));
> +      TE->setOperandsInOrder();
> +      buildTree_rec(Operands, Depth + 1, {TE, 0});
>         return;
>       }
>       case Instruction::Call: {
> @@ -3231,23 +3181,15 @@ int BoUpSLP::getEntryCost(TreeEntry *E) {
>       }
>       case Instruction::Store: {
>         // We know that we can merge the stores. Calculate the cost.
> -      bool IsReorder = !E->ReorderIndices.empty();
> -      auto *SI =
> -          cast<StoreInst>(IsReorder ? VL[E->ReorderIndices.front()] : VL0);
> -      MaybeAlign Alignment(SI->getAlignment());
> +      MaybeAlign alignment(cast<StoreInst>(VL0)->getAlignment());
>         int ScalarEltCost =
> -          TTI->getMemoryOpCost(Instruction::Store, ScalarTy, Alignment, 0, VL0);
> +          TTI->getMemoryOpCost(Instruction::Store, ScalarTy, alignment, 0, VL0);
>         if (NeedToShuffleReuses) {
>           ReuseShuffleCost -= (ReuseShuffleNumbers - VL.size()) * ScalarEltCost;
>         }
>         int ScalarStCost = VecTy->getNumElements() * ScalarEltCost;
> -      int VecStCost = TTI->getMemoryOpCost(Instruction::Store,
> -                                           VecTy, Alignment, 0, VL0);
> -      if (IsReorder) {
> -        // TODO: Merge this shuffle with the ReuseShuffleCost.
> -        VecStCost += TTI->getShuffleCost(
> -            TargetTransformInfo::SK_PermuteSingleSrc, VecTy);
> -      }
> +      int VecStCost =
> +          TTI->getMemoryOpCost(Instruction::Store, VecTy, alignment, 0, VL0);
>         return ReuseShuffleCost + VecStCost - ScalarStCost;
>       }
>       case Instruction::Call: {
> @@ -4109,22 +4051,13 @@ Value *BoUpSLP::vectorizeTree(TreeEntry *E) {
>         return V;
>       }
>       case Instruction::Store: {
> -      bool IsReorder = !E->ReorderIndices.empty();
> -      auto *SI = cast<StoreInst>(
> -          IsReorder ? E->Scalars[E->ReorderIndices.front()] : VL0);
> +      StoreInst *SI = cast<StoreInst>(VL0);
>         unsigned Alignment = SI->getAlignment();
>         unsigned AS = SI->getPointerAddressSpace();
>   
>         setInsertPointAfterBundle(E);
>   
>         Value *VecValue = vectorizeTree(E->getOperand(0));
> -      if (IsReorder) {
> -        OrdersType Mask;
> -        inversePermutation(E->ReorderIndices, Mask);
> -        VecValue = Builder.CreateShuffleVector(
> -            VecValue, UndefValue::get(VecValue->getType()), E->ReorderIndices,
> -            "reorder_shuffle");
> -      }
>         Value *ScalarPtr = SI->getPointerOperand();
>         Value *VecPtr = Builder.CreateBitCast(ScalarPtr, VecTy->getPointerTo(AS));
>         StoreInst *ST = Builder.CreateStore(VecValue, VecPtr);
> @@ -5414,14 +5347,6 @@ bool SLPVectorizerPass::vectorizeStoreChain(ArrayRef<Value *> Chain, BoUpSLP &R,
>                       << "\n");
>   
>     R.buildTree(Chain);
> -  Optional<ArrayRef<unsigned>> Order = R.bestOrder();
> -  if (Order) {
> -    // TODO: reorder tree nodes without tree rebuilding.
> -    SmallVector<Value *, 4> ReorderedOps(Chain.rbegin(), Chain.rend());
> -    llvm::transform(*Order, ReorderedOps.begin(),
> -                    [Chain](const unsigned Idx) { return Chain[Idx]; });
> -    R.buildTree(ReorderedOps);
> -  }
>     if (R.isTreeTinyAndNotFullyVectorizable())
>       return false;
>   
>
> diff  --git a/llvm/test/Transforms/SLPVectorizer/X86/store-jumbled.ll b/llvm/test/Transforms/SLPVectorizer/X86/store-jumbled.ll
> index e99864205bfc..2255a12342f4 100644
> --- a/llvm/test/Transforms/SLPVectorizer/X86/store-jumbled.ll
> +++ b/llvm/test/Transforms/SLPVectorizer/X86/store-jumbled.ll
> @@ -11,20 +11,21 @@ define i32 @jumbled-load(i32* noalias nocapture %in, i32* noalias nocapture %inn
>   ; CHECK-NEXT:    [[GEP_3:%.*]] = getelementptr inbounds i32, i32* [[IN_ADDR]], i64 3
>   ; CHECK-NEXT:    [[TMP1:%.*]] = bitcast i32* [[IN_ADDR]] to <4 x i32>*
>   ; CHECK-NEXT:    [[TMP2:%.*]] = load <4 x i32>, <4 x i32>* [[TMP1]], align 4
> +; CHECK-NEXT:    [[REORDER_SHUFFLE:%.*]] = shufflevector <4 x i32> [[TMP2]], <4 x i32> undef, <4 x i32> <i32 1, i32 3, i32 0, i32 2>
>   ; CHECK-NEXT:    [[INN_ADDR:%.*]] = getelementptr inbounds i32, i32* [[INN:%.*]], i64 0
>   ; CHECK-NEXT:    [[GEP_4:%.*]] = getelementptr inbounds i32, i32* [[INN_ADDR]], i64 1
>   ; CHECK-NEXT:    [[GEP_5:%.*]] = getelementptr inbounds i32, i32* [[INN_ADDR]], i64 2
>   ; CHECK-NEXT:    [[GEP_6:%.*]] = getelementptr inbounds i32, i32* [[INN_ADDR]], i64 3
>   ; CHECK-NEXT:    [[TMP3:%.*]] = bitcast i32* [[INN_ADDR]] to <4 x i32>*
>   ; CHECK-NEXT:    [[TMP4:%.*]] = load <4 x i32>, <4 x i32>* [[TMP3]], align 4
> -; CHECK-NEXT:    [[TMP5:%.*]] = mul <4 x i32> [[TMP2]], [[TMP4]]
> +; CHECK-NEXT:    [[REORDER_SHUFFLE1:%.*]] = shufflevector <4 x i32> [[TMP4]], <4 x i32> undef, <4 x i32> <i32 1, i32 3, i32 0, i32 2>
> +; CHECK-NEXT:    [[TMP5:%.*]] = mul <4 x i32> [[REORDER_SHUFFLE]], [[REORDER_SHUFFLE1]]
>   ; CHECK-NEXT:    [[GEP_7:%.*]] = getelementptr inbounds i32, i32* [[OUT:%.*]], i64 0
>   ; CHECK-NEXT:    [[GEP_8:%.*]] = getelementptr inbounds i32, i32* [[OUT]], i64 1
>   ; CHECK-NEXT:    [[GEP_9:%.*]] = getelementptr inbounds i32, i32* [[OUT]], i64 2
>   ; CHECK-NEXT:    [[GEP_10:%.*]] = getelementptr inbounds i32, i32* [[OUT]], i64 3
> -; CHECK-NEXT:    [[REORDER_SHUFFLE:%.*]] = shufflevector <4 x i32> [[TMP5]], <4 x i32> undef, <4 x i32> <i32 1, i32 3, i32 0, i32 2>
>   ; CHECK-NEXT:    [[TMP6:%.*]] = bitcast i32* [[GEP_7]] to <4 x i32>*
> -; CHECK-NEXT:    store <4 x i32> [[REORDER_SHUFFLE]], <4 x i32>* [[TMP6]], align 4
> +; CHECK-NEXT:    store <4 x i32> [[TMP5]], <4 x i32>* [[TMP6]], align 4
>   ; CHECK-NEXT:    ret i32 undef
>   ;
>     %in.addr = getelementptr inbounds i32, i32* %in, i64 0
>
> diff  --git a/llvm/test/Transforms/SLPVectorizer/X86/stores_vectorize.ll b/llvm/test/Transforms/SLPVectorizer/X86/stores_vectorize.ll
> index fd697f1a7ca5..425f3e634167 100644
> --- a/llvm/test/Transforms/SLPVectorizer/X86/stores_vectorize.ll
> +++ b/llvm/test/Transforms/SLPVectorizer/X86/stores_vectorize.ll
> @@ -92,14 +92,15 @@ define void @store_reverse(i64* %p3) {
>   ; CHECK-NEXT:    [[ARRAYIDX11:%.*]] = getelementptr inbounds i64, i64* [[P3]], i64 3
>   ; CHECK-NEXT:    [[TMP0:%.*]] = bitcast i64* [[P3]] to <4 x i64>*
>   ; CHECK-NEXT:    [[TMP1:%.*]] = load <4 x i64>, <4 x i64>* [[TMP0]], align 8
> +; CHECK-NEXT:    [[REORDER_SHUFFLE:%.*]] = shufflevector <4 x i64> [[TMP1]], <4 x i64> undef, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
>   ; CHECK-NEXT:    [[ARRAYIDX12:%.*]] = getelementptr inbounds i64, i64* [[P3]], i64 11
>   ; CHECK-NEXT:    [[TMP2:%.*]] = bitcast i64* [[ARRAYIDX1]] to <4 x i64>*
>   ; CHECK-NEXT:    [[TMP3:%.*]] = load <4 x i64>, <4 x i64>* [[TMP2]], align 8
> -; CHECK-NEXT:    [[TMP4:%.*]] = shl <4 x i64> [[TMP1]], [[TMP3]]
> +; CHECK-NEXT:    [[REORDER_SHUFFLE1:%.*]] = shufflevector <4 x i64> [[TMP3]], <4 x i64> undef, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
> +; CHECK-NEXT:    [[TMP4:%.*]] = shl <4 x i64> [[REORDER_SHUFFLE]], [[REORDER_SHUFFLE1]]
>   ; CHECK-NEXT:    [[ARRAYIDX14:%.*]] = getelementptr inbounds i64, i64* [[P3]], i64 4
> -; CHECK-NEXT:    [[TMP5:%.*]] = shufflevector <4 x i64> [[TMP4]], <4 x i64> undef, <4 x i32> <i32 3, i32 2, i32 1, i32 0>
> -; CHECK-NEXT:    [[TMP6:%.*]] = bitcast i64* [[ARRAYIDX14]] to <4 x i64>*
> -; CHECK-NEXT:    store <4 x i64> [[TMP5]], <4 x i64>* [[TMP6]], align 8
> +; CHECK-NEXT:    [[TMP5:%.*]] = bitcast i64* [[ARRAYIDX14]] to <4 x i64>*
> +; CHECK-NEXT:    store <4 x i64> [[TMP4]], <4 x i64>* [[TMP5]], align 8
>   ; CHECK-NEXT:    ret void
>   ;
>   entry:
>
>
>          
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at lists.llvm.org
> https://lists.llvm.org/cgi-bin/mailman/listinfo/llvm-commits


More information about the llvm-commits mailing list