[llvm] 4979c90 - [LV] Account for tripcount when calculation vectorization profitability

David Green via llvm-commits llvm-commits at lists.llvm.org
Thu May 6 04:36:58 PDT 2021


Author: David Green
Date: 2021-05-06T12:36:46+01:00
New Revision: 4979c90458628c9463815d81c637f8787f72fff0

URL: https://github.com/llvm/llvm-project/commit/4979c90458628c9463815d81c637f8787f72fff0
DIFF: https://github.com/llvm/llvm-project/commit/4979c90458628c9463815d81c637f8787f72fff0.diff

LOG: [LV] Account for tripcount when calculation vectorization profitability

The loop vectorizer will currently assume a large trip count when
calculating which of several vectorization factors are more profitable.
That is often not a terrible assumption to make as small trip count
loops will usually have been fully unrolled. There are cases however
where we will try to vectorize them, and especially when folding the
tail by masking can incorrectly choose to vectorize loops that are not
beneficial, due to the folded tail rounding the iteration count up for
the vectorized loop.

The motivating example here has a trip count of 5, so either performs 5
scalar iterations or 2 vector iterations (with VF=4). At a high enough
trip count the vectorization becomes profitable, but the rounding up to
2 vector iterations vs only 5 scalar makes it unprofitable.

This adds an alternative cost calculation when we know the max trip
count and are folding tail by masking, rounding the iteration count up
to the correct number for the vector width. We still do not account for
anything like setup cost or the mixture of vector and scalar loops, but
this is at least an improvement in a few cases that we have had
reported.

Differential Revision: https://reviews.llvm.org/D101726

Added: 
    llvm/test/Transforms/LoopVectorize/ARM/mve-known-trip-count.ll

Modified: 
    llvm/lib/Transforms/Vectorize/LoopVectorize.cpp

Removed: 
    


################################################################################
diff  --git a/llvm/lib/Transforms/Vectorize/LoopVectorize.cpp b/llvm/lib/Transforms/Vectorize/LoopVectorize.cpp
index e53d111a9bad1..bba8e6f57a219 100644
--- a/llvm/lib/Transforms/Vectorize/LoopVectorize.cpp
+++ b/llvm/lib/Transforms/Vectorize/LoopVectorize.cpp
@@ -5969,6 +5969,22 @@ bool LoopVectorizationCostModel::isMoreProfitable(
   InstructionCost::CostType CostA = *A.Cost.getValue();
   InstructionCost::CostType CostB = *B.Cost.getValue();
 
+  unsigned MaxTripCount = PSE.getSE()->getSmallConstantMaxTripCount(TheLoop);
+
+  if (!A.Width.isScalable() && !B.Width.isScalable() && FoldTailByMasking &&
+      MaxTripCount) {
+    // If we are folding the tail and the trip count is a known (possibly small)
+    // constant, the trip count will be rounded up to an integer number of
+    // iterations. The total cost will be PerIterationCost*ceil(TripCount/VF),
+    // which we compare directly. When not folding the tail, the total cost will
+    // be PerIterationCost*floor(TC/VF) + Scalar remainder cost, and so is
+    // approximated with the per-lane cost below instead of using the tripcount
+    // as here.
+    int64_t RTCostA = CostA * divideCeil(MaxTripCount, A.Width.getFixedValue());
+    int64_t RTCostB = CostB * divideCeil(MaxTripCount, B.Width.getFixedValue());
+    return RTCostA < RTCostB;
+  }
+
   // To avoid the need for FP division:
   //      (CostA / A.Width) < (CostB / B.Width)
   // <=>  (CostA * B.Width) < (CostB * A.Width)

diff  --git a/llvm/test/Transforms/LoopVectorize/ARM/mve-known-trip-count.ll b/llvm/test/Transforms/LoopVectorize/ARM/mve-known-trip-count.ll
new file mode 100644
index 0000000000000..28dba6af29c67
--- /dev/null
+++ b/llvm/test/Transforms/LoopVectorize/ARM/mve-known-trip-count.ll
@@ -0,0 +1,389 @@
+; RUN: opt -loop-vectorize -debug-only=loop-vectorize -disable-output < %s 2>&1 | FileCheck %s
+; REQUIRES: asserts
+
+target datalayout = "e-m:e-p:32:32-Fi8-i64:64-v128:64:128-a:0:32-n32-S64"
+target triple = "thumbv8.1m.main-arm-none-eabi"
+
+; Trip count of 5 - shouldn't be vectorized.
+; CHECK-LABEL: tripcount5
+; CHECK: LV: Selecting VF: 1
+define void @tripcount5(i16* nocapture readonly %in, i32* nocapture %out, i16* nocapture readonly %consts, i32 %n) #0 {
+entry:
+  %arrayidx20 = getelementptr inbounds i32, i32* %out, i32 1
+  %arrayidx38 = getelementptr inbounds i32, i32* %out, i32 2
+  %arrayidx56 = getelementptr inbounds i32, i32* %out, i32 3
+  %arrayidx74 = getelementptr inbounds i32, i32* %out, i32 4
+  %arrayidx92 = getelementptr inbounds i32, i32* %out, i32 5
+  %arrayidx110 = getelementptr inbounds i32, i32* %out, i32 6
+  %arrayidx128 = getelementptr inbounds i32, i32* %out, i32 7
+  %out.promoted = load i32, i32* %out, align 4
+  %arrayidx20.promoted = load i32, i32* %arrayidx20, align 4
+  %arrayidx38.promoted = load i32, i32* %arrayidx38, align 4
+  %arrayidx56.promoted = load i32, i32* %arrayidx56, align 4
+  %arrayidx74.promoted = load i32, i32* %arrayidx74, align 4
+  %arrayidx92.promoted = load i32, i32* %arrayidx92, align 4
+  %arrayidx110.promoted = load i32, i32* %arrayidx110, align 4
+  %arrayidx128.promoted = load i32, i32* %arrayidx128, align 4
+  br label %for.body
+
+for.cond.cleanup:                                 ; preds = %for.body
+  store i32 %add12, i32* %out, align 4
+  store i32 %add30, i32* %arrayidx20, align 4
+  store i32 %add48, i32* %arrayidx38, align 4
+  store i32 %add66, i32* %arrayidx56, align 4
+  store i32 %add84, i32* %arrayidx74, align 4
+  store i32 %add102, i32* %arrayidx92, align 4
+  store i32 %add120, i32* %arrayidx110, align 4
+  store i32 %add138, i32* %arrayidx128, align 4
+  ret void
+
+for.body:                                         ; preds = %entry, %for.body
+  %hop.0236 = phi i32 [ 0, %entry ], [ %add139, %for.body ]
+  %add12220235 = phi i32 [ %out.promoted, %entry ], [ %add12, %for.body ]
+  %add30221234 = phi i32 [ %arrayidx20.promoted, %entry ], [ %add30, %for.body ]
+  %add48222233 = phi i32 [ %arrayidx38.promoted, %entry ], [ %add48, %for.body ]
+  %add66223232 = phi i32 [ %arrayidx56.promoted, %entry ], [ %add66, %for.body ]
+  %add84224231 = phi i32 [ %arrayidx74.promoted, %entry ], [ %add84, %for.body ]
+  %add102225230 = phi i32 [ %arrayidx92.promoted, %entry ], [ %add102, %for.body ]
+  %add120226229 = phi i32 [ %arrayidx110.promoted, %entry ], [ %add120, %for.body ]
+  %add138227228 = phi i32 [ %arrayidx128.promoted, %entry ], [ %add138, %for.body ]
+  %arrayidx = getelementptr inbounds i16, i16* %in, i32 %hop.0236
+  %0 = load i16, i16* %arrayidx, align 2
+  %conv = sext i16 %0 to i32
+  %arrayidx1 = getelementptr inbounds i16, i16* %consts, i32 %hop.0236
+  %1 = load i16, i16* %arrayidx1, align 2
+  %conv2 = sext i16 %1 to i32
+  %mul = mul nsw i32 %conv2, %conv
+  %add = add nsw i32 %mul, %add12220235
+  %add4 = or i32 %hop.0236, 1
+  %arrayidx5 = getelementptr inbounds i16, i16* %in, i32 %add4
+  %2 = load i16, i16* %arrayidx5, align 2
+  %conv6 = sext i16 %2 to i32
+  %arrayidx8 = getelementptr inbounds i16, i16* %consts, i32 %add4
+  %3 = load i16, i16* %arrayidx8, align 2
+  %conv9 = sext i16 %3 to i32
+  %mul10 = mul nsw i32 %conv9, %conv6
+  %add12 = add nsw i32 %mul10, %add
+  %add13 = or i32 %hop.0236, 2
+  %arrayidx14 = getelementptr inbounds i16, i16* %in, i32 %add13
+  %4 = load i16, i16* %arrayidx14, align 2
+  %conv15 = sext i16 %4 to i32
+  %arrayidx17 = getelementptr inbounds i16, i16* %consts, i32 %add13
+  %5 = load i16, i16* %arrayidx17, align 2
+  %conv18 = sext i16 %5 to i32
+  %mul19 = mul nsw i32 %conv18, %conv15
+  %add21 = add nsw i32 %mul19, %add30221234
+  %add22 = or i32 %hop.0236, 3
+  %arrayidx23 = getelementptr inbounds i16, i16* %in, i32 %add22
+  %6 = load i16, i16* %arrayidx23, align 2
+  %conv24 = sext i16 %6 to i32
+  %arrayidx26 = getelementptr inbounds i16, i16* %consts, i32 %add22
+  %7 = load i16, i16* %arrayidx26, align 2
+  %conv27 = sext i16 %7 to i32
+  %mul28 = mul nsw i32 %conv27, %conv24
+  %add30 = add nsw i32 %mul28, %add21
+  %add31 = or i32 %hop.0236, 4
+  %arrayidx32 = getelementptr inbounds i16, i16* %in, i32 %add31
+  %8 = load i16, i16* %arrayidx32, align 2
+  %conv33 = sext i16 %8 to i32
+  %arrayidx35 = getelementptr inbounds i16, i16* %consts, i32 %add31
+  %9 = load i16, i16* %arrayidx35, align 2
+  %conv36 = sext i16 %9 to i32
+  %mul37 = mul nsw i32 %conv36, %conv33
+  %add39 = add nsw i32 %mul37, %add48222233
+  %add40 = or i32 %hop.0236, 5
+  %arrayidx41 = getelementptr inbounds i16, i16* %in, i32 %add40
+  %10 = load i16, i16* %arrayidx41, align 2
+  %conv42 = sext i16 %10 to i32
+  %arrayidx44 = getelementptr inbounds i16, i16* %consts, i32 %add40
+  %11 = load i16, i16* %arrayidx44, align 2
+  %conv45 = sext i16 %11 to i32
+  %mul46 = mul nsw i32 %conv45, %conv42
+  %add48 = add nsw i32 %mul46, %add39
+  %add49 = or i32 %hop.0236, 6
+  %arrayidx50 = getelementptr inbounds i16, i16* %in, i32 %add49
+  %12 = load i16, i16* %arrayidx50, align 2
+  %conv51 = sext i16 %12 to i32
+  %arrayidx53 = getelementptr inbounds i16, i16* %consts, i32 %add49
+  %13 = load i16, i16* %arrayidx53, align 2
+  %conv54 = sext i16 %13 to i32
+  %mul55 = mul nsw i32 %conv54, %conv51
+  %add57 = add nsw i32 %mul55, %add66223232
+  %add58 = or i32 %hop.0236, 7
+  %arrayidx59 = getelementptr inbounds i16, i16* %in, i32 %add58
+  %14 = load i16, i16* %arrayidx59, align 2
+  %conv60 = sext i16 %14 to i32
+  %arrayidx62 = getelementptr inbounds i16, i16* %consts, i32 %add58
+  %15 = load i16, i16* %arrayidx62, align 2
+  %conv63 = sext i16 %15 to i32
+  %mul64 = mul nsw i32 %conv63, %conv60
+  %add66 = add nsw i32 %mul64, %add57
+  %add67 = or i32 %hop.0236, 8
+  %arrayidx68 = getelementptr inbounds i16, i16* %in, i32 %add67
+  %16 = load i16, i16* %arrayidx68, align 2
+  %conv69 = sext i16 %16 to i32
+  %arrayidx71 = getelementptr inbounds i16, i16* %consts, i32 %add67
+  %17 = load i16, i16* %arrayidx71, align 2
+  %conv72 = sext i16 %17 to i32
+  %mul73 = mul nsw i32 %conv72, %conv69
+  %add75 = add nsw i32 %mul73, %add84224231
+  %add76 = or i32 %hop.0236, 9
+  %arrayidx77 = getelementptr inbounds i16, i16* %in, i32 %add76
+  %18 = load i16, i16* %arrayidx77, align 2
+  %conv78 = sext i16 %18 to i32
+  %arrayidx80 = getelementptr inbounds i16, i16* %consts, i32 %add76
+  %19 = load i16, i16* %arrayidx80, align 2
+  %conv81 = sext i16 %19 to i32
+  %mul82 = mul nsw i32 %conv81, %conv78
+  %add84 = add nsw i32 %mul82, %add75
+  %add85 = or i32 %hop.0236, 10
+  %arrayidx86 = getelementptr inbounds i16, i16* %in, i32 %add85
+  %20 = load i16, i16* %arrayidx86, align 2
+  %conv87 = sext i16 %20 to i32
+  %arrayidx89 = getelementptr inbounds i16, i16* %consts, i32 %add85
+  %21 = load i16, i16* %arrayidx89, align 2
+  %conv90 = sext i16 %21 to i32
+  %mul91 = mul nsw i32 %conv90, %conv87
+  %add93 = add nsw i32 %mul91, %add102225230
+  %add94 = or i32 %hop.0236, 11
+  %arrayidx95 = getelementptr inbounds i16, i16* %in, i32 %add94
+  %22 = load i16, i16* %arrayidx95, align 2
+  %conv96 = sext i16 %22 to i32
+  %arrayidx98 = getelementptr inbounds i16, i16* %consts, i32 %add94
+  %23 = load i16, i16* %arrayidx98, align 2
+  %conv99 = sext i16 %23 to i32
+  %mul100 = mul nsw i32 %conv99, %conv96
+  %add102 = add nsw i32 %mul100, %add93
+  %add103 = or i32 %hop.0236, 12
+  %arrayidx104 = getelementptr inbounds i16, i16* %in, i32 %add103
+  %24 = load i16, i16* %arrayidx104, align 2
+  %conv105 = sext i16 %24 to i32
+  %arrayidx107 = getelementptr inbounds i16, i16* %consts, i32 %add103
+  %25 = load i16, i16* %arrayidx107, align 2
+  %conv108 = sext i16 %25 to i32
+  %mul109 = mul nsw i32 %conv108, %conv105
+  %add111 = add nsw i32 %mul109, %add120226229
+  %add112 = or i32 %hop.0236, 13
+  %arrayidx113 = getelementptr inbounds i16, i16* %in, i32 %add112
+  %26 = load i16, i16* %arrayidx113, align 2
+  %conv114 = sext i16 %26 to i32
+  %arrayidx116 = getelementptr inbounds i16, i16* %consts, i32 %add112
+  %27 = load i16, i16* %arrayidx116, align 2
+  %conv117 = sext i16 %27 to i32
+  %mul118 = mul nsw i32 %conv117, %conv114
+  %add120 = add nsw i32 %mul118, %add111
+  %add121 = or i32 %hop.0236, 14
+  %arrayidx122 = getelementptr inbounds i16, i16* %in, i32 %add121
+  %28 = load i16, i16* %arrayidx122, align 2
+  %conv123 = sext i16 %28 to i32
+  %arrayidx125 = getelementptr inbounds i16, i16* %consts, i32 %add121
+  %29 = load i16, i16* %arrayidx125, align 2
+  %conv126 = sext i16 %29 to i32
+  %mul127 = mul nsw i32 %conv126, %conv123
+  %add129 = add nsw i32 %mul127, %add138227228
+  %add130 = or i32 %hop.0236, 15
+  %arrayidx131 = getelementptr inbounds i16, i16* %in, i32 %add130
+  %30 = load i16, i16* %arrayidx131, align 2
+  %conv132 = sext i16 %30 to i32
+  %arrayidx134 = getelementptr inbounds i16, i16* %consts, i32 %add130
+  %31 = load i16, i16* %arrayidx134, align 2
+  %conv135 = sext i16 %31 to i32
+  %mul136 = mul nsw i32 %conv135, %conv132
+  %add138 = add nsw i32 %mul136, %add129
+  %add139 = add nuw nsw i32 %hop.0236, 16
+  %cmp = icmp ult i32 %hop.0236, 64
+  br i1 %cmp, label %for.body, label %for.cond.cleanup
+}
+
+; Trip count of 8 - does get vectorized
+; CHECK-LABEL: tripcount8
+; CHECK: LV: Selecting VF: 4
+define void @tripcount8(i16* nocapture readonly %in, i32* nocapture %out, i16* nocapture readonly %consts, i32 %n) #0 {
+entry:
+  %arrayidx20 = getelementptr inbounds i32, i32* %out, i32 1
+  %arrayidx38 = getelementptr inbounds i32, i32* %out, i32 2
+  %arrayidx56 = getelementptr inbounds i32, i32* %out, i32 3
+  %arrayidx74 = getelementptr inbounds i32, i32* %out, i32 4
+  %arrayidx92 = getelementptr inbounds i32, i32* %out, i32 5
+  %arrayidx110 = getelementptr inbounds i32, i32* %out, i32 6
+  %arrayidx128 = getelementptr inbounds i32, i32* %out, i32 7
+  %out.promoted = load i32, i32* %out, align 4
+  %arrayidx20.promoted = load i32, i32* %arrayidx20, align 4
+  %arrayidx38.promoted = load i32, i32* %arrayidx38, align 4
+  %arrayidx56.promoted = load i32, i32* %arrayidx56, align 4
+  %arrayidx74.promoted = load i32, i32* %arrayidx74, align 4
+  %arrayidx92.promoted = load i32, i32* %arrayidx92, align 4
+  %arrayidx110.promoted = load i32, i32* %arrayidx110, align 4
+  %arrayidx128.promoted = load i32, i32* %arrayidx128, align 4
+  br label %for.body
+
+for.cond.cleanup:                                 ; preds = %for.body
+  store i32 %add12, i32* %out, align 4
+  store i32 %add30, i32* %arrayidx20, align 4
+  store i32 %add48, i32* %arrayidx38, align 4
+  store i32 %add66, i32* %arrayidx56, align 4
+  store i32 %add84, i32* %arrayidx74, align 4
+  store i32 %add102, i32* %arrayidx92, align 4
+  store i32 %add120, i32* %arrayidx110, align 4
+  store i32 %add138, i32* %arrayidx128, align 4
+  ret void
+
+for.body:                                         ; preds = %entry, %for.body
+  %hop.0236 = phi i32 [ 0, %entry ], [ %add139, %for.body ]
+  %add12220235 = phi i32 [ %out.promoted, %entry ], [ %add12, %for.body ]
+  %add30221234 = phi i32 [ %arrayidx20.promoted, %entry ], [ %add30, %for.body ]
+  %add48222233 = phi i32 [ %arrayidx38.promoted, %entry ], [ %add48, %for.body ]
+  %add66223232 = phi i32 [ %arrayidx56.promoted, %entry ], [ %add66, %for.body ]
+  %add84224231 = phi i32 [ %arrayidx74.promoted, %entry ], [ %add84, %for.body ]
+  %add102225230 = phi i32 [ %arrayidx92.promoted, %entry ], [ %add102, %for.body ]
+  %add120226229 = phi i32 [ %arrayidx110.promoted, %entry ], [ %add120, %for.body ]
+  %add138227228 = phi i32 [ %arrayidx128.promoted, %entry ], [ %add138, %for.body ]
+  %arrayidx = getelementptr inbounds i16, i16* %in, i32 %hop.0236
+  %0 = load i16, i16* %arrayidx, align 2
+  %conv = sext i16 %0 to i32
+  %arrayidx1 = getelementptr inbounds i16, i16* %consts, i32 %hop.0236
+  %1 = load i16, i16* %arrayidx1, align 2
+  %conv2 = sext i16 %1 to i32
+  %mul = mul nsw i32 %conv2, %conv
+  %add = add nsw i32 %mul, %add12220235
+  %add4 = or i32 %hop.0236, 1
+  %arrayidx5 = getelementptr inbounds i16, i16* %in, i32 %add4
+  %2 = load i16, i16* %arrayidx5, align 2
+  %conv6 = sext i16 %2 to i32
+  %arrayidx8 = getelementptr inbounds i16, i16* %consts, i32 %add4
+  %3 = load i16, i16* %arrayidx8, align 2
+  %conv9 = sext i16 %3 to i32
+  %mul10 = mul nsw i32 %conv9, %conv6
+  %add12 = add nsw i32 %mul10, %add
+  %add13 = or i32 %hop.0236, 2
+  %arrayidx14 = getelementptr inbounds i16, i16* %in, i32 %add13
+  %4 = load i16, i16* %arrayidx14, align 2
+  %conv15 = sext i16 %4 to i32
+  %arrayidx17 = getelementptr inbounds i16, i16* %consts, i32 %add13
+  %5 = load i16, i16* %arrayidx17, align 2
+  %conv18 = sext i16 %5 to i32
+  %mul19 = mul nsw i32 %conv18, %conv15
+  %add21 = add nsw i32 %mul19, %add30221234
+  %add22 = or i32 %hop.0236, 3
+  %arrayidx23 = getelementptr inbounds i16, i16* %in, i32 %add22
+  %6 = load i16, i16* %arrayidx23, align 2
+  %conv24 = sext i16 %6 to i32
+  %arrayidx26 = getelementptr inbounds i16, i16* %consts, i32 %add22
+  %7 = load i16, i16* %arrayidx26, align 2
+  %conv27 = sext i16 %7 to i32
+  %mul28 = mul nsw i32 %conv27, %conv24
+  %add30 = add nsw i32 %mul28, %add21
+  %add31 = or i32 %hop.0236, 4
+  %arrayidx32 = getelementptr inbounds i16, i16* %in, i32 %add31
+  %8 = load i16, i16* %arrayidx32, align 2
+  %conv33 = sext i16 %8 to i32
+  %arrayidx35 = getelementptr inbounds i16, i16* %consts, i32 %add31
+  %9 = load i16, i16* %arrayidx35, align 2
+  %conv36 = sext i16 %9 to i32
+  %mul37 = mul nsw i32 %conv36, %conv33
+  %add39 = add nsw i32 %mul37, %add48222233
+  %add40 = or i32 %hop.0236, 5
+  %arrayidx41 = getelementptr inbounds i16, i16* %in, i32 %add40
+  %10 = load i16, i16* %arrayidx41, align 2
+  %conv42 = sext i16 %10 to i32
+  %arrayidx44 = getelementptr inbounds i16, i16* %consts, i32 %add40
+  %11 = load i16, i16* %arrayidx44, align 2
+  %conv45 = sext i16 %11 to i32
+  %mul46 = mul nsw i32 %conv45, %conv42
+  %add48 = add nsw i32 %mul46, %add39
+  %add49 = or i32 %hop.0236, 6
+  %arrayidx50 = getelementptr inbounds i16, i16* %in, i32 %add49
+  %12 = load i16, i16* %arrayidx50, align 2
+  %conv51 = sext i16 %12 to i32
+  %arrayidx53 = getelementptr inbounds i16, i16* %consts, i32 %add49
+  %13 = load i16, i16* %arrayidx53, align 2
+  %conv54 = sext i16 %13 to i32
+  %mul55 = mul nsw i32 %conv54, %conv51
+  %add57 = add nsw i32 %mul55, %add66223232
+  %add58 = or i32 %hop.0236, 7
+  %arrayidx59 = getelementptr inbounds i16, i16* %in, i32 %add58
+  %14 = load i16, i16* %arrayidx59, align 2
+  %conv60 = sext i16 %14 to i32
+  %arrayidx62 = getelementptr inbounds i16, i16* %consts, i32 %add58
+  %15 = load i16, i16* %arrayidx62, align 2
+  %conv63 = sext i16 %15 to i32
+  %mul64 = mul nsw i32 %conv63, %conv60
+  %add66 = add nsw i32 %mul64, %add57
+  %add67 = or i32 %hop.0236, 8
+  %arrayidx68 = getelementptr inbounds i16, i16* %in, i32 %add67
+  %16 = load i16, i16* %arrayidx68, align 2
+  %conv69 = sext i16 %16 to i32
+  %arrayidx71 = getelementptr inbounds i16, i16* %consts, i32 %add67
+  %17 = load i16, i16* %arrayidx71, align 2
+  %conv72 = sext i16 %17 to i32
+  %mul73 = mul nsw i32 %conv72, %conv69
+  %add75 = add nsw i32 %mul73, %add84224231
+  %add76 = or i32 %hop.0236, 9
+  %arrayidx77 = getelementptr inbounds i16, i16* %in, i32 %add76
+  %18 = load i16, i16* %arrayidx77, align 2
+  %conv78 = sext i16 %18 to i32
+  %arrayidx80 = getelementptr inbounds i16, i16* %consts, i32 %add76
+  %19 = load i16, i16* %arrayidx80, align 2
+  %conv81 = sext i16 %19 to i32
+  %mul82 = mul nsw i32 %conv81, %conv78
+  %add84 = add nsw i32 %mul82, %add75
+  %add85 = or i32 %hop.0236, 10
+  %arrayidx86 = getelementptr inbounds i16, i16* %in, i32 %add85
+  %20 = load i16, i16* %arrayidx86, align 2
+  %conv87 = sext i16 %20 to i32
+  %arrayidx89 = getelementptr inbounds i16, i16* %consts, i32 %add85
+  %21 = load i16, i16* %arrayidx89, align 2
+  %conv90 = sext i16 %21 to i32
+  %mul91 = mul nsw i32 %conv90, %conv87
+  %add93 = add nsw i32 %mul91, %add102225230
+  %add94 = or i32 %hop.0236, 11
+  %arrayidx95 = getelementptr inbounds i16, i16* %in, i32 %add94
+  %22 = load i16, i16* %arrayidx95, align 2
+  %conv96 = sext i16 %22 to i32
+  %arrayidx98 = getelementptr inbounds i16, i16* %consts, i32 %add94
+  %23 = load i16, i16* %arrayidx98, align 2
+  %conv99 = sext i16 %23 to i32
+  %mul100 = mul nsw i32 %conv99, %conv96
+  %add102 = add nsw i32 %mul100, %add93
+  %add103 = or i32 %hop.0236, 12
+  %arrayidx104 = getelementptr inbounds i16, i16* %in, i32 %add103
+  %24 = load i16, i16* %arrayidx104, align 2
+  %conv105 = sext i16 %24 to i32
+  %arrayidx107 = getelementptr inbounds i16, i16* %consts, i32 %add103
+  %25 = load i16, i16* %arrayidx107, align 2
+  %conv108 = sext i16 %25 to i32
+  %mul109 = mul nsw i32 %conv108, %conv105
+  %add111 = add nsw i32 %mul109, %add120226229
+  %add112 = or i32 %hop.0236, 13
+  %arrayidx113 = getelementptr inbounds i16, i16* %in, i32 %add112
+  %26 = load i16, i16* %arrayidx113, align 2
+  %conv114 = sext i16 %26 to i32
+  %arrayidx116 = getelementptr inbounds i16, i16* %consts, i32 %add112
+  %27 = load i16, i16* %arrayidx116, align 2
+  %conv117 = sext i16 %27 to i32
+  %mul118 = mul nsw i32 %conv117, %conv114
+  %add120 = add nsw i32 %mul118, %add111
+  %add121 = or i32 %hop.0236, 14
+  %arrayidx122 = getelementptr inbounds i16, i16* %in, i32 %add121
+  %28 = load i16, i16* %arrayidx122, align 2
+  %conv123 = sext i16 %28 to i32
+  %arrayidx125 = getelementptr inbounds i16, i16* %consts, i32 %add121
+  %29 = load i16, i16* %arrayidx125, align 2
+  %conv126 = sext i16 %29 to i32
+  %mul127 = mul nsw i32 %conv126, %conv123
+  %add129 = add nsw i32 %mul127, %add138227228
+  %add130 = or i32 %hop.0236, 15
+  %arrayidx131 = getelementptr inbounds i16, i16* %in, i32 %add130
+  %30 = load i16, i16* %arrayidx131, align 2
+  %conv132 = sext i16 %30 to i32
+  %arrayidx134 = getelementptr inbounds i16, i16* %consts, i32 %add130
+  %31 = load i16, i16* %arrayidx134, align 2
+  %conv135 = sext i16 %31 to i32
+  %mul136 = mul nsw i32 %conv135, %conv132
+  %add138 = add nsw i32 %mul136, %add129
+  %add139 = add nuw nsw i32 %hop.0236, 16
+  %cmp = icmp ult i32 %hop.0236, 112
+  br i1 %cmp, label %for.body, label %for.cond.cleanup
+}
+
+attributes #0 = { "target-features"="+mve" }
\ No newline at end of file


        


More information about the llvm-commits mailing list