[llvm] 565b98e - [LV] Convert consecutive-ptr-uniforms.ll to use opaque pointers (NFC).

Florian Hahn via llvm-commits llvm-commits at lists.llvm.org
Fri Mar 17 07:07:38 PDT 2023


Author: Florian Hahn
Date: 2023-03-17T14:07:11Z
New Revision: 565b98e793422efbffaa2928f03bc943a1346a3e

URL: https://github.com/llvm/llvm-project/commit/565b98e793422efbffaa2928f03bc943a1346a3e
DIFF: https://github.com/llvm/llvm-project/commit/565b98e793422efbffaa2928f03bc943a1346a3e.diff

LOG: [LV] Convert consecutive-ptr-uniforms.ll to use opaque pointers (NFC).

Added: 
    

Modified: 
    llvm/test/Transforms/LoopVectorize/consecutive-ptr-uniforms.ll

Removed: 
    


################################################################################
diff  --git a/llvm/test/Transforms/LoopVectorize/consecutive-ptr-uniforms.ll b/llvm/test/Transforms/LoopVectorize/consecutive-ptr-uniforms.ll
index 40337e7da7a46..e4129747898b0 100644
--- a/llvm/test/Transforms/LoopVectorize/consecutive-ptr-uniforms.ll
+++ b/llvm/test/Transforms/LoopVectorize/consecutive-ptr-uniforms.ll
@@ -1,6 +1,6 @@
 ; REQUIRES: asserts
-; RUN: opt -opaque-pointers=0 < %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s
-; RUN: opt -opaque-pointers=0 < %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -enable-interleaved-mem-accesses -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s --check-prefix=INTER
+; RUN: opt %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s
+; RUN: opt %s -passes=loop-vectorize,instcombine -force-vector-width=4 -force-vector-interleave=1 -enable-interleaved-mem-accesses -debug-only=loop-vectorize -disable-output -print-after=instcombine 2>&1 | FileCheck %s --check-prefix=INTER
 
 target datalayout = "e-m:e-i64:64-i128:128-n32:64-S128"
 
@@ -11,23 +11,23 @@ target datalayout = "e-m:e-i64:64-i128:128-n32:64-S128"
 ; Check that a forward consecutive pointer is recognized as uniform and remains
 ; uniform after vectorization.
 ;
-; CHECK:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, i32* %a, i64 %i
+; CHECK:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK-NOT:   getelementptr
-; CHECK:       getelementptr inbounds i32, i32* %a, i64 %index
+; CHECK:       getelementptr inbounds i32, ptr %a, i64 %index
 ; CHECK-NOT:   getelementptr
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define i32 @consecutive_ptr_forward(i32* %a, i64 %n) {
+define i32 @consecutive_ptr_forward(ptr %a, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
   %tmp0 = phi i32 [ %tmp3, %for.body ], [ 0, %entry ]
-  %tmp1 = getelementptr inbounds i32, i32* %a, i64 %i
-  %tmp2 = load i32, i32* %tmp1, align 8
+  %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i
+  %tmp2 = load i32, ptr %tmp1, align 8
   %tmp3 = add i32 %tmp0, %tmp2
   %i.next = add nuw nsw i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
@@ -43,25 +43,25 @@ for.end:
 ; Check that a reverse consecutive pointer is recognized as uniform and remains
 ; uniform after vectorization.
 ;
-; CHECK:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, i32* %a, i64 %i
+; CHECK:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK:       %offset.idx = sub i64 %n, %index
 ; CHECK-NOT:   getelementptr
-; CHECK:       %[[G0:.+]] = getelementptr i32, i32* %a, i64 -3
-; CHECK:       getelementptr i32, i32* %[[G0]], i64 %offset.idx
+; CHECK:       %[[G0:.+]] = getelementptr i32, ptr %a, i64 -3
+; CHECK:       getelementptr i32, ptr %[[G0]], i64 %offset.idx
 ; CHECK-NOT:   getelementptr
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define i32 @consecutive_ptr_reverse(i32* %a, i64 %n) {
+define i32 @consecutive_ptr_reverse(ptr %a, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ %n, %entry ]
   %tmp0 = phi i32 [ %tmp3, %for.body ], [ 0, %entry ]
-  %tmp1 = getelementptr inbounds i32, i32* %a, i64 %i
-  %tmp2 = load i32, i32* %tmp1, align 8
+  %tmp1 = getelementptr inbounds i32, ptr %a, i64 %i
+  %tmp2 = load i32, ptr %tmp1, align 8
   %tmp3 = add i32 %tmp0, %tmp2
   %i.next = add nsw i64 %i, -1
   %cond = icmp sgt i64 %i.next, 0
@@ -80,43 +80,43 @@ for.end:
 ; interleaved memory accesses aren't enabled, the pointer should not be
 ; recognized as uniform, and it should not be uniform after vectorization.
 ;
-; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
+; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK:       %[[I1:.+]] = or i64 %index, 1
 ; CHECK:       %[[I2:.+]] = or i64 %index, 2
 ; CHECK:       %[[I3:.+]] = or i64 %index, 3
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %index, i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I1]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I2]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I3]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %index, i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I1]], i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I2]], i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I3]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %index, i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %index, i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 1
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-; INTER:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-; INTER:     LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
+; INTER:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+; INTER:     LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
 ; INTER:     vector.body
 ; INTER:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; INTER-NOT:   getelementptr
-; INTER:       getelementptr inbounds %pair, %pair* %p, i64 %index, i32 0
+; INTER:       getelementptr inbounds %pair, ptr %p, i64 %index, i32 0
 ; INTER-NOT:   getelementptr
 ; INTER:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define i32 @interleaved_access_forward(%pair* %p, i64 %n) {
+define i32 @interleaved_access_forward(ptr %p, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
   %tmp0 = phi i32 [ %tmp6, %for.body ], [ 0, %entry ]
-  %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-  %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
-  %tmp3 = load i32, i32* %tmp1, align 8
-  %tmp4 = load i32, i32* %tmp2, align 8
+  %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+  %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
+  %tmp3 = load i32, ptr %tmp1, align 8
+  %tmp4 = load i32, ptr %tmp2, align 8
   %tmp5 = add i32 %tmp3, %tmp4
   %tmp6 = add i32 %tmp0, %tmp5
   %i.next = add nuw nsw i64 %i, 1
@@ -137,46 +137,46 @@ for.end:
 ; recognized as uniform, and it should not be uniform after vectorization.
 ;
 ; recognized as uniform, and it should not be uniform after vectorization.
-; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
+; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+; CHECK-NOT: LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK:       %offset.idx = sub i64 %n, %index
 ; CHECK:       %[[I1:.+]] = add i64 %offset.idx, -1
 ; CHECK:       %[[I2:.+]] = add i64 %offset.idx, -2
 ; CHECK:       %[[I3:.+]] = add i64 %offset.idx, -3
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %offset.idx, i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I1]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I2]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I3]], i32 0
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %offset.idx, i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I1]], i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I2]], i32 1
-; CHECK:       getelementptr inbounds %pair, %pair* %p, i64 %[[I3]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %offset.idx, i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %offset.idx, i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 1
+; CHECK:       getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 1
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-; INTER:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-; INTER:     LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
+; INTER:     LV: Found uniform instruction: %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+; INTER:     LV: Found uniform instruction: %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
 ; INTER:     vector.body
 ; INTER:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; INTER:       %offset.idx = sub i64 %n, %index
 ; INTER-NOT:   getelementptr
-; INTER:       %[[G0:.+]] = getelementptr inbounds %pair, %pair* %p, i64 %offset.idx, i32 0
-; INTER:       getelementptr inbounds i32, i32* %[[G0]], i64 -6
+; INTER:       %[[G0:.+]] = getelementptr inbounds %pair, ptr %p, i64 %offset.idx, i32 0
+; INTER:       getelementptr inbounds i32, ptr %[[G0]], i64 -6
 ; INTER-NOT:   getelementptr
 ; INTER:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define i32 @interleaved_access_reverse(%pair* %p, i64 %n) {
+define i32 @interleaved_access_reverse(ptr %p, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ %n, %entry ]
   %tmp0 = phi i32 [ %tmp6, %for.body ], [ 0, %entry ]
-  %tmp1 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-  %tmp2 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 1
-  %tmp3 = load i32, i32* %tmp1, align 8
-  %tmp4 = load i32, i32* %tmp2, align 8
+  %tmp1 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+  %tmp2 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 1
+  %tmp3 = load i32, ptr %tmp1, align 8
+  %tmp4 = load i32, ptr %tmp2, align 8
   %tmp5 = add i32 %tmp3, %tmp4
   %tmp6 = add i32 %tmp0, %tmp5
   %i.next = add nsw i64 %i, -1
@@ -197,18 +197,17 @@ for.end:
 ; as if it were uniform, the store is scalarized, making the pointer
 ; non-uniform.
 ;
-; INTER-NOT: LV: Found uniform instruction: %tmp0 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
+; INTER-NOT: LV: Found uniform instruction: %tmp0 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
 ; INTER:     vector.body
 ; INTER:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, {{.*}} ]
-; INTER:       %[[G0:.+]] = getelementptr inbounds %pair, %pair* %p, i64 %index, i32 0
-; INTER:       %[[B0:.+]] = bitcast i32* %[[G0]] to <8 x i32>*
-; INTER:       %wide.vec = load <8 x i32>, <8 x i32>* %[[B0]], align 8
+; INTER:       %[[G0:.+]] = getelementptr inbounds %pair, ptr %p, i64 %index, i32 0
+; INTER:       %wide.vec = load <8 x i32>, ptr %[[G0]], align 8
 ; INTER:       %[[I1:.+]] = or i64 %index, 1
-; INTER:       getelementptr inbounds %pair, %pair* %p, i64 %[[I1]], i32 0
+; INTER:       getelementptr inbounds %pair, ptr %p, i64 %[[I1]], i32 0
 ; INTER:       %[[I2:.+]] = or i64 %index, 2
-; INTER:       getelementptr inbounds %pair, %pair* %p, i64 %[[I2]], i32 0
+; INTER:       getelementptr inbounds %pair, ptr %p, i64 %[[I2]], i32 0
 ; INTER:       %[[I3:.+]] = or i64 %index, 3
-; INTER:       getelementptr inbounds %pair, %pair* %p, i64 %[[I3]], i32 0
+; INTER:       getelementptr inbounds %pair, ptr %p, i64 %[[I3]], i32 0
 ; INTER:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
 define void @predicated_store(%pair *%p, i32 %x, i64 %n) {
@@ -217,13 +216,13 @@ entry:
 
 for.body:
   %i  = phi i64 [ %i.next, %if.merge ], [ 0, %entry ]
-  %tmp0 = getelementptr inbounds %pair, %pair* %p, i64 %i, i32 0
-  %tmp1 = load i32, i32* %tmp0, align 8
+  %tmp0 = getelementptr inbounds %pair, ptr %p, i64 %i, i32 0
+  %tmp1 = load i32, ptr %tmp0, align 8
   %tmp2 = icmp eq i32 %tmp1, %x
   br i1 %tmp2, label %if.then, label %if.merge
 
 if.then:
-  store i32 %tmp1, i32* %tmp0, align 8
+  store i32 %tmp1, ptr %tmp0, align 8
   br label %if.merge
 
 if.merge:
@@ -241,27 +240,27 @@ for.end:
 ; as uniform and is not uniform after vectorization. The store is scalarized
 ; because the stored type may required padding.
 ;
-; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %i
+; CHECK-NOT: LV: Found uniform instruction: %tmp1 = getelementptr inbounds x86_fp80, ptr %a, i64 %i
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK:       %[[I1:.+]] = or i64 %index, 1
 ; CHECK:       %[[I2:.+]] = or i64 %index, 2
 ; CHECK:       %[[I3:.+]] = or i64 %index, 3
-; CHECK:       getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %index
-; CHECK:       getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %[[I1]]
-; CHECK:       getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %[[I2]]
-; CHECK:       getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %[[I3]]
+; CHECK:       getelementptr inbounds x86_fp80, ptr %a, i64 %index
+; CHECK:       getelementptr inbounds x86_fp80, ptr %a, i64 %[[I1]]
+; CHECK:       getelementptr inbounds x86_fp80, ptr %a, i64 %[[I2]]
+; CHECK:       getelementptr inbounds x86_fp80, ptr %a, i64 %[[I3]]
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define void @irregular_type(x86_fp80* %a, i64 %n) {
+define void @irregular_type(ptr %a, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ]
   %tmp0 = sitofp i32 1 to x86_fp80
-  %tmp1 = getelementptr inbounds x86_fp80, x86_fp80* %a, i64 %i
-  store x86_fp80 %tmp0, x86_fp80* %tmp1, align 16
+  %tmp1 = getelementptr inbounds x86_fp80, ptr %a, i64 %i
+  store x86_fp80 %tmp0, ptr %tmp1, align 16
   %i.next = add i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end
@@ -275,23 +274,24 @@ for.end:
 ; Check that a pointer induction variable is recognized as uniform and remains
 ; uniform after vectorization.
 ;
-; CHECK:     LV: Found uniform instruction: %p = phi i32* [ %tmp03, %for.body ], [ %a, %entry ]
+; CHECK:     LV: Found uniform instruction: %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ]
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
 ; CHECK-NOT:   getelementptr
-; CHECK:       %next.gep = getelementptr i32, i32* %a, i64 %index
+; CHECK:       [[SHL:%.+]] = shl i64 %index, 2
+; CHECK:       %next.gep = getelementptr i8, ptr %a, i64 [[SHL]]
 ; CHECK-NOT:   getelementptr
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define void @pointer_iv_uniform(i32* %a, i32 %x, i64 %n) {
+define void @pointer_iv_uniform(ptr %a, i32 %x, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
-  %p = phi i32* [ %tmp03, %for.body ], [ %a, %entry ]
-  store i32 %x, i32* %p, align 8
-  %tmp03 = getelementptr inbounds i32, i32* %p, i32 1
+  %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ]
+  store i32 %x, ptr %p, align 8
+  %tmp03 = getelementptr inbounds i32, ptr %p, i32 1
   %i.next = add nuw nsw i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end
@@ -307,42 +307,42 @@ for.end:
 ; induction variable is used by getelementptr instructions that are non-uniform
 ; due to scalarization of the stores.
 ;
-; INTER-NOT: LV: Found uniform instruction: %p = phi i32* [ %tmp03, %for.body ], [ %a, %entry ]
+; INTER-NOT: LV: Found uniform instruction: %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ]
 ; INTER:     vector.body
 ; INTER:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
-; INTER:       %[[I0:.+]] = shl i64 %index, 2
-; INTER:       %next.gep = getelementptr i32, i32* %a, i64 %[[I0]]
-; INTER:       %[[S1:.+]] = shl i64 %index, 2
-; INTER:       %[[I1:.+]] = or i64 %[[S1]], 4
-; INTER:       %next.gep2 = getelementptr i32, i32* %a, i64 %[[I1]]
-; INTER:       %[[S2:.+]] = shl i64 %index, 2
-; INTER:       %[[I2:.+]] = or i64 %[[S2]], 8
-; INTER:       %next.gep3 = getelementptr i32, i32* %a, i64 %[[I2]]
-; INTER:       %[[S3:.+]] = shl i64 %index, 2
-; INTER:       %[[I3:.+]] = or i64 %[[S3]], 12
-; INTER:       %next.gep4 = getelementptr i32, i32* %a, i64 %[[I3]]
+; INTER:       %[[I0:.+]] = shl i64 %index, 4
+; INTER:       %next.gep = getelementptr i8, ptr %a, i64 %[[I0]]
+; INTER:       %[[S1:.+]] = shl i64 %index, 4
+; INTER:       %[[I1:.+]] = or i64 %[[S1]], 16
+; INTER:       %next.gep2 = getelementptr i8, ptr %a, i64 %[[I1]]
+; INTER:       %[[S2:.+]] = shl i64 %index, 4
+; INTER:       %[[I2:.+]] = or i64 %[[S2]], 32
+; INTER:       %next.gep3 = getelementptr i8, ptr %a, i64 %[[I2]]
+; INTER:       %[[S3:.+]] = shl i64 %index, 4
+; INTER:       %[[I3:.+]] = or i64 %[[S3]], 48
+; INTER:       %next.gep4 = getelementptr i8, ptr %a, i64 %[[I3]]
 ; INTER:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define void @pointer_iv_non_uniform_0(i32* %a, i64 %n) {
+define void @pointer_iv_non_uniform_0(ptr %a, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
-  %p = phi i32* [ %tmp03, %for.body ], [ %a, %entry ]
-  %tmp00 = load i32, i32* %p, align 8
-  %tmp01 = getelementptr inbounds i32, i32* %p, i32 1
-  %tmp02 = load i32, i32* %tmp01, align 8
-  %tmp03 = getelementptr inbounds i32, i32* %p, i32 4
-  %tmp04 = load i32, i32* %tmp03, align 8
-  %tmp05 = getelementptr inbounds i32, i32* %p, i32 5
-  %tmp06 = load i32, i32* %tmp05, align 8
+  %p = phi ptr [ %tmp03, %for.body ], [ %a, %entry ]
+  %tmp00 = load i32, ptr %p, align 8
+  %tmp01 = getelementptr inbounds i32, ptr %p, i32 1
+  %tmp02 = load i32, ptr %tmp01, align 8
+  %tmp03 = getelementptr inbounds i32, ptr %p, i32 4
+  %tmp04 = load i32, ptr %tmp03, align 8
+  %tmp05 = getelementptr inbounds i32, ptr %p, i32 5
+  %tmp06 = load i32, ptr %tmp05, align 8
   %tmp07 = sub i32 %tmp04, %tmp00
   %tmp08 = sub i32 %tmp02, %tmp02
-  %tmp09 = getelementptr inbounds i32, i32* %p, i32 2
-  store i32 %tmp07, i32* %tmp09, align 8
-  %tmp10 = getelementptr inbounds i32, i32* %p, i32 3
-  store i32 %tmp08, i32* %tmp10, align 8
+  %tmp09 = getelementptr inbounds i32, ptr %p, i32 2
+  store i32 %tmp07, ptr %tmp09, align 8
+  %tmp10 = getelementptr inbounds i32, ptr %p, i32 3
+  store i32 %tmp08, ptr %tmp10, align 8
   %i.next = add nuw nsw i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end
@@ -357,28 +357,32 @@ for.end:
 ; recognized as uniform and is not uniform after vectorization. The pointer
 ; induction variable is used by a store that will be scalarized.
 ;
-; CHECK-NOT: LV: Found uniform instruction: %p = phi x86_fp80* [%tmp1, %for.body], [%a, %entry]
+; CHECK-NOT: LV: Found uniform instruction: %p = phi ptr [%tmp1, %for.body], [%a, %entry]
 ; CHECK:     vector.body
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
-; CHECK:       %next.gep = getelementptr x86_fp80, x86_fp80* %a, i64 %index
-; CHECK:       %[[I1:.+]] = or i64 %index, 1
-; CHECK:       %next.gep2 = getelementptr x86_fp80, x86_fp80* %a, i64 %[[I1]]
-; CHECK:       %[[I2:.+]] = or i64 %index, 2
-; CHECK:       %next.gep3 = getelementptr x86_fp80, x86_fp80* %a, i64 %[[I2]]
-; CHECK:       %[[I3:.+]] = or i64 %index, 3
-; CHECK:       %next.gep4 = getelementptr x86_fp80, x86_fp80* %a, i64 %[[I3]]
+; CHECK:       [[SHL1:%.+]] = shl i64 %index, 4
+; CHECK:       %next.gep = getelementptr i8, ptr %a, i64 [[SHL1]]
+; CHECK:       [[SHL2:%.+]] = shl i64 %index, 4
+; CHECK:       %[[I1:.+]] = or i64 [[SHL2]], 16
+; CHECK:       %next.gep2 = getelementptr i8, ptr %a, i64 %[[I1]]
+; CHECK:       [[SHL3:%.+]] = shl i64 %index, 4
+; CHECK:       %[[I2:.+]] = or i64 [[SHL3]], 32
+; CHECK:       %next.gep3 = getelementptr i8, ptr %a, i64 %[[I2]]
+; CHECK:       [[SHL4:%.+]] = shl i64 %index, 4
+; CHECK:       %[[I3:.+]] = or i64 [[SHL4]], 48
+; CHECK:       %next.gep4 = getelementptr i8, ptr %a, i64 %[[I3]]
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define void @pointer_iv_non_uniform_1(x86_fp80* %a, i64 %n) {
+define void @pointer_iv_non_uniform_1(ptr %a, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
-  %p = phi x86_fp80* [%tmp1, %for.body], [%a, %entry]
+  %p = phi ptr [%tmp1, %for.body], [%a, %entry]
   %tmp0 = sitofp i32 1 to x86_fp80
-  store x86_fp80 %tmp0, x86_fp80* %p, align 16
-  %tmp1 = getelementptr inbounds x86_fp80, x86_fp80* %p, i32 1
+  store x86_fp80 %tmp0, ptr %p, align 16
+  %tmp1 = getelementptr inbounds x86_fp80, ptr %p, i32 1
   %i.next = add i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end
@@ -394,32 +398,32 @@ for.end:
 ; variable is not recognized as uniform and is not uniform after vectorization
 ; because it is stored to memory.
 ;
-; CHECK-NOT: LV: Found uniform instruction: %p = phi i32* [ %tmp3, %for.body ], [ %a, %entry ]
-; CHECK:     LV: Found uniform instruction: %q = phi i32** [ %tmp4, %for.body ], [ %b, %entry ]
+; CHECK-NOT: LV: Found uniform instruction: %p = phi ptr [ %tmp3, %for.body ], [ %a, %entry ]
+; CHECK:     LV: Found uniform instruction: %q = phi ptr [ %tmp4, %for.body ], [ %b, %entry ]
 ; CHECK:     vector.body
-; CHECK:       %pointer.phi = phi i32* [ %a, %vector.ph ], [ %ptr.ind, %vector.body ]
+; CHECK:       %pointer.phi = phi ptr [ %a, %vector.ph ], [ %ptr.ind, %vector.body ]
 ; CHECK:       %index = phi i64 [ 0, %vector.ph ], [ %index.next, %vector.body ]
-; CHECK:       %[[PTRVEC:.+]] = getelementptr i32, i32* %pointer.phi, <4 x i64> <i64 0, i64 1, i64 2, i64 3>
-; CHECK:       %next.gep = getelementptr i32*, i32** %b, i64 %index
-; CHECK:       %[[NEXTGEPBC:.+]] = bitcast i32** %next.gep to <4 x i32*>*
-; CHECK:       store <4 x i32*> %[[PTRVEC]], <4 x i32*>* %[[NEXTGEPBC]], align 8
-; CHECK:       %ptr.ind = getelementptr i32, i32* %pointer.phi, i64 4
+; CHECK:       %[[PTRVEC:.+]] = getelementptr i8, ptr %pointer.phi, <4 x i64> <i64 0, i64 4, i64 8, i64 12>
+; CHECK:       [[SHL:%.+]] = shl i64 %index, 3
+; CHECK:       %next.gep = getelementptr i8, ptr %b, i64 [[SHL]]
+; CHECK:       store <4 x ptr> %[[PTRVEC]], ptr %next.gep, align 8
+; CHECK:       %ptr.ind = getelementptr i8, ptr %pointer.phi, i64 16
 ; CHECK:       br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define i32 @pointer_iv_mixed(i32* %a, i32** %b, i64 %n) {
+define i32 @pointer_iv_mixed(ptr %a, ptr %b, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ %i.next, %for.body ], [ 0, %entry ]
-  %p = phi i32* [ %tmp3, %for.body ], [ %a, %entry ]
-  %q = phi i32** [ %tmp4, %for.body ], [ %b, %entry ]
+  %p = phi ptr [ %tmp3, %for.body ], [ %a, %entry ]
+  %q = phi ptr [ %tmp4, %for.body ], [ %b, %entry ]
   %tmp0 = phi i32 [ %tmp2, %for.body ], [ 0, %entry ]
-  %tmp1 = load i32, i32* %p, align 8
+  %tmp1 = load i32, ptr %p, align 8
   %tmp2 = add i32 %tmp1, %tmp0
-  store i32* %p, i32** %q, align 8
-  %tmp3 = getelementptr inbounds i32, i32* %p, i32 1
-  %tmp4 = getelementptr inbounds i32*, i32** %q, i32 1
+  store ptr %p, ptr %q, align 8
+  %tmp3 = getelementptr inbounds i32, ptr %p, i32 1
+  %tmp4 = getelementptr inbounds ptr, ptr %q, i32 1
   %i.next = add nuw nsw i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end
@@ -429,49 +433,47 @@ for.end:
   ret i32 %tmp5
 }
 
-; INTER-LABEL: bitcast_pointer_operand
+; INTER-LABEL: pointer_operand_geps_with_
diff erent_indexed_types
 ;
 ; Check that a pointer operand having a user other than a memory access is
-; recognized as uniform after vectorization. In this test case, %tmp1 is a
-; bitcast that is used by a load and a getelementptr instruction (%tmp2). Once
-; %tmp2 is marked uniform, %tmp1 should be marked uniform as well.
+; recognized as uniform after vectorization. In this test case, %tmp0 is a
+; GEP that is used by a load and a getelementptr instruction (%tmp2). Once
+; %tmp2 is marked uniform, %tmp0 should be marked uniform as well.
 ;
 ; INTER:       LV: Found uniform instruction: %cond = icmp slt i64 %i.next, %n
-; INTER-NEXT:  LV: Found uniform instruction: %tmp2 = getelementptr inbounds i8, i8* %tmp1, i64 3
-; INTER-NEXT:  LV: Found uniform instruction: %tmp6 = getelementptr inbounds i8, i8* %B, i64 %i
-; INTER-NEXT:  LV: Found uniform instruction: %tmp1 = bitcast i64* %tmp0 to i8*
-; INTER-NEXT:  LV: Found uniform instruction: %tmp0 = getelementptr inbounds i64, i64* %A, i64 %i
+; INTER-NEXT:  LV: Found uniform instruction: %tmp2 = getelementptr inbounds i8, ptr %tmp0, i64 3
+; INTER-NEXT:  LV: Found uniform instruction: %tmp6 = getelementptr inbounds i8, ptr %B, i64 %i
+; INTER-NEXT:  LV: Found uniform instruction: %tmp0 = getelementptr inbounds i64, ptr %A, i64 %i
 ; INTER-NEXT:  LV: Found uniform instruction: %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ]
 ; INTER-NEXT:  LV: Found uniform instruction: %i.next = add nuw nsw i64 %i, 1
-; INTER:       define void @bitcast_pointer_operand(
+; INTER:       define void @pointer_operand_geps_with_
diff erent_indexed_types(
 ; INTER:       vector.body:
 ; INTER-NEXT:    [[INDEX:%.*]] = phi i64 [ 0, %vector.ph ], [ [[INDEX_NEXT:%.*]], %vector.body ]
-; INTER-NEXT:    [[TMP4:%.*]] = getelementptr inbounds i64, i64* %A, i64 [[INDEX]]
-; INTER-NEXT:    [[TMP5:%.*]] = bitcast i64* [[TMP4]] to <32 x i8>*
-; INTER-NEXT:    [[WIDE_VEC:%.*]] = load <32 x i8>, <32 x i8>* [[TMP5]], align 1
+; INTER-NEXT:    [[TMP4:%.*]] = getelementptr inbounds i64, ptr %A, i64 [[INDEX]]
+; INTER-NEXT:    [[TMP5:%.*]] = getelementptr inbounds i8, ptr [[TMP4]], i64 3
+; INTER-NEXT:    [[TMP6:%.*]] = getelementptr inbounds i8, ptr [[TMP5]], i64 -3
+; INTER-NEXT:    [[WIDE_VEC:%.*]] = load <32 x i8>, ptr [[TMP6]], align 1
 ; INTER-NEXT:    [[STRIDED_VEC:%.*]] = shufflevector <32 x i8> [[WIDE_VEC]], <32 x i8> poison, <4 x i32> <i32 0, i32 8, i32 16, i32 24>
 ; INTER-NEXT:    [[STRIDED_VEC5:%.*]] = shufflevector <32 x i8> [[WIDE_VEC]], <32 x i8> poison, <4 x i32> <i32 3, i32 11, i32 19, i32 27>
 ; INTER-NEXT:    [[TMP6:%.*]] = xor <4 x i8> [[STRIDED_VEC5]], [[STRIDED_VEC]]
-; INTER-NEXT:    [[TMP7:%.*]] = getelementptr inbounds i8, i8* %B, i64 [[INDEX]]
-; INTER-NEXT:    [[TMP8:%.*]] = bitcast i8* [[TMP7]] to <4 x i8>*
-; INTER-NEXT:    store <4 x i8> [[TMP6]], <4 x i8>* [[TMP8]], align 1
+; INTER-NEXT:    [[TMP7:%.*]] = getelementptr inbounds i8, ptr %B, i64 [[INDEX]]
+; INTER-NEXT:    store <4 x i8> [[TMP6]], ptr [[TMP7]], align 1
 ; INTER-NEXT:    [[INDEX_NEXT]] = add nuw i64 [[INDEX]], 4
 ; INTER:         br i1 {{.*}}, label %middle.block, label %vector.body
 ;
-define void @bitcast_pointer_operand(i64* %A, i8* %B, i64 %n) {
+define void @pointer_operand_geps_with_
diff erent_indexed_types(i64* %A, i8* %B, i64 %n) {
 entry:
   br label %for.body
 
 for.body:
   %i = phi i64 [ 0, %entry ], [ %i.next, %for.body ]
-  %tmp0 = getelementptr inbounds i64, i64* %A, i64 %i
-  %tmp1 = bitcast i64* %tmp0 to i8*
-  %tmp2 = getelementptr inbounds i8, i8* %tmp1, i64 3
-  %tmp3 = load i8, i8* %tmp2, align 1
-  %tmp4 = load i8, i8* %tmp1, align 1
+  %tmp0 = getelementptr inbounds i64, ptr %A, i64 %i
+  %tmp2 = getelementptr inbounds i8, ptr %tmp0, i64 3
+  %tmp3 = load i8, ptr %tmp2, align 1
+  %tmp4 = load i8, ptr %tmp0, align 1
   %tmp5 = xor i8 %tmp3, %tmp4
-  %tmp6 = getelementptr inbounds i8, i8* %B, i64 %i
-  store i8 %tmp5, i8* %tmp6
+  %tmp6 = getelementptr inbounds i8, ptr %B, i64 %i
+  store i8 %tmp5, ptr %tmp6
   %i.next = add nuw nsw i64 %i, 1
   %cond = icmp slt i64 %i.next, %n
   br i1 %cond, label %for.body, label %for.end


        


More information about the llvm-commits mailing list