[llvm] r365828 - [InstCombine][NFCI] Add more test coverage to onehot_merge.ll

Huihui Zhang via llvm-commits llvm-commits at lists.llvm.org
Thu Jul 11 14:28:25 PDT 2019


Author: huihuiz
Date: Thu Jul 11 14:28:25 2019
New Revision: 365828

URL: http://llvm.org/viewvc/llvm-project?rev=365828&view=rev
Log:
[InstCombine][NFCI] Add more test coverage to onehot_merge.ll

Prep work for upcoming patch D64275.

Modified:
    llvm/trunk/test/Transforms/InstCombine/onehot_merge.ll

Modified: llvm/trunk/test/Transforms/InstCombine/onehot_merge.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/Transforms/InstCombine/onehot_merge.ll?rev=365828&r1=365827&r2=365828&view=diff
==============================================================================
--- llvm/trunk/test/Transforms/InstCombine/onehot_merge.ll (original)
+++ llvm/trunk/test/Transforms/InstCombine/onehot_merge.ll Thu Jul 11 14:28:25 2019
@@ -224,6 +224,145 @@ define i1 @foo1_or_signbit_lshr_without_
 ; Extra use
 
 ; Expect to fold
+define i1 @foo1_and_extra_use_shl(i32 %k, i32 %c1, i32 %c2, i32* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_shl(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    store i32 [[T0]], i32* [[P:%.*]], align 4
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K:%.*]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  store i32 %t0, i32* %p  ; extra use of shl
+  %t1 = shl i32 1, %c2
+  %t2 = and i32 %t0, %k
+  %t3 = icmp eq i32 %t2, 0
+  %t4 = and i32 %t1, %k
+  %t5 = icmp eq i32 %t4, 0
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Should not fold
+define i1 @foo1_and_extra_use_and(i32 %k, i32 %c1, i32 %c2, i32* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_and(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    [[T2:%.*]] = and i32 [[T0]], [[K:%.*]]
+; CHECK-NEXT:    store i32 [[T2]], i32* [[P:%.*]], align 4
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  %t1 = shl i32 1, %c2
+  %t2 = and i32 %t0, %k
+  store i32 %t2, i32* %p  ; extra use of and
+  %t3 = icmp eq i32 %t2, 0
+  %t4 = and i32 %t1, %k
+  %t5 = icmp eq i32 %t4, 0
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Should not fold
+define i1 @foo1_and_extra_use_cmp(i32 %k, i32 %c1, i32 %c2, i1* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_cmp(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    [[T2:%.*]] = and i32 [[T0]], [[K:%.*]]
+; CHECK-NEXT:    [[T3:%.*]] = icmp eq i32 [[T2]], 0
+; CHECK-NEXT:    store i1 [[T3]], i1* [[P:%.*]], align 1
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  %t1 = shl i32 1, %c2
+  %t2 = and i32 %t0, %k
+  %t3 = icmp eq i32 %t2, 0
+  store i1 %t3, i1* %p  ; extra use of cmp
+  %t4 = and i32 %t1, %k
+  %t5 = icmp eq i32 %t4, 0
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Expect to fold
+define i1 @foo1_and_extra_use_shl2(i32 %k, i32 %c1, i32 %c2, i32* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_shl2(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    store i32 [[T1]], i32* [[P:%.*]], align 4
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K:%.*]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  %t1 = shl i32 1, %c2
+  store i32 %t1, i32* %p  ; extra use of shl
+  %t2 = and i32 %t0, %k
+  %t3 = icmp eq i32 %t2, 0
+  %t4 = and i32 %t1, %k
+  %t5 = icmp eq i32 %t4, 0
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Should not fold
+define i1 @foo1_and_extra_use_and2(i32 %k, i32 %c1, i32 %c2, i32* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_and2(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    [[T4:%.*]] = and i32 [[T1]], [[K:%.*]]
+; CHECK-NEXT:    store i32 [[T4]], i32* [[P:%.*]], align 4
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  %t1 = shl i32 1, %c2
+  %t2 = and i32 %t0, %k
+  %t3 = icmp eq i32 %t2, 0
+  %t4 = and i32 %t1, %k
+  store i32 %t4, i32* %p  ; extra use of and
+  %t5 = icmp eq i32 %t4, 0
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Should not fold
+define i1 @foo1_and_extra_use_cmp2(i32 %k, i32 %c1, i32 %c2, i1* %p) {
+; CHECK-LABEL: @foo1_and_extra_use_cmp2(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = shl i32 1, [[C2:%.*]]
+; CHECK-NEXT:    [[T4:%.*]] = and i32 [[T1]], [[K:%.*]]
+; CHECK-NEXT:    [[T5:%.*]] = icmp eq i32 [[T4]], 0
+; CHECK-NEXT:    store i1 [[T5]], i1* [[P:%.*]], align 1
+; CHECK-NEXT:    [[TMP1:%.*]] = or i32 [[T0]], [[T1]]
+; CHECK-NEXT:    [[TMP2:%.*]] = and i32 [[TMP1]], [[K]]
+; CHECK-NEXT:    [[TMP3:%.*]] = icmp ne i32 [[TMP2]], [[TMP1]]
+; CHECK-NEXT:    ret i1 [[TMP3]]
+;
+  %t0 = shl i32 1, %c1
+  %t1 = shl i32 1, %c2
+  %t2 = and i32 %t0, %k
+  %t3 = icmp eq i32 %t2, 0
+  %t4 = and i32 %t1, %k
+  %t5 = icmp eq i32 %t4, 0
+  store i1 %t5, i1* %p  ; extra use of cmp
+  %or = or i1 %t3, %t5
+  ret i1 %or
+}
+
+; Shift-of-signbit replaced with 'icmp s*'
+; Expect to fold
 define i1 @foo1_and_signbit_lshr_without_shifting_signbit_extra_use_shl1(i32 %k, i32 %c1, i32 %c2, i32* %p) {
 ; CHECK-LABEL: @foo1_and_signbit_lshr_without_shifting_signbit_extra_use_shl1(
 ; CHECK-NEXT:    [[T0:%.*]] = shl i32 1, [[C1:%.*]]
@@ -332,3 +471,25 @@ define i1 @foo1_and_signbit_lshr_without
   %or = or i1 %t2, %t4
   ret i1 %or
 }
+
+; Negative tests
+
+; This test checks that we are not creating additional shift instruction when fold fails.
+define i1 @foo1_and_signbit_lshr_without_shifting_signbit_not_pwr2(i32 %k, i32 %c1, i32 %c2) {
+; CHECK-LABEL: @foo1_and_signbit_lshr_without_shifting_signbit_not_pwr2(
+; CHECK-NEXT:    [[T0:%.*]] = shl i32 3, [[C1:%.*]]
+; CHECK-NEXT:    [[T1:%.*]] = and i32 [[T0]], [[K:%.*]]
+; CHECK-NEXT:    [[T2:%.*]] = icmp eq i32 [[T1]], 0
+; CHECK-NEXT:    [[T3:%.*]] = shl i32 [[K]], [[C2:%.*]]
+; CHECK-NEXT:    [[T4:%.*]] = icmp sgt i32 [[T3]], -1
+; CHECK-NEXT:    [[OR:%.*]] = or i1 [[T2]], [[T4]]
+; CHECK-NEXT:    ret i1 [[OR]]
+;
+  %t0 = shl i32 3, %c1
+  %t1 = and i32 %t0, %k
+  %t2 = icmp eq i32 %t1, 0
+  %t3 = shl i32 %k, %c2
+  %t4 = icmp sgt i32 %t3, -1
+  %or = or i1 %t2, %t4
+  ret i1 %or
+}




More information about the llvm-commits mailing list