[llvm] 61d1847 - [InstCombine] Add 8/16/32/64 bitreverse test coverage
Simon Pilgrim via llvm-commits
llvm-commits at lists.llvm.org
Fri Oct 23 07:50:40 PDT 2020
Author: Simon Pilgrim
Date: 2020-10-23T15:39:56+01:00
New Revision: 61d1847b12a91e148a7f6bc32a7a2d490bc19ffd
URL: https://github.com/llvm/llvm-project/commit/61d1847b12a91e148a7f6bc32a7a2d490bc19ffd
DIFF: https://github.com/llvm/llvm-project/commit/61d1847b12a91e148a7f6bc32a7a2d490bc19ffd.diff
LOG: [InstCombine] Add 8/16/32/64 bitreverse test coverage
Use typical codegen for the traditional pairwise lgN bitreverse algorithm
Added:
Modified:
llvm/test/Transforms/InstCombine/bitreverse.ll
Removed:
################################################################################
diff --git a/llvm/test/Transforms/InstCombine/bitreverse.ll b/llvm/test/Transforms/InstCombine/bitreverse.ll
index 1b84119bcb42..72b95aa894e4 100644
--- a/llvm/test/Transforms/InstCombine/bitreverse.ll
+++ b/llvm/test/Transforms/InstCombine/bitreverse.ll
@@ -3,15 +3,203 @@
target datalayout = "e-p:32:32:32-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:32:64-f32:32:32-f64:32:64-v64:64:64-v128:128:128-a0:0:64-f80:128:128-n8:16:32"
-;unsigned char rev8(unsigned char x) {
+;pairwise reverse
+;template <typename T>
+;T reverse(T v) {
+; T s = sizeof(v) * 8;
+; T mask = ~(T)0;
+; while ((s >>= 1) > 0) {
+; mask ^= (mask << s);
+; v = ((v >> s) & mask) | ((v << s) & ~mask);
+; }
+; return v;
+;}
+define i8 @rev8(i8 %v) {
+; CHECK-LABEL: @rev8(
+; CHECK-NEXT: entry:
+; CHECK-NEXT: [[OR:%.*]] = call i8 @llvm.fshl.i8(i8 [[V:%.*]], i8 [[V]], i8 4)
+; CHECK-NEXT: [[SHR4_1:%.*]] = lshr i8 [[OR]], 2
+; CHECK-NEXT: [[AND_1:%.*]] = and i8 [[SHR4_1]], 51
+; CHECK-NEXT: [[SHL7_1:%.*]] = shl i8 [[OR]], 2
+; CHECK-NEXT: [[AND9_1:%.*]] = and i8 [[SHL7_1]], -52
+; CHECK-NEXT: [[OR_1:%.*]] = or i8 [[AND_1]], [[AND9_1]]
+; CHECK-NEXT: [[SHR4_2:%.*]] = lshr i8 [[OR_1]], 1
+; CHECK-NEXT: [[AND_2:%.*]] = and i8 [[SHR4_2]], 85
+; CHECK-NEXT: [[SHL7_2:%.*]] = shl i8 [[OR_1]], 1
+; CHECK-NEXT: [[AND9_2:%.*]] = and i8 [[SHL7_2]], -86
+; CHECK-NEXT: [[OR_2:%.*]] = or i8 [[AND_2]], [[AND9_2]]
+; CHECK-NEXT: ret i8 [[OR_2]]
+;
+entry:
+ %shr4 = lshr i8 %v, 4
+ %shl7 = shl i8 %v, 4
+ %or = or i8 %shr4, %shl7
+ %shr4.1 = lshr i8 %or, 2
+ %and.1 = and i8 %shr4.1, 51
+ %shl7.1 = shl i8 %or, 2
+ %and9.1 = and i8 %shl7.1, -52
+ %or.1 = or i8 %and.1, %and9.1
+ %shr4.2 = lshr i8 %or.1, 1
+ %and.2 = and i8 %shr4.2, 85
+ %shl7.2 = shl i8 %or.1, 1
+ %and9.2 = and i8 %shl7.2, -86
+ %or.2 = or i8 %and.2, %and9.2
+ ret i8 %or.2
+}
+
+define i16 @rev16(i16 %v) {
+; CHECK-LABEL: @rev16(
+; CHECK-NEXT: entry:
+; CHECK-NEXT: [[OR:%.*]] = call i16 @llvm.bswap.i16(i16 [[V:%.*]])
+; CHECK-NEXT: [[SHR4_1:%.*]] = lshr i16 [[OR]], 4
+; CHECK-NEXT: [[AND_1:%.*]] = and i16 [[SHR4_1]], 3855
+; CHECK-NEXT: [[SHL7_1:%.*]] = shl i16 [[OR]], 4
+; CHECK-NEXT: [[AND9_1:%.*]] = and i16 [[SHL7_1]], -3856
+; CHECK-NEXT: [[OR_1:%.*]] = or i16 [[AND_1]], [[AND9_1]]
+; CHECK-NEXT: [[SHR4_2:%.*]] = lshr i16 [[OR_1]], 2
+; CHECK-NEXT: [[AND_2:%.*]] = and i16 [[SHR4_2]], 13107
+; CHECK-NEXT: [[SHL7_2:%.*]] = shl i16 [[OR_1]], 2
+; CHECK-NEXT: [[AND9_2:%.*]] = and i16 [[SHL7_2]], -13108
+; CHECK-NEXT: [[OR_2:%.*]] = or i16 [[AND_2]], [[AND9_2]]
+; CHECK-NEXT: [[SHR4_3:%.*]] = lshr i16 [[OR_2]], 1
+; CHECK-NEXT: [[AND_3:%.*]] = and i16 [[SHR4_3]], 21845
+; CHECK-NEXT: [[SHL7_3:%.*]] = shl i16 [[OR_2]], 1
+; CHECK-NEXT: [[AND9_3:%.*]] = and i16 [[SHL7_3]], -21846
+; CHECK-NEXT: [[OR_3:%.*]] = or i16 [[AND_3]], [[AND9_3]]
+; CHECK-NEXT: ret i16 [[OR_3]]
+;
+entry:
+ %shr4 = lshr i16 %v, 8
+ %shl7 = shl i16 %v, 8
+ %or = or i16 %shr4, %shl7
+ %shr4.1 = lshr i16 %or, 4
+ %and.1 = and i16 %shr4.1, 3855
+ %shl7.1 = shl i16 %or, 4
+ %and9.1 = and i16 %shl7.1, -3856
+ %or.1 = or i16 %and.1, %and9.1
+ %shr4.2 = lshr i16 %or.1, 2
+ %and.2 = and i16 %shr4.2, 13107
+ %shl7.2 = shl i16 %or.1, 2
+ %and9.2 = and i16 %shl7.2, -13108
+ %or.2 = or i16 %and.2, %and9.2
+ %shr4.3 = lshr i16 %or.2, 1
+ %and.3 = and i16 %shr4.3, 21845
+ %shl7.3 = shl i16 %or.2, 1
+ %and9.3 = and i16 %shl7.3, -21846
+ %or.3 = or i16 %and.3, %and9.3
+ ret i16 %or.3
+}
+
+define i32 @rev32(i32 %v) {
+; CHECK-LABEL: @rev32(
+; CHECK-NEXT: entry:
+; CHECK-NEXT: [[OR_1:%.*]] = call i32 @llvm.bswap.i32(i32 [[V:%.*]])
+; CHECK-NEXT: [[SHR1_2:%.*]] = lshr i32 [[OR_1]], 4
+; CHECK-NEXT: [[AND_2:%.*]] = and i32 [[SHR1_2]], 252645135
+; CHECK-NEXT: [[SHL2_2:%.*]] = shl i32 [[OR_1]], 4
+; CHECK-NEXT: [[AND3_2:%.*]] = and i32 [[SHL2_2]], -252645136
+; CHECK-NEXT: [[OR_2:%.*]] = or i32 [[AND_2]], [[AND3_2]]
+; CHECK-NEXT: [[SHR1_3:%.*]] = lshr i32 [[OR_2]], 2
+; CHECK-NEXT: [[AND_3:%.*]] = and i32 [[SHR1_3]], 858993459
+; CHECK-NEXT: [[SHL2_3:%.*]] = shl i32 [[OR_2]], 2
+; CHECK-NEXT: [[AND3_3:%.*]] = and i32 [[SHL2_3]], -858993460
+; CHECK-NEXT: [[OR_3:%.*]] = or i32 [[AND_3]], [[AND3_3]]
+; CHECK-NEXT: [[SHR1_4:%.*]] = lshr i32 [[OR_3]], 1
+; CHECK-NEXT: [[AND_4:%.*]] = and i32 [[SHR1_4]], 1431655765
+; CHECK-NEXT: [[SHL2_4:%.*]] = shl i32 [[OR_3]], 1
+; CHECK-NEXT: [[AND3_4:%.*]] = and i32 [[SHL2_4]], -1431655766
+; CHECK-NEXT: [[OR_4:%.*]] = or i32 [[AND_4]], [[AND3_4]]
+; CHECK-NEXT: ret i32 [[OR_4]]
+;
+entry:
+ %shr1 = lshr i32 %v, 16
+ %shl2 = shl i32 %v, 16
+ %or = or i32 %shr1, %shl2
+ %shr1.1 = lshr i32 %or, 8
+ %and.1 = and i32 %shr1.1, 16711935
+ %shl2.1 = shl i32 %or, 8
+ %and3.1 = and i32 %shl2.1, -16711936
+ %or.1 = or i32 %and.1, %and3.1
+ %shr1.2 = lshr i32 %or.1, 4
+ %and.2 = and i32 %shr1.2, 252645135
+ %shl2.2 = shl i32 %or.1, 4
+ %and3.2 = and i32 %shl2.2, -252645136
+ %or.2 = or i32 %and.2, %and3.2
+ %shr1.3 = lshr i32 %or.2, 2
+ %and.3 = and i32 %shr1.3, 858993459
+ %shl2.3 = shl i32 %or.2, 2
+ %and3.3 = and i32 %shl2.3, -858993460
+ %or.3 = or i32 %and.3, %and3.3
+ %shr1.4 = lshr i32 %or.3, 1
+ %and.4 = and i32 %shr1.4, 1431655765
+ %shl2.4 = shl i32 %or.3, 1
+ %and3.4 = and i32 %shl2.4, -1431655766
+ %or.4 = or i32 %and.4, %and3.4
+ ret i32 %or.4
+}
+
+define i64 @rev64(i64 %v) {
+; CHECK-LABEL: @rev64(
+; CHECK-NEXT: entry:
+; CHECK-NEXT: [[OR_2:%.*]] = call i64 @llvm.bswap.i64(i64 [[V:%.*]])
+; CHECK-NEXT: [[SHR2_3:%.*]] = lshr i64 [[OR_2]], 4
+; CHECK-NEXT: [[AND_3:%.*]] = and i64 [[SHR2_3]], 1085102592571150095
+; CHECK-NEXT: [[SHL4_3:%.*]] = shl i64 [[OR_2]], 4
+; CHECK-NEXT: [[AND5_3:%.*]] = and i64 [[SHL4_3]], -1085102592571150096
+; CHECK-NEXT: [[OR_3:%.*]] = or i64 [[AND_3]], [[AND5_3]]
+; CHECK-NEXT: [[SHR2_4:%.*]] = lshr i64 [[OR_3]], 2
+; CHECK-NEXT: [[AND_4:%.*]] = and i64 [[SHR2_4]], 3689348814741910323
+; CHECK-NEXT: [[SHL4_4:%.*]] = shl i64 [[OR_3]], 2
+; CHECK-NEXT: [[AND5_4:%.*]] = and i64 [[SHL4_4]], -3689348814741910324
+; CHECK-NEXT: [[OR_4:%.*]] = or i64 [[AND_4]], [[AND5_4]]
+; CHECK-NEXT: [[SHR2_5:%.*]] = lshr i64 [[OR_4]], 1
+; CHECK-NEXT: [[AND_5:%.*]] = and i64 [[SHR2_5]], 6148914691236517205
+; CHECK-NEXT: [[SHL4_5:%.*]] = shl i64 [[OR_4]], 1
+; CHECK-NEXT: [[AND5_5:%.*]] = and i64 [[SHL4_5]], -6148914691236517206
+; CHECK-NEXT: [[OR_5:%.*]] = or i64 [[AND_5]], [[AND5_5]]
+; CHECK-NEXT: ret i64 [[OR_5]]
+;
+entry:
+ %shr2 = lshr i64 %v, 32
+ %shl4 = shl i64 %v, 32
+ %or = or i64 %shr2, %shl4
+ %shr2.1 = lshr i64 %or, 16
+ %and.1 = and i64 %shr2.1, 281470681808895
+ %shl4.1 = shl i64 %or, 16
+ %and5.1 = and i64 %shl4.1, -281470681808896
+ %or.1 = or i64 %and.1, %and5.1
+ %shr2.2 = lshr i64 %or.1, 8
+ %and.2 = and i64 %shr2.2, 71777214294589695
+ %shl4.2 = shl i64 %or.1, 8
+ %and5.2 = and i64 %shl4.2, -71777214294589696
+ %or.2 = or i64 %and.2, %and5.2
+ %shr2.3 = lshr i64 %or.2, 4
+ %and.3 = and i64 %shr2.3, 1085102592571150095
+ %shl4.3 = shl i64 %or.2, 4
+ %and5.3 = and i64 %shl4.3, -1085102592571150096
+ %or.3 = or i64 %and.3, %and5.3
+ %shr2.4 = lshr i64 %or.3, 2
+ %and.4 = and i64 %shr2.4, 3689348814741910323
+ %shl4.4 = shl i64 %or.3, 2
+ %and5.4 = and i64 %shl4.4, -3689348814741910324
+ %or.4 = or i64 %and.4, %and5.4
+ %shr2.5 = lshr i64 %or.4, 1
+ %and.5 = and i64 %shr2.5, 6148914691236517205
+ %shl4.5 = shl i64 %or.4, 1
+ %and5.5 = and i64 %shl4.5, -6148914691236517206
+ %or.5 = or i64 %and.5, %and5.5
+ ret i64 %or.5
+}
+
+;unsigned char rev8_xor(unsigned char x) {
; unsigned char y;
; y = x&0x55; x ^= y; x |= (y<<2)|(y>>6);
; y = x&0x66; x ^= y; x |= (y<<4)|(y>>4);
; return (x<<1)|(x>>7);
;}
-define i8 @test1(i8 %0) {
-; CHECK-LABEL: @test1(
+define i8 @rev8_xor(i8 %0) {
+; CHECK-LABEL: @rev8_xor(
; CHECK-NEXT: [[TMP2:%.*]] = and i8 [[TMP0:%.*]], 85
; CHECK-NEXT: [[TMP3:%.*]] = and i8 [[TMP0]], -86
; CHECK-NEXT: [[TMP4:%.*]] = shl i8 [[TMP2]], 2
@@ -47,8 +235,8 @@ define i8 @test1(i8 %0) {
ret i8 %16
}
-define <2 x i8> @test1_vector(<2 x i8> %0) {
-; CHECK-LABEL: @test1_vector(
+define <2 x i8> @rev8_xor_vector(<2 x i8> %0) {
+; CHECK-LABEL: @rev8_xor_vector(
; CHECK-NEXT: [[TMP2:%.*]] = and <2 x i8> [[TMP0:%.*]], <i8 85, i8 85>
; CHECK-NEXT: [[TMP3:%.*]] = and <2 x i8> [[TMP0]], <i8 -86, i8 -86>
; CHECK-NEXT: [[TMP4:%.*]] = shl <2 x i8> [[TMP2]], <i8 2, i8 2>
More information about the llvm-commits
mailing list