[llvm] 9606c69 - [InstCombine] Fold sub(Y,and(lshr(X,C),1)) --> add(ashr(shl(X,(BW-1)-C),BW-1),Y) (PR53610)
Simon Pilgrim via llvm-commits
llvm-commits at lists.llvm.org
Tue Feb 15 05:24:34 PST 2022
Author: Simon Pilgrim
Date: 2022-02-15T13:24:20Z
New Revision: 9606c690873fab93d895b2b763d8e2ed781c05af
URL: https://github.com/llvm/llvm-project/commit/9606c690873fab93d895b2b763d8e2ed781c05af
DIFF: https://github.com/llvm/llvm-project/commit/9606c690873fab93d895b2b763d8e2ed781c05af.diff
LOG: [InstCombine] Fold sub(Y,and(lshr(X,C),1)) --> add(ashr(shl(X,(BW-1)-C),BW-1),Y) (PR53610)
As noted on PR53610, we can fold a 'bit splat' negation of a shifted bitmask pattern into a pair of shifts.
https://alive2.llvm.org/ce/z/eGrsoN
Differential Revision: https://reviews.llvm.org/D119715
Added:
Modified:
llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
llvm/test/Transforms/InstCombine/negated-bitmask.ll
Removed:
################################################################################
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp b/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
index 42ba4a34a5a9d..c573b03f31a62 100644
--- a/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
+++ b/llvm/lib/Transforms/InstCombine/InstCombineNegator.cpp
@@ -248,6 +248,20 @@ LLVM_NODISCARD Value *Negator::visitImpl(Value *V, unsigned Depth) {
return nullptr;
switch (I->getOpcode()) {
+ case Instruction::And: {
+ Constant *ShAmt;
+ // sub(y,and(lshr(x,C),1)) --> add(ashr(shl(x,(BW-1)-C),BW-1),y)
+ if (match(I, m_c_And(m_OneUse(m_TruncOrSelf(
+ m_LShr(m_Value(X), m_ImmConstant(ShAmt)))),
+ m_One()))) {
+ unsigned BW = X->getType()->getScalarSizeInBits();
+ Constant *BWMinusOne = ConstantInt::get(X->getType(), BW - 1);
+ Value *R = Builder.CreateShl(X, Builder.CreateSub(BWMinusOne, ShAmt));
+ R = Builder.CreateAShr(R, BWMinusOne);
+ return Builder.CreateTruncOrBitCast(R, I->getType());
+ }
+ break;
+ }
case Instruction::SDiv:
// `sdiv` is negatible if divisor is not undef/INT_MIN/1.
// While this is normally not behind a use-check,
diff --git a/llvm/test/Transforms/InstCombine/negated-bitmask.ll b/llvm/test/Transforms/InstCombine/negated-bitmask.ll
index 5dd6a30aac969..c8270fe009850 100644
--- a/llvm/test/Transforms/InstCombine/negated-bitmask.ll
+++ b/llvm/test/Transforms/InstCombine/negated-bitmask.ll
@@ -6,10 +6,9 @@
define i8 @neg_mask1_lshr(i8 %a0) {
; CHECK-LABEL: @neg_mask1_lshr(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr i8 [[A0:%.*]], 3
-; CHECK-NEXT: [[MASK:%.*]] = and i8 [[SHIFT]], 1
-; CHECK-NEXT: [[NEG:%.*]] = sub nsw i8 0, [[MASK]]
-; CHECK-NEXT: ret i8 [[NEG]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[A0:%.*]], 4
+; CHECK-NEXT: [[TMP2:%.*]] = ashr i8 [[TMP1]], 7
+; CHECK-NEXT: ret i8 [[TMP2]]
;
%shift = lshr i8 %a0, 3
%mask = and i8 %shift, 1
@@ -19,9 +18,9 @@ define i8 @neg_mask1_lshr(i8 %a0) {
define i8 @sub_mask1_lshr(i8 %a0) {
; CHECK-LABEL: @sub_mask1_lshr(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr i8 [[A0:%.*]], 1
-; CHECK-NEXT: [[MASK:%.*]] = and i8 [[SHIFT]], 1
-; CHECK-NEXT: [[NEG:%.*]] = sub nuw nsw i8 10, [[MASK]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl i8 [[A0:%.*]], 6
+; CHECK-NEXT: [[TMP2:%.*]] = ashr i8 [[TMP1]], 7
+; CHECK-NEXT: [[NEG:%.*]] = add nsw i8 [[TMP2]], 10
; CHECK-NEXT: ret i8 [[NEG]]
;
%shift = lshr i8 %a0, 1
@@ -32,10 +31,9 @@ define i8 @sub_mask1_lshr(i8 %a0) {
define <4 x i32> @neg_mask1_lshr_vector_uniform(<4 x i32> %a0) {
; CHECK-LABEL: @neg_mask1_lshr_vector_uniform(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr <4 x i32> [[A0:%.*]], <i32 3, i32 3, i32 3, i32 3>
-; CHECK-NEXT: [[MASK:%.*]] = and <4 x i32> [[SHIFT]], <i32 1, i32 1, i32 1, i32 1>
-; CHECK-NEXT: [[NEG:%.*]] = sub nsw <4 x i32> zeroinitializer, [[MASK]]
-; CHECK-NEXT: ret <4 x i32> [[NEG]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl <4 x i32> [[A0:%.*]], <i32 28, i32 28, i32 28, i32 28>
+; CHECK-NEXT: [[TMP2:%.*]] = ashr <4 x i32> [[TMP1]], <i32 31, i32 31, i32 31, i32 31>
+; CHECK-NEXT: ret <4 x i32> [[TMP2]]
;
%shift = lshr <4 x i32> %a0, <i32 3, i32 3, i32 3, i32 3>
%mask = and <4 x i32> %shift, <i32 1, i32 1, i32 1, i32 1>
@@ -45,10 +43,9 @@ define <4 x i32> @neg_mask1_lshr_vector_uniform(<4 x i32> %a0) {
define <4 x i32> @neg_mask1_lshr_vector_nonuniform(<4 x i32> %a0) {
; CHECK-LABEL: @neg_mask1_lshr_vector_nonuniform(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr <4 x i32> [[A0:%.*]], <i32 3, i32 4, i32 5, i32 6>
-; CHECK-NEXT: [[MASK:%.*]] = and <4 x i32> [[SHIFT]], <i32 1, i32 1, i32 1, i32 1>
-; CHECK-NEXT: [[NEG:%.*]] = sub nsw <4 x i32> zeroinitializer, [[MASK]]
-; CHECK-NEXT: ret <4 x i32> [[NEG]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl <4 x i32> [[A0:%.*]], <i32 28, i32 27, i32 26, i32 25>
+; CHECK-NEXT: [[TMP2:%.*]] = ashr <4 x i32> [[TMP1]], <i32 31, i32 31, i32 31, i32 31>
+; CHECK-NEXT: ret <4 x i32> [[TMP2]]
;
%shift = lshr <4 x i32> %a0, <i32 3, i32 4, i32 5, i32 6>
%mask = and <4 x i32> %shift, <i32 1, i32 1, i32 1, i32 1>
@@ -58,9 +55,9 @@ define <4 x i32> @neg_mask1_lshr_vector_nonuniform(<4 x i32> %a0) {
define <4 x i32> @sub_mask1_lshr_vector_nonuniform(<4 x i32> %a0) {
; CHECK-LABEL: @sub_mask1_lshr_vector_nonuniform(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr <4 x i32> [[A0:%.*]], <i32 3, i32 4, i32 5, i32 6>
-; CHECK-NEXT: [[MASK:%.*]] = and <4 x i32> [[SHIFT]], <i32 1, i32 1, i32 1, i32 1>
-; CHECK-NEXT: [[NEG:%.*]] = sub nsw <4 x i32> <i32 5, i32 0, i32 -1, i32 65556>, [[MASK]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl <4 x i32> [[A0:%.*]], <i32 28, i32 27, i32 26, i32 25>
+; CHECK-NEXT: [[TMP2:%.*]] = ashr <4 x i32> [[TMP1]], <i32 31, i32 31, i32 31, i32 31>
+; CHECK-NEXT: [[NEG:%.*]] = add nsw <4 x i32> [[TMP2]], <i32 5, i32 0, i32 -1, i32 65556>
; CHECK-NEXT: ret <4 x i32> [[NEG]]
;
%shift = lshr <4 x i32> %a0, <i32 3, i32 4, i32 5, i32 6>
@@ -71,10 +68,10 @@ define <4 x i32> @sub_mask1_lshr_vector_nonuniform(<4 x i32> %a0) {
define i8 @sub_mask1_trunc_lshr(i64 %a0) {
; CHECK-LABEL: @sub_mask1_trunc_lshr(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr i64 [[A0:%.*]], 15
-; CHECK-NEXT: [[TRUNC:%.*]] = trunc i64 [[SHIFT]] to i8
-; CHECK-NEXT: [[MASK:%.*]] = and i8 [[TRUNC]], 1
-; CHECK-NEXT: [[NEG:%.*]] = sub nuw nsw i8 10, [[MASK]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl i64 [[A0:%.*]], 48
+; CHECK-NEXT: [[TMP2:%.*]] = ashr i64 [[TMP1]], 63
+; CHECK-NEXT: [[TMP3:%.*]] = trunc i64 [[TMP2]] to i8
+; CHECK-NEXT: [[NEG:%.*]] = add i8 [[TMP3]], 10
; CHECK-NEXT: ret i8 [[NEG]]
;
%shift = lshr i64 %a0, 15
@@ -86,10 +83,10 @@ define i8 @sub_mask1_trunc_lshr(i64 %a0) {
define i32 @sub_sext_mask1_trunc_lshr(i64 %a0) {
; CHECK-LABEL: @sub_sext_mask1_trunc_lshr(
-; CHECK-NEXT: [[SHIFT:%.*]] = lshr i64 [[A0:%.*]], 15
-; CHECK-NEXT: [[TRUNC:%.*]] = trunc i64 [[SHIFT]] to i8
-; CHECK-NEXT: [[MASK:%.*]] = and i8 [[TRUNC]], 1
-; CHECK-NEXT: [[NARROW:%.*]] = sub nuw nsw i8 10, [[MASK]]
+; CHECK-NEXT: [[TMP1:%.*]] = shl i64 [[A0:%.*]], 48
+; CHECK-NEXT: [[TMP2:%.*]] = ashr i64 [[TMP1]], 63
+; CHECK-NEXT: [[TMP3:%.*]] = trunc i64 [[TMP2]] to i8
+; CHECK-NEXT: [[NARROW:%.*]] = add i8 [[TMP3]], 10
; CHECK-NEXT: [[NEG:%.*]] = zext i8 [[NARROW]] to i32
; CHECK-NEXT: ret i32 [[NEG]]
;
@@ -104,9 +101,9 @@ define i32 @sub_sext_mask1_trunc_lshr(i64 %a0) {
define i32 @sub_zext_trunc_lshr(i64 %a0) {
; CHECK-LABEL: @sub_zext_trunc_lshr(
; CHECK-NEXT: [[TMP1:%.*]] = trunc i64 [[A0:%.*]] to i32
-; CHECK-NEXT: [[TMP2:%.*]] = lshr i32 [[TMP1]], 15
-; CHECK-NEXT: [[TMP3:%.*]] = and i32 [[TMP2]], 1
-; CHECK-NEXT: [[NEG:%.*]] = sub nuw nsw i32 10, [[TMP3]]
+; CHECK-NEXT: [[TMP2:%.*]] = shl i32 [[TMP1]], 16
+; CHECK-NEXT: [[TMP3:%.*]] = ashr i32 [[TMP2]], 31
+; CHECK-NEXT: [[NEG:%.*]] = add nsw i32 [[TMP3]], 10
; CHECK-NEXT: ret i32 [[NEG]]
;
%shift = lshr i64 %a0, 15
More information about the llvm-commits
mailing list