[llvm] [InstCombine] Generate better code for `std::bit_floor` from libstdc++ (PR #144842)
Iris Shi via llvm-commits
llvm-commits at lists.llvm.org
Wed Jun 18 22:35:38 PDT 2025
https://github.com/el-ev created https://github.com/llvm/llvm-project/pull/144842
Closes #61183.
This reverts commit 8a1373d3087b1103b402458a3118418d642bb3f5.
Added a mask to the result of ctlz to avoid poison from out-of-range lshr.
Alive2: https://alive2.llvm.org/ce/z/WJiLdq
>From 4072d91fb6d80012839203f8ac1829db75c8a210 Mon Sep 17 00:00:00 2001
From: Iris Shi <0.0 at owo.li>
Date: Thu, 19 Jun 2025 13:02:53 +0800
Subject: [PATCH] [InstCombine] Generate better code for std::bit_floor from
libstdc++
This reverts commit 8a1373d3087b1103b402458a3118418d642bb3f5.
---
.../InstCombine/InstCombineSelect.cpp | 77 +++++++++++++++++++
llvm/test/Transforms/InstCombine/bit_floor.ll | 28 +++----
2 files changed, 89 insertions(+), 16 deletions(-)
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineSelect.cpp b/llvm/lib/Transforms/InstCombine/InstCombineSelect.cpp
index 73ba0f78e8053..72fa225045662 100644
--- a/llvm/lib/Transforms/InstCombine/InstCombineSelect.cpp
+++ b/llvm/lib/Transforms/InstCombine/InstCombineSelect.cpp
@@ -3911,6 +3911,80 @@ static Value *foldSelectBitTest(SelectInst &Sel, Value *CondVal, Value *TrueVal,
return nullptr;
}
+// Transform:
+//
+// 1 << (C - ctlz(X >> 1))
+//
+// into
+//
+// (1 << (C - 1)) >> ctlz(X)
+//
+// The caller must guarantee that X is nonzero.
+//
+// TODO: Relax the requirement that X be nonzero. We just need to require X to
+// be nonzero or the second argument of CTLZ to be true (that is, returning
+// poison on zero).
+static Instruction *foldBitFloorNonzero(Value *N, Value *X,
+ InstCombiner::BuilderTy &Builder) {
+ Type *NType = N->getType();
+ unsigned BitWidth = NType->getScalarSizeInBits();
+
+ // Match C - ctlz(X >> 1), where C is in (0, BitWidth].
+ // TODO: Handle C in [0, BitWidth] (with 0 included in the range), in which
+ // case 1 << C - ctlz(X >> 1) is equivalent to
+ // (1 << ((C - 1) & (BitWidth - 1))) >> ctlz(X).
+ const APInt *C = nullptr;
+ Value *CTLZ;
+ if (!match(N, m_OneUse(m_Shl(m_One(),
+ m_OneUse(m_Sub(m_APInt(C), m_Value(CTLZ)))))) ||
+ !(C->ugt(0) && C->ule(BitWidth)) ||
+ !match(CTLZ, m_OneUse(m_Intrinsic<Intrinsic::ctlz>(
+ m_OneUse(m_LShr(m_Specific(X), m_One())), m_Zero()))))
+ return nullptr;
+
+ APInt ShiftedBit = APInt::getOneBitSet(BitWidth, C->getZExtValue() - 1);
+
+ Value *NewCTLZ =
+ Builder.CreateIntrinsic(Intrinsic::ctlz, {CTLZ->getType()},
+ {X, cast<Instruction>(CTLZ)->getOperand(1)});
+ Value *Masked =
+ Builder.CreateAnd(NewCTLZ, ConstantInt::get(NType, BitWidth - 1));
+ auto *Shift = cast<Instruction>(
+ Builder.CreateLShr(ConstantInt::get(NType, ShiftedBit), Masked));
+ Shift->setIsExact();
+ return Shift;
+}
+
+// Transform:
+//
+// X == 0 ? 0 : (1 << (C1 - ctlz(X >> 1)))
+//
+// into
+//
+// X == 0 ? 0 : (C2 >> ctlz(X))
+//
+// where C2 is computed by foldBitFloorNonzero based on C1. The caller is
+// responsible for replacing one of the select operands.
+static Instruction *foldBitFloor(SelectInst &SI,
+ InstCombiner::BuilderTy &Builder) {
+ Value *TrueVal = SI.getTrueValue();
+ Value *FalseVal = SI.getFalseValue();
+
+ CmpPredicate Pred;
+ Value *Cond0;
+ if (!match(SI.getCondition(), m_ICmp(Pred, m_Value(Cond0), m_Zero())) ||
+ !ICmpInst::isEquality(Pred))
+ return nullptr;
+
+ if (Pred == ICmpInst::ICMP_NE)
+ std::swap(TrueVal, FalseVal);
+
+ if (!match(TrueVal, m_Zero()))
+ return nullptr;
+
+ return foldBitFloorNonzero(FalseVal, Cond0, Builder);
+}
+
Instruction *InstCombinerImpl::visitSelectInst(SelectInst &SI) {
Value *CondVal = SI.getCondition();
Value *TrueVal = SI.getTrueValue();
@@ -4392,6 +4466,9 @@ Instruction *InstCombinerImpl::visitSelectInst(SelectInst &SI) {
if (Instruction *I = foldBitCeil(SI, Builder, *this))
return I;
+ if (Instruction *I = foldBitFloor(SI, Builder))
+ return replaceOperand(SI, match(SI.getTrueValue(), m_Zero()) ? 2 : 1, I);
+
if (Instruction *I = foldSelectToCmp(SI))
return I;
diff --git a/llvm/test/Transforms/InstCombine/bit_floor.ll b/llvm/test/Transforms/InstCombine/bit_floor.ll
index 2872221e8aa87..bce7e3fb45eba 100644
--- a/llvm/test/Transforms/InstCombine/bit_floor.ll
+++ b/llvm/test/Transforms/InstCombine/bit_floor.ll
@@ -4,10 +4,9 @@
define i32 @bit_floor_32(i32 %x) {
; CHECK-LABEL: @bit_floor_32(
; CHECK-NEXT: [[EQ0:%.*]] = icmp eq i32 [[X:%.*]], 0
-; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 [[X]], 1
-; CHECK-NEXT: [[CTLZ:%.*]] = tail call range(i32 1, 33) i32 @llvm.ctlz.i32(i32 [[LSHR]], i1 false)
-; CHECK-NEXT: [[SUB:%.*]] = sub nuw nsw i32 32, [[CTLZ]]
-; CHECK-NEXT: [[SHL:%.*]] = shl nuw i32 1, [[SUB]]
+; CHECK-NEXT: [[TMP1:%.*]] = call range(i32 0, 33) i32 @llvm.ctlz.i32(i32 [[X]], i1 false)
+; CHECK-NEXT: [[TMP2:%.*]] = and i32 [[TMP1]], 31
+; CHECK-NEXT: [[SHL:%.*]] = lshr exact i32 -2147483648, [[TMP2]]
; CHECK-NEXT: [[SEL:%.*]] = select i1 [[EQ0]], i32 0, i32 [[SHL]]
; CHECK-NEXT: ret i32 [[SEL]]
;
@@ -23,10 +22,9 @@ define i32 @bit_floor_32(i32 %x) {
define i64 @bit_floor_64(i64 %x) {
; CHECK-LABEL: @bit_floor_64(
; CHECK-NEXT: [[EQ0:%.*]] = icmp eq i64 [[X:%.*]], 0
-; CHECK-NEXT: [[LSHR:%.*]] = lshr i64 [[X]], 1
-; CHECK-NEXT: [[CTLZ:%.*]] = tail call range(i64 1, 65) i64 @llvm.ctlz.i64(i64 [[LSHR]], i1 false)
-; CHECK-NEXT: [[SUB:%.*]] = sub nuw nsw i64 64, [[CTLZ]]
-; CHECK-NEXT: [[SHL:%.*]] = shl nuw i64 1, [[SUB]]
+; CHECK-NEXT: [[TMP1:%.*]] = call range(i64 0, 65) i64 @llvm.ctlz.i64(i64 [[X]], i1 false)
+; CHECK-NEXT: [[TMP2:%.*]] = and i64 [[TMP1]], 63
+; CHECK-NEXT: [[SHL:%.*]] = lshr exact i64 -9223372036854775808, [[TMP2]]
; CHECK-NEXT: [[SEL:%.*]] = select i1 [[EQ0]], i64 0, i64 [[SHL]]
; CHECK-NEXT: ret i64 [[SEL]]
;
@@ -43,10 +41,9 @@ define i64 @bit_floor_64(i64 %x) {
define i32 @bit_floor_commuted_operands(i32 %x) {
; CHECK-LABEL: @bit_floor_commuted_operands(
; CHECK-NEXT: [[NE0_NOT:%.*]] = icmp eq i32 [[X:%.*]], 0
-; CHECK-NEXT: [[LSHR:%.*]] = lshr i32 [[X]], 1
-; CHECK-NEXT: [[CTLZ:%.*]] = tail call range(i32 1, 33) i32 @llvm.ctlz.i32(i32 [[LSHR]], i1 false)
-; CHECK-NEXT: [[SUB:%.*]] = sub nuw nsw i32 32, [[CTLZ]]
-; CHECK-NEXT: [[SHL:%.*]] = shl nuw i32 1, [[SUB]]
+; CHECK-NEXT: [[TMP1:%.*]] = call range(i32 0, 33) i32 @llvm.ctlz.i32(i32 [[X]], i1 false)
+; CHECK-NEXT: [[TMP2:%.*]] = and i32 [[TMP1]], 31
+; CHECK-NEXT: [[SHL:%.*]] = lshr exact i32 -2147483648, [[TMP2]]
; CHECK-NEXT: [[SEL:%.*]] = select i1 [[NE0_NOT]], i32 0, i32 [[SHL]]
; CHECK-NEXT: ret i32 [[SEL]]
;
@@ -151,10 +148,9 @@ define i32 @bit_floor_shl_used_twice(i32 %x, ptr %p) {
define <4 x i32> @bit_floor_v4i32(<4 x i32> %x) {
; CHECK-LABEL: @bit_floor_v4i32(
; CHECK-NEXT: [[EQ0:%.*]] = icmp eq <4 x i32> [[X:%.*]], zeroinitializer
-; CHECK-NEXT: [[LSHR:%.*]] = lshr <4 x i32> [[X]], splat (i32 1)
-; CHECK-NEXT: [[CTLZ:%.*]] = tail call range(i32 1, 33) <4 x i32> @llvm.ctlz.v4i32(<4 x i32> [[LSHR]], i1 false)
-; CHECK-NEXT: [[SUB:%.*]] = sub nuw nsw <4 x i32> splat (i32 32), [[CTLZ]]
-; CHECK-NEXT: [[SHL:%.*]] = shl nuw <4 x i32> splat (i32 1), [[SUB]]
+; CHECK-NEXT: [[TMP1:%.*]] = call range(i32 0, 33) <4 x i32> @llvm.ctlz.v4i32(<4 x i32> [[X]], i1 false)
+; CHECK-NEXT: [[TMP2:%.*]] = and <4 x i32> [[TMP1]], splat (i32 31)
+; CHECK-NEXT: [[SHL:%.*]] = lshr exact <4 x i32> splat (i32 -2147483648), [[TMP2]]
; CHECK-NEXT: [[SEL:%.*]] = select <4 x i1> [[EQ0]], <4 x i32> zeroinitializer, <4 x i32> [[SHL]]
; CHECK-NEXT: ret <4 x i32> [[SEL]]
;
More information about the llvm-commits
mailing list