[llvm] Combine (X ^ Y) and (X == Y) where appropriate (PR #130922)
Ryan Buchner via llvm-commits
llvm-commits at lists.llvm.org
Wed Mar 12 13:43:03 PDT 2025
https://github.com/bababuck updated https://github.com/llvm/llvm-project/pull/130922
>From cce371d7039358f7a9bd2fef0cc4f79c31cd4158 Mon Sep 17 00:00:00 2001
From: bababuck <buchner.ryan at gmail.com>
Date: Wed, 12 Mar 2025 12:54:37 -0700
Subject: [PATCH 1/4] Add test for folding sign extensions with comparison's to
0 in RISCV
---
.../test/CodeGen/RISCV/select-constant-xor.ll | 37 +++++++++++++++++++
1 file changed, 37 insertions(+)
diff --git a/llvm/test/CodeGen/RISCV/select-constant-xor.ll b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
index 2e26ae78e2dd8..938eda58637e6 100644
--- a/llvm/test/CodeGen/RISCV/select-constant-xor.ll
+++ b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
@@ -239,3 +239,40 @@ define i32 @oneusecmp(i32 %a, i32 %b, i32 %d) {
%x = add i32 %s, %s2
ret i32 %x
}
+
+define i32 @branch_sext_ret(i32 %x) {
+; RV32-LABEL: branch_sext_ret:
+; RV32: # %bb.0: # %entry
+; RV32-NEXT: beqz a0, .LBB11_2
+; RV32-NEXT: # %bb.1: # %if.then
+; RV32-NEXT: ret
+; RV32-NEXT: .LBB11_2: # %if.end
+; RV32-NEXT: addi sp, sp, -16
+; RV32-NEXT: .cfi_def_cfa_offset 16
+; RV32-NEXT: sw ra, 12(sp) # 4-byte Folded Spill
+; RV32-NEXT: .cfi_offset ra, -4
+; RV32-NEXT: call abort
+;
+; RV64-LABEL: branch_sext_ret:
+; RV64: # %bb.0: # %entry
+; RV64-NEXT: sext.w a1, a0
+; RV64-NEXT: beqz a1, .LBB11_2
+; RV64-NEXT: # %bb.1: # %if.then
+; RV64-NEXT: ret
+; RV64-NEXT: .LBB11_2: # %if.end
+; RV64-NEXT: addi sp, sp, -16
+; RV64-NEXT: .cfi_def_cfa_offset 16
+; RV64-NEXT: sd ra, 8(sp) # 8-byte Folded Spill
+; RV64-NEXT: .cfi_offset ra, -8
+; RV64-NEXT: call abort
+entry:
+ %cmp.not = icmp eq i32 %x, 0
+ br i1 %cmp.not, label %if.end, label %if.then
+if.then:
+ ret i32 %x
+if.end:
+ tail call void @abort() #2
+ unreachable
+}
+
+declare void @abort()
>From e93170897d25c26f7f941b0e9811d675c4bb7a20 Mon Sep 17 00:00:00 2001
From: bababuck <buchner.ryan at gmail.com>
Date: Wed, 12 Mar 2025 12:44:41 -0700
Subject: [PATCH 2/4] Fold RISCV sign extension operations is their result will
compared with zero.
sext_ext(0) == 0.
---
llvm/lib/Target/RISCV/RISCVISelLowering.cpp | 6 ++++++
llvm/test/CodeGen/RISCV/select-constant-xor.ll | 3 +--
2 files changed, 7 insertions(+), 2 deletions(-)
diff --git a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
index 27a4bbce1f5fc..d0c0db929a081 100644
--- a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
+++ b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
@@ -17201,6 +17201,12 @@ static bool combine_CC(SDValue &LHS, SDValue &RHS, SDValue &CC, const SDLoc &DL,
return true;
}
+ // Fold ((sext X), 0, eq/ne) -> (X, 0, eq/ne)
+ if (LHS.getOpcode() == ISD::SIGN_EXTEND_INREG && isNullConstant(RHS)) {
+ LHS = LHS.getOperand(0);
+ return true;
+ }
+
// Fold ((srl (and X, 1<<C), C), 0, eq/ne) -> ((shl X, XLen-1-C), 0, ge/lt)
if (isNullConstant(RHS) && LHS.getOpcode() == ISD::SRL && LHS.hasOneUse() &&
LHS.getOperand(1).getOpcode() == ISD::Constant) {
diff --git a/llvm/test/CodeGen/RISCV/select-constant-xor.ll b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
index 938eda58637e6..fc8a505b264ef 100644
--- a/llvm/test/CodeGen/RISCV/select-constant-xor.ll
+++ b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
@@ -255,8 +255,7 @@ define i32 @branch_sext_ret(i32 %x) {
;
; RV64-LABEL: branch_sext_ret:
; RV64: # %bb.0: # %entry
-; RV64-NEXT: sext.w a1, a0
-; RV64-NEXT: beqz a1, .LBB11_2
+; RV64-NEXT: beqz a0, .LBB11_2
; RV64-NEXT: # %bb.1: # %if.then
; RV64-NEXT: ret
; RV64-NEXT: .LBB11_2: # %if.end
>From 7d6a77276dc4334d75c9f7027ef18792cc1d7ed3 Mon Sep 17 00:00:00 2001
From: bababuck <buchner.ryan at gmail.com>
Date: Wed, 12 Mar 2025 09:23:06 -0700
Subject: [PATCH 3/4] Add new test that re-produces issue #130510
---
.../test/CodeGen/RISCV/select-constant-xor.ll | 82 +++++++++++++++++++
1 file changed, 82 insertions(+)
diff --git a/llvm/test/CodeGen/RISCV/select-constant-xor.ll b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
index fc8a505b264ef..3336cc6b96ad4 100644
--- a/llvm/test/CodeGen/RISCV/select-constant-xor.ll
+++ b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
@@ -274,4 +274,86 @@ if.end:
unreachable
}
+define i32 @xor_branch_imm_ret(i32 %x) {
+; RV32-LABEL: xor_branch_imm_ret:
+; RV32: # %bb.0: # %entry
+; RV32-NEXT: li a1, -1365
+; RV32-NEXT: beq a0, a1, .LBB12_2
+; RV32-NEXT: # %bb.1: # %if.then
+; RV32-NEXT: xori a0, a0, -1365
+; RV32-NEXT: ret
+; RV32-NEXT: .LBB12_2: # %if.end
+; RV32-NEXT: addi sp, sp, -16
+; RV32-NEXT: .cfi_def_cfa_offset 16
+; RV32-NEXT: sw ra, 12(sp) # 4-byte Folded Spill
+; RV32-NEXT: .cfi_offset ra, -4
+; RV32-NEXT: call abort
+;
+; RV64-LABEL: xor_branch_imm_ret:
+; RV64: # %bb.0: # %entry
+; RV64-NEXT: sext.w a1, a0
+; RV64-NEXT: li a2, -1365
+; RV64-NEXT: beq a1, a2, .LBB12_2
+; RV64-NEXT: # %bb.1: # %if.then
+; RV64-NEXT: xori a0, a0, -1365
+; RV64-NEXT: ret
+; RV64-NEXT: .LBB12_2: # %if.end
+; RV64-NEXT: addi sp, sp, -16
+; RV64-NEXT: .cfi_def_cfa_offset 16
+; RV64-NEXT: sd ra, 8(sp) # 8-byte Folded Spill
+; RV64-NEXT: .cfi_offset ra, -8
+; RV64-NEXT: call abort
+entry:
+ %cmp.not = icmp eq i32 %x, -1365
+ br i1 %cmp.not, label %if.end, label %if.then
+if.then:
+ %xor = xor i32 %x, -1365
+ ret i32 %xor
+if.end:
+ tail call void @abort() #2
+ unreachable
+}
+
+define i32 @xor_branch_ret(i32 %x) {
+; RV32-LABEL: xor_branch_ret:
+; RV32: # %bb.0: # %entry
+; RV32-NEXT: li a1, 1
+; RV32-NEXT: slli a1, a1, 11
+; RV32-NEXT: beq a0, a1, .LBB13_2
+; RV32-NEXT: # %bb.1: # %if.then
+; RV32-NEXT: xor a0, a0, a1
+; RV32-NEXT: ret
+; RV32-NEXT: .LBB13_2: # %if.end
+; RV32-NEXT: addi sp, sp, -16
+; RV32-NEXT: .cfi_def_cfa_offset 16
+; RV32-NEXT: sw ra, 12(sp) # 4-byte Folded Spill
+; RV32-NEXT: .cfi_offset ra, -4
+; RV32-NEXT: call abort
+;
+; RV64-LABEL: xor_branch_ret:
+; RV64: # %bb.0: # %entry
+; RV64-NEXT: sext.w a2, a0
+; RV64-NEXT: li a1, 1
+; RV64-NEXT: slli a1, a1, 11
+; RV64-NEXT: beq a2, a1, .LBB13_2
+; RV64-NEXT: # %bb.1: # %if.then
+; RV64-NEXT: xor a0, a0, a1
+; RV64-NEXT: ret
+; RV64-NEXT: .LBB13_2: # %if.end
+; RV64-NEXT: addi sp, sp, -16
+; RV64-NEXT: .cfi_def_cfa_offset 16
+; RV64-NEXT: sd ra, 8(sp) # 8-byte Folded Spill
+; RV64-NEXT: .cfi_offset ra, -8
+; RV64-NEXT: call abort
+entry:
+ %cmp.not = icmp eq i32 %x, 2048
+ br i1 %cmp.not, label %if.end, label %if.then
+if.then:
+ %xor = xor i32 %x, 2048
+ ret i32 %xor
+if.end:
+ tail call void @abort() #2
+ unreachable
+}
+
declare void @abort()
>From aa0d2fb1069745dc72da059250fa5f42697f3727 Mon Sep 17 00:00:00 2001
From: bababuck <buchner.ryan at gmail.com>
Date: Tue, 11 Mar 2025 09:47:21 -0700
Subject: [PATCH 4/4] Combine (X ^ Y) and (X == Y) where appropriate
In RISCV, modify the folding of (X ^ Y == 0) -> (X == Y) to account for
cases where the (X ^ Y) will be re-used.
Fixes #130510.
---
llvm/lib/CodeGen/CodeGenPrepare.cpp | 3 +-
llvm/lib/Target/RISCV/RISCVISelLowering.cpp | 35 ++++++++++++++++++-
.../test/CodeGen/RISCV/select-constant-xor.ll | 14 +++-----
3 files changed, 41 insertions(+), 11 deletions(-)
diff --git a/llvm/lib/CodeGen/CodeGenPrepare.cpp b/llvm/lib/CodeGen/CodeGenPrepare.cpp
index d5fbd4c380746..2acb7cb321d07 100644
--- a/llvm/lib/CodeGen/CodeGenPrepare.cpp
+++ b/llvm/lib/CodeGen/CodeGenPrepare.cpp
@@ -8578,7 +8578,8 @@ static bool optimizeBranch(BranchInst *Branch, const TargetLowering &TLI,
}
if (Cmp->isEquality() &&
(match(UI, m_Add(m_Specific(X), m_SpecificInt(-CmpC))) ||
- match(UI, m_Sub(m_Specific(X), m_SpecificInt(CmpC))))) {
+ match(UI, m_Sub(m_Specific(X), m_SpecificInt(CmpC))) ||
+ match(UI, m_Xor(m_Specific(X), m_SpecificInt(CmpC))))) {
IRBuilder<> Builder(Branch);
if (UI->getParent() != Branch->getParent())
UI->moveBefore(Branch->getIterator());
diff --git a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
index d0c0db929a081..5570b7ee6ae14 100644
--- a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
+++ b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
@@ -17194,8 +17194,41 @@ static bool combine_CC(SDValue &LHS, SDValue &RHS, SDValue &CC, const SDLoc &DL,
return true;
}
+ // If XOR is reused and has an immediate that will fit in XORI,
+ // do not fold
+ auto IsXorImmediate = [](const SDValue &Op) -> bool {
+ if (const auto XorCnst = dyn_cast<ConstantSDNode>(Op)) {
+ auto isLegalXorImmediate = [](int64_t Imm) -> bool {
+ return isInt<12>(Imm);
+ };
+ return isLegalXorImmediate(XorCnst->getSExtValue());
+ }
+ return false;
+ };
+ // Fold (X(i1) ^ 1) == 0 -> X != 0
+ auto SingleBitOp = [&DAG](const SDValue &VarOp,
+ const SDValue &ConstOp) -> bool {
+ if (const auto XorCnst = dyn_cast<ConstantSDNode>(ConstOp)) {
+ const APInt Mask = APInt::getBitsSetFrom(VarOp.getValueSizeInBits(), 1);
+ return (XorCnst->getSExtValue() == 1) && DAG.MaskedValueIsZero(VarOp, Mask);
+ }
+ return false;
+ };
+ auto OnlyUsedBySelectOrBR = [](const SDValue &Op) -> bool {
+ for (const SDNode *UserNode : Op->users()) {
+ const unsigned Opcode = UserNode->getOpcode();
+ if (Opcode != RISCVISD::SELECT_CC && Opcode != RISCVISD::BR_CC) {
+ return false;
+ }
+ }
+ return true;
+ };
+
// Fold ((xor X, Y), 0, eq/ne) -> (X, Y, eq/ne)
- if (LHS.getOpcode() == ISD::XOR && isNullConstant(RHS)) {
+ if (LHS.getOpcode() == ISD::XOR && isNullConstant(RHS) &&
+ (!IsXorImmediate(LHS.getOperand(1)) ||
+ SingleBitOp(LHS.getOperand(0), LHS.getOperand(1)) ||
+ OnlyUsedBySelectOrBR(LHS))) {
RHS = LHS.getOperand(1);
LHS = LHS.getOperand(0);
return true;
diff --git a/llvm/test/CodeGen/RISCV/select-constant-xor.ll b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
index 3336cc6b96ad4..b85686090182c 100644
--- a/llvm/test/CodeGen/RISCV/select-constant-xor.ll
+++ b/llvm/test/CodeGen/RISCV/select-constant-xor.ll
@@ -277,10 +277,9 @@ if.end:
define i32 @xor_branch_imm_ret(i32 %x) {
; RV32-LABEL: xor_branch_imm_ret:
; RV32: # %bb.0: # %entry
-; RV32-NEXT: li a1, -1365
-; RV32-NEXT: beq a0, a1, .LBB12_2
-; RV32-NEXT: # %bb.1: # %if.then
; RV32-NEXT: xori a0, a0, -1365
+; RV32-NEXT: beqz a0, .LBB12_2
+; RV32-NEXT: # %bb.1: # %if.then
; RV32-NEXT: ret
; RV32-NEXT: .LBB12_2: # %if.end
; RV32-NEXT: addi sp, sp, -16
@@ -291,11 +290,9 @@ define i32 @xor_branch_imm_ret(i32 %x) {
;
; RV64-LABEL: xor_branch_imm_ret:
; RV64: # %bb.0: # %entry
-; RV64-NEXT: sext.w a1, a0
-; RV64-NEXT: li a2, -1365
-; RV64-NEXT: beq a1, a2, .LBB12_2
-; RV64-NEXT: # %bb.1: # %if.then
; RV64-NEXT: xori a0, a0, -1365
+; RV64-NEXT: beqz a0, .LBB12_2
+; RV64-NEXT: # %bb.1: # %if.then
; RV64-NEXT: ret
; RV64-NEXT: .LBB12_2: # %if.end
; RV64-NEXT: addi sp, sp, -16
@@ -332,10 +329,9 @@ define i32 @xor_branch_ret(i32 %x) {
;
; RV64-LABEL: xor_branch_ret:
; RV64: # %bb.0: # %entry
-; RV64-NEXT: sext.w a2, a0
; RV64-NEXT: li a1, 1
; RV64-NEXT: slli a1, a1, 11
-; RV64-NEXT: beq a2, a1, .LBB13_2
+; RV64-NEXT: beq a0, a1, .LBB13_2
; RV64-NEXT: # %bb.1: # %if.then
; RV64-NEXT: xor a0, a0, a1
; RV64-NEXT: ret
More information about the llvm-commits
mailing list