[llvm] [RISCV] Combine (and (select cond, x, -1), c) to (select cond, x, (and x, c)) with Zicond. (PR #69563)
via llvm-commits
llvm-commits at lists.llvm.org
Wed Oct 18 22:41:41 PDT 2023
- Previous message: [llvm] [RISCV] Combine (and (select cond, x, -1), c) to (select cond, x, (and x, c)) with Zicond. (PR #69563)
- Next message: [llvm] [RISCV] Combine (and (select cond, x, -1), c) to (select cond, x, (and x, c)) with Zicond. (PR #69563)
- Messages sorted by:
[ date ]
[ thread ]
[ subject ]
[ author ]
llvmbot wrote:
<!--LLVM PR SUMMARY COMMENT-->
@llvm/pr-subscribers-backend-risc-v
Author: Yeting Kuo (yetingk)
<details>
<summary>Changes</summary>
It's only beneficial when cond is setcc with integer equality condition code. For other case, it has same instruction count as the original.
---
Full diff: https://github.com/llvm/llvm-project/pull/69563.diff
2 Files Affected:
- (modified) llvm/lib/Target/RISCV/RISCVISelLowering.cpp (+17-2)
- (modified) llvm/test/CodeGen/RISCV/select-binop-identity.ll (+323-74)
``````````diff
diff --git a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
index e8f001e491cdcaa..40f845c8dcb9265 100644
--- a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
+++ b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
@@ -11584,8 +11584,23 @@ static SDValue combineSelectAndUse(SDNode *N, SDValue Slct, SDValue OtherOp,
if (VT.isVector())
return SDValue();
- if (!Subtarget.hasShortForwardBranchOpt() ||
- (Slct.getOpcode() != ISD::SELECT &&
+ if (!Subtarget.hasShortForwardBranchOpt()) {
+ // (select cond, x, (and x, c)) has custom lowering with Zicond.
+ if ((!Subtarget.hasStdExtZicond() &&
+ !Subtarget.hasVendorXVentanaCondOps()) ||
+ N->getOpcode() != ISD::AND)
+ return SDValue();
+
+ // Maybe harmful when condition code has multip use.
+ if (Slct.getOpcode() == ISD::SELECT && !Slct.getOperand(0).hasOneUse())
+ return SDValue();
+
+ // Maybe harmful when VT is wider than XLen.
+ if (VT.getSizeInBits() > Subtarget.getXLen())
+ return SDValue();
+ }
+
+ if ((Slct.getOpcode() != ISD::SELECT &&
Slct.getOpcode() != RISCVISD::SELECT_CC) ||
!Slct.hasOneUse())
return SDValue();
diff --git a/llvm/test/CodeGen/RISCV/select-binop-identity.ll b/llvm/test/CodeGen/RISCV/select-binop-identity.ll
index 61344bc8979abb7..4c0af2956a0161e 100644
--- a/llvm/test/CodeGen/RISCV/select-binop-identity.ll
+++ b/llvm/test/CodeGen/RISCV/select-binop-identity.ll
@@ -6,11 +6,11 @@
; RUN: llc -mtriple=riscv64 -mcpu=sifive-u74 -verify-machineinstrs < %s \
; RUN: | FileCheck -check-prefix=SFB64 %s
; RUN: llc -mtriple=riscv64 -mattr=+xventanacondops -verify-machineinstrs < %s \
-; RUN: | FileCheck -check-prefixes=RV64,VTCONDOPS64 %s
+; RUN: | FileCheck -check-prefixes=VTCONDOPS64 %s
; RUN: llc -mtriple=riscv32 -mattr=+experimental-zicond -verify-machineinstrs < %s \
; RUN: | FileCheck -check-prefixes=RV32,ZICOND,ZICOND32 %s
; RUN: llc -mtriple=riscv64 -mattr=+experimental-zicond -verify-machineinstrs < %s \
-; RUN: | FileCheck -check-prefixes=RV64,ZICOND,ZICOND64 %s
+; RUN: | FileCheck -check-prefixes=ZICOND,ZICOND64 %s
; InstCombine canonicalizes (c ? x | y : x) to (x | (c ? y : 0)) similar for
; other binary operations using their identity value as the constant.
@@ -19,12 +19,12 @@
; the basic block we create when we expand select.
define signext i32 @and_select_all_ones_i32(i1 zeroext %c, i32 signext %x, i32 signext %y) {
-; RV32-LABEL: and_select_all_ones_i32:
-; RV32: # %bb.0:
-; RV32-NEXT: addi a0, a0, -1
-; RV32-NEXT: or a0, a0, a1
-; RV32-NEXT: and a0, a0, a2
-; RV32-NEXT: ret
+; RV32I-LABEL: and_select_all_ones_i32:
+; RV32I: # %bb.0:
+; RV32I-NEXT: addi a0, a0, -1
+; RV32I-NEXT: or a0, a0, a1
+; RV32I-NEXT: and a0, a0, a2
+; RV32I-NEXT: ret
;
; RV64-LABEL: and_select_all_ones_i32:
; RV64: # %bb.0:
@@ -41,6 +41,118 @@ define signext i32 @and_select_all_ones_i32(i1 zeroext %c, i32 signext %x, i32 s
; SFB64-NEXT: .LBB0_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i32:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: vt.maskcn a0, a2, a0
+; VTCONDOPS64-NEXT: and a1, a2, a1
+; VTCONDOPS64-NEXT: or a0, a1, a0
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND-LABEL: and_select_all_ones_i32:
+; ZICOND: # %bb.0:
+; ZICOND-NEXT: czero.nez a0, a2, a0
+; ZICOND-NEXT: and a1, a2, a1
+; ZICOND-NEXT: or a0, a1, a0
+; ZICOND-NEXT: ret
+ %a = select i1 %c, i32 %x, i32 -1
+ %b = and i32 %a, %y
+ ret i32 %b
+}
+
+define signext i32 @and_select_all_ones_i32_cmp(i32 signext %x, i32 signext %y, i32 signext %z) {
+; RV32I-LABEL: and_select_all_ones_i32_cmp:
+; RV32I: # %bb.0:
+; RV32I-NEXT: addi a2, a2, -4
+; RV32I-NEXT: seqz a2, a2
+; RV32I-NEXT: addi a2, a2, -1
+; RV32I-NEXT: or a0, a2, a0
+; RV32I-NEXT: and a0, a0, a1
+; RV32I-NEXT: ret
+;
+; RV64-LABEL: and_select_all_ones_i32_cmp:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a2, a2, -4
+; RV64-NEXT: seqz a2, a2
+; RV64-NEXT: addi a2, a2, -1
+; RV64-NEXT: or a0, a2, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: ret
+;
+; SFB64-LABEL: and_select_all_ones_i32_cmp:
+; SFB64: # %bb.0:
+; SFB64-NEXT: li a3, 4
+; SFB64-NEXT: bne a2, a3, .LBB1_2
+; SFB64-NEXT: # %bb.1:
+; SFB64-NEXT: and a1, a1, a0
+; SFB64-NEXT: .LBB1_2:
+; SFB64-NEXT: mv a0, a1
+; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i32_cmp:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: addi a2, a2, -4
+; VTCONDOPS64-NEXT: and a0, a1, a0
+; VTCONDOPS64-NEXT: vt.maskc a1, a1, a2
+; VTCONDOPS64-NEXT: or a0, a0, a1
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND-LABEL: and_select_all_ones_i32_cmp:
+; ZICOND: # %bb.0:
+; ZICOND-NEXT: addi a2, a2, -4
+; ZICOND-NEXT: and a0, a1, a0
+; ZICOND-NEXT: czero.eqz a1, a1, a2
+; ZICOND-NEXT: or a0, a0, a1
+; ZICOND-NEXT: ret
+ %c = icmp eq i32 %z, 4
+ %a = select i1 %c, i32 %x, i32 -1
+ %b = and i32 %a, %y
+ ret i32 %b
+}
+
+define signext i32 @and_select_all_ones_i32_cmp2(i32 signext %x, i32 signext %y, i32 signext %z) {
+; RV32I-LABEL: and_select_all_ones_i32_cmp2:
+; RV32I: # %bb.0:
+; RV32I-NEXT: slti a2, a2, 4
+; RV32I-NEXT: addi a2, a2, -1
+; RV32I-NEXT: or a0, a2, a0
+; RV32I-NEXT: and a0, a0, a1
+; RV32I-NEXT: ret
+;
+; RV64-LABEL: and_select_all_ones_i32_cmp2:
+; RV64: # %bb.0:
+; RV64-NEXT: slti a2, a2, 4
+; RV64-NEXT: addi a2, a2, -1
+; RV64-NEXT: or a0, a2, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: ret
+;
+; SFB64-LABEL: and_select_all_ones_i32_cmp2:
+; SFB64: # %bb.0:
+; SFB64-NEXT: li a3, 4
+; SFB64-NEXT: bge a2, a3, .LBB2_2
+; SFB64-NEXT: # %bb.1:
+; SFB64-NEXT: and a1, a1, a0
+; SFB64-NEXT: .LBB2_2:
+; SFB64-NEXT: mv a0, a1
+; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i32_cmp2:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: slti a2, a2, 4
+; VTCONDOPS64-NEXT: and a0, a1, a0
+; VTCONDOPS64-NEXT: vt.maskcn a1, a1, a2
+; VTCONDOPS64-NEXT: or a0, a0, a1
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND-LABEL: and_select_all_ones_i32_cmp2:
+; ZICOND: # %bb.0:
+; ZICOND-NEXT: slti a2, a2, 4
+; ZICOND-NEXT: and a0, a1, a0
+; ZICOND-NEXT: czero.nez a1, a1, a2
+; ZICOND-NEXT: or a0, a0, a1
+; ZICOND-NEXT: ret
+ %c = icmp slt i32 %z, 4
%a = select i1 %c, i32 %x, i32 -1
%b = and i32 %a, %y
ret i32 %b
@@ -65,17 +177,154 @@ define i64 @and_select_all_ones_i64(i1 zeroext %c, i64 %x, i64 %y) {
;
; SFB64-LABEL: and_select_all_ones_i64:
; SFB64: # %bb.0:
-; SFB64-NEXT: bnez a0, .LBB1_2
+; SFB64-NEXT: bnez a0, .LBB3_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: and a2, a2, a1
-; SFB64-NEXT: .LBB1_2:
+; SFB64-NEXT: .LBB3_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i64:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: vt.maskc a0, a2, a0
+; VTCONDOPS64-NEXT: and a1, a2, a1
+; VTCONDOPS64-NEXT: or a0, a1, a0
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND64-LABEL: and_select_all_ones_i64:
+; ZICOND64: # %bb.0:
+; ZICOND64-NEXT: czero.eqz a0, a2, a0
+; ZICOND64-NEXT: and a1, a2, a1
+; ZICOND64-NEXT: or a0, a1, a0
+; ZICOND64-NEXT: ret
%a = select i1 %c, i64 -1, i64 %x
%b = and i64 %y, %a
ret i64 %b
}
+define i64 @and_select_all_ones_i64_cmp(i64 %x, i64 %y, i64 %z) {
+; RV32-LABEL: and_select_all_ones_i64_cmp:
+; RV32: # %bb.0:
+; RV32-NEXT: xori a4, a4, 4
+; RV32-NEXT: or a4, a4, a5
+; RV32-NEXT: seqz a4, a4
+; RV32-NEXT: addi a4, a4, -1
+; RV32-NEXT: or a1, a4, a1
+; RV32-NEXT: or a0, a4, a0
+; RV32-NEXT: and a0, a0, a2
+; RV32-NEXT: and a1, a1, a3
+; RV32-NEXT: ret
+;
+; RV64-LABEL: and_select_all_ones_i64_cmp:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a2, a2, -4
+; RV64-NEXT: seqz a2, a2
+; RV64-NEXT: addi a2, a2, -1
+; RV64-NEXT: or a0, a2, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: ret
+;
+; SFB64-LABEL: and_select_all_ones_i64_cmp:
+; SFB64: # %bb.0:
+; SFB64-NEXT: li a3, 4
+; SFB64-NEXT: bne a2, a3, .LBB4_2
+; SFB64-NEXT: # %bb.1:
+; SFB64-NEXT: and a1, a1, a0
+; SFB64-NEXT: .LBB4_2:
+; SFB64-NEXT: mv a0, a1
+; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i64_cmp:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: addi a2, a2, -4
+; VTCONDOPS64-NEXT: and a0, a1, a0
+; VTCONDOPS64-NEXT: vt.maskc a1, a1, a2
+; VTCONDOPS64-NEXT: or a0, a0, a1
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND64-LABEL: and_select_all_ones_i64_cmp:
+; ZICOND64: # %bb.0:
+; ZICOND64-NEXT: addi a2, a2, -4
+; ZICOND64-NEXT: and a0, a1, a0
+; ZICOND64-NEXT: czero.eqz a1, a1, a2
+; ZICOND64-NEXT: or a0, a0, a1
+; ZICOND64-NEXT: ret
+ %c = icmp eq i64 %z, 4
+ %a = select i1 %c, i64 %x, i64 -1
+ %b = and i64 %a, %y
+ ret i64 %b
+}
+
+define i64 @and_select_all_ones_i64_cmp2(i64 %x, i64 %y, i64 %z) {
+; RV32I-LABEL: and_select_all_ones_i64_cmp2:
+; RV32I: # %bb.0:
+; RV32I-NEXT: beqz a5, .LBB5_2
+; RV32I-NEXT: # %bb.1:
+; RV32I-NEXT: slti a4, a5, 0
+; RV32I-NEXT: j .LBB5_3
+; RV32I-NEXT: .LBB5_2:
+; RV32I-NEXT: sltiu a4, a4, 4
+; RV32I-NEXT: .LBB5_3:
+; RV32I-NEXT: addi a4, a4, -1
+; RV32I-NEXT: or a1, a4, a1
+; RV32I-NEXT: or a0, a4, a0
+; RV32I-NEXT: and a0, a0, a2
+; RV32I-NEXT: and a1, a1, a3
+; RV32I-NEXT: ret
+;
+; RV64-LABEL: and_select_all_ones_i64_cmp2:
+; RV64: # %bb.0:
+; RV64-NEXT: slti a2, a2, 4
+; RV64-NEXT: addi a2, a2, -1
+; RV64-NEXT: or a0, a2, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: ret
+;
+; SFB64-LABEL: and_select_all_ones_i64_cmp2:
+; SFB64: # %bb.0:
+; SFB64-NEXT: li a3, 4
+; SFB64-NEXT: bge a2, a3, .LBB5_2
+; SFB64-NEXT: # %bb.1:
+; SFB64-NEXT: and a1, a1, a0
+; SFB64-NEXT: .LBB5_2:
+; SFB64-NEXT: mv a0, a1
+; SFB64-NEXT: ret
+;
+; VTCONDOPS64-LABEL: and_select_all_ones_i64_cmp2:
+; VTCONDOPS64: # %bb.0:
+; VTCONDOPS64-NEXT: slti a2, a2, 4
+; VTCONDOPS64-NEXT: and a0, a1, a0
+; VTCONDOPS64-NEXT: vt.maskcn a1, a1, a2
+; VTCONDOPS64-NEXT: or a0, a0, a1
+; VTCONDOPS64-NEXT: ret
+;
+; ZICOND32-LABEL: and_select_all_ones_i64_cmp2:
+; ZICOND32: # %bb.0:
+; ZICOND32-NEXT: slti a6, a5, 0
+; ZICOND32-NEXT: czero.eqz a6, a6, a5
+; ZICOND32-NEXT: sltiu a4, a4, 4
+; ZICOND32-NEXT: czero.nez a4, a4, a5
+; ZICOND32-NEXT: or a4, a4, a6
+; ZICOND32-NEXT: addi a4, a4, -1
+; ZICOND32-NEXT: or a1, a4, a1
+; ZICOND32-NEXT: or a0, a4, a0
+; ZICOND32-NEXT: and a0, a0, a2
+; ZICOND32-NEXT: and a1, a1, a3
+; ZICOND32-NEXT: ret
+;
+; ZICOND64-LABEL: and_select_all_ones_i64_cmp2:
+; ZICOND64: # %bb.0:
+; ZICOND64-NEXT: slti a2, a2, 4
+; ZICOND64-NEXT: and a0, a1, a0
+; ZICOND64-NEXT: czero.nez a1, a1, a2
+; ZICOND64-NEXT: or a0, a0, a1
+; ZICOND64-NEXT: ret
+ %c = icmp slt i64 %z, 4
+ %a = select i1 %c, i64 %x, i64 -1
+ %b = and i64 %a, %y
+ ret i64 %b
+}
+
define signext i32 @or_select_all_zeros_i32(i1 zeroext %c, i32 signext %x, i32 signext %y) {
; RV32I-LABEL: or_select_all_zeros_i32:
; RV32I: # %bb.0:
@@ -84,19 +333,19 @@ define signext i32 @or_select_all_zeros_i32(i1 zeroext %c, i32 signext %x, i32 s
; RV32I-NEXT: or a0, a2, a0
; RV32I-NEXT: ret
;
-; RV64I-LABEL: or_select_all_zeros_i32:
-; RV64I: # %bb.0:
-; RV64I-NEXT: neg a0, a0
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: or a0, a2, a0
-; RV64I-NEXT: ret
+; RV64-LABEL: or_select_all_zeros_i32:
+; RV64: # %bb.0:
+; RV64-NEXT: neg a0, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: or a0, a2, a0
+; RV64-NEXT: ret
;
; SFB64-LABEL: or_select_all_zeros_i32:
; SFB64: # %bb.0:
-; SFB64-NEXT: beqz a0, .LBB2_2
+; SFB64-NEXT: beqz a0, .LBB6_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: or a2, a2, a1
-; SFB64-NEXT: .LBB2_2:
+; SFB64-NEXT: .LBB6_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -126,19 +375,19 @@ define i64 @or_select_all_zeros_i64(i1 zeroext %c, i64 %x, i64 %y) {
; RV32I-NEXT: or a1, a2, a4
; RV32I-NEXT: ret
;
-; RV64I-LABEL: or_select_all_zeros_i64:
-; RV64I: # %bb.0:
-; RV64I-NEXT: addi a0, a0, -1
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: or a0, a0, a2
-; RV64I-NEXT: ret
+; RV64-LABEL: or_select_all_zeros_i64:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a0, a0, -1
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: or a0, a0, a2
+; RV64-NEXT: ret
;
; SFB64-LABEL: or_select_all_zeros_i64:
; SFB64: # %bb.0:
-; SFB64-NEXT: bnez a0, .LBB3_2
+; SFB64-NEXT: bnez a0, .LBB7_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: or a2, a2, a1
-; SFB64-NEXT: .LBB3_2:
+; SFB64-NEXT: .LBB7_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -174,19 +423,19 @@ define signext i32 @xor_select_all_zeros_i32(i1 zeroext %c, i32 signext %x, i32
; RV32I-NEXT: xor a0, a2, a0
; RV32I-NEXT: ret
;
-; RV64I-LABEL: xor_select_all_zeros_i32:
-; RV64I: # %bb.0:
-; RV64I-NEXT: addi a0, a0, -1
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: xor a0, a2, a0
-; RV64I-NEXT: ret
+; RV64-LABEL: xor_select_all_zeros_i32:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a0, a0, -1
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: xor a0, a2, a0
+; RV64-NEXT: ret
;
; SFB64-LABEL: xor_select_all_zeros_i32:
; SFB64: # %bb.0:
-; SFB64-NEXT: bnez a0, .LBB4_2
+; SFB64-NEXT: bnez a0, .LBB8_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: xor a2, a2, a1
-; SFB64-NEXT: .LBB4_2:
+; SFB64-NEXT: .LBB8_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -216,19 +465,19 @@ define i64 @xor_select_all_zeros_i64(i1 zeroext %c, i64 %x, i64 %y) {
; RV32I-NEXT: xor a1, a2, a4
; RV32I-NEXT: ret
;
-; RV64I-LABEL: xor_select_all_zeros_i64:
-; RV64I: # %bb.0:
-; RV64I-NEXT: neg a0, a0
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: xor a0, a0, a2
-; RV64I-NEXT: ret
+; RV64-LABEL: xor_select_all_zeros_i64:
+; RV64: # %bb.0:
+; RV64-NEXT: neg a0, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: xor a0, a0, a2
+; RV64-NEXT: ret
;
; SFB64-LABEL: xor_select_all_zeros_i64:
; SFB64: # %bb.0:
-; SFB64-NEXT: beqz a0, .LBB5_2
+; SFB64-NEXT: beqz a0, .LBB9_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: xor a2, a2, a1
-; SFB64-NEXT: .LBB5_2:
+; SFB64-NEXT: .LBB9_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -264,19 +513,19 @@ define signext i32 @add_select_all_zeros_i32(i1 zeroext %c, i32 signext %x, i32
; RV32I-NEXT: add a0, a2, a0
; RV32I-NEXT: ret
;
-; RV64I-LABEL: add_select_all_zeros_i32:
-; RV64I: # %bb.0:
-; RV64I-NEXT: addi a0, a0, -1
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: addw a0, a2, a0
-; RV64I-NEXT: ret
+; RV64-LABEL: add_select_all_zeros_i32:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a0, a0, -1
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: addw a0, a2, a0
+; RV64-NEXT: ret
;
; SFB64-LABEL: add_select_all_zeros_i32:
; SFB64: # %bb.0:
-; SFB64-NEXT: bnez a0, .LBB6_2
+; SFB64-NEXT: bnez a0, .LBB10_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: addw a2, a2, a1
-; SFB64-NEXT: .LBB6_2:
+; SFB64-NEXT: .LBB10_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -314,19 +563,19 @@ define i64 @add_select_all_zeros_i64(i1 zeroext %c, i64 %x, i64 %y) {
; RV32I-NEXT: add a1, a2, a1
; RV32I-NEXT: ret
;
-; RV64I-LABEL: add_select_all_zeros_i64:
-; RV64I: # %bb.0:
-; RV64I-NEXT: neg a0, a0
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: add a0, a0, a2
-; RV64I-NEXT: ret
+; RV64-LABEL: add_select_all_zeros_i64:
+; RV64: # %bb.0:
+; RV64-NEXT: neg a0, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: add a0, a0, a2
+; RV64-NEXT: ret
;
; SFB64-LABEL: add_select_all_zeros_i64:
; SFB64: # %bb.0:
-; SFB64-NEXT: beqz a0, .LBB7_2
+; SFB64-NEXT: beqz a0, .LBB11_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: add a2, a2, a1
-; SFB64-NEXT: .LBB7_2:
+; SFB64-NEXT: .LBB11_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -364,19 +613,19 @@ define signext i32 @sub_select_all_zeros_i32(i1 zeroext %c, i32 signext %x, i32
; RV32I-NEXT: sub a0, a2, a0
; RV32I-NEXT: ret
;
-; RV64I-LABEL: sub_select_all_zeros_i32:
-; RV64I: # %bb.0:
-; RV64I-NEXT: addi a0, a0, -1
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: subw a0, a2, a0
-; RV64I-NEXT: ret
+; RV64-LABEL: sub_select_all_zeros_i32:
+; RV64: # %bb.0:
+; RV64-NEXT: addi a0, a0, -1
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: subw a0, a2, a0
+; RV64-NEXT: ret
;
; SFB64-LABEL: sub_select_all_zeros_i32:
; SFB64: # %bb.0:
-; SFB64-NEXT: bnez a0, .LBB8_2
+; SFB64-NEXT: bnez a0, .LBB12_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: subw a2, a2, a1
-; SFB64-NEXT: .LBB8_2:
+; SFB64-NEXT: .LBB12_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
@@ -414,19 +663,19 @@ define i64 @sub_select_all_zeros_i64(i1 zeroext %c, i64 %x, i64 %y) {
; RV32I-NEXT: sub a0, a3, a0
; RV32I-NEXT: ret
;
-; RV64I-LABEL: sub_select_all_zeros_i64:
-; RV64I: # %bb.0:
-; RV64I-NEXT: neg a0, a0
-; RV64I-NEXT: and a0, a0, a1
-; RV64I-NEXT: sub a0, a2, a0
-; RV64I-NEXT: ret
+; RV64-LABEL: sub_select_all_zeros_i64:
+; RV64: # %bb.0:
+; RV64-NEXT: neg a0, a0
+; RV64-NEXT: and a0, a0, a1
+; RV64-NEXT: sub a0, a2, a0
+; RV64-NEXT: ret
;
; SFB64-LABEL: sub_select_all_zeros_i64:
; SFB64: # %bb.0:
-; SFB64-NEXT: beqz a0, .LBB9_2
+; SFB64-NEXT: beqz a0, .LBB13_2
; SFB64-NEXT: # %bb.1:
; SFB64-NEXT: sub a2, a2, a1
-; SFB64-NEXT: .LBB9_2:
+; SFB64-NEXT: .LBB13_2:
; SFB64-NEXT: mv a0, a2
; SFB64-NEXT: ret
;
``````````
</details>
https://github.com/llvm/llvm-project/pull/69563
- Previous message: [llvm] [RISCV] Combine (and (select cond, x, -1), c) to (select cond, x, (and x, c)) with Zicond. (PR #69563)
- Next message: [llvm] [RISCV] Combine (and (select cond, x, -1), c) to (select cond, x, (and x, c)) with Zicond. (PR #69563)
- Messages sorted by:
[ date ]
[ thread ]
[ subject ]
[ author ]
More information about the llvm-commits
mailing list