[llvm] [LoongArch][NFC] Pre-commit for conditional branch optimization (PR #151788)

via llvm-commits llvm-commits at lists.llvm.org
Fri Aug 1 20:57:33 PDT 2025


llvmbot wrote:


<!--LLVM PR SUMMARY COMMENT-->

@llvm/pr-subscribers-backend-loongarch

Author: hev (heiher)

<details>
<summary>Changes</summary>



---

Patch is 87.90 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/151788.diff


2 Files Affected:

- (added) llvm/test/CodeGen/LoongArch/bittest.ll (+3366) 
- (modified) llvm/test/CodeGen/LoongArch/select-const.ll (+25) 


``````````diff
diff --git a/llvm/test/CodeGen/LoongArch/bittest.ll b/llvm/test/CodeGen/LoongArch/bittest.ll
new file mode 100644
index 0000000000000..210e4edbb38ff
--- /dev/null
+++ b/llvm/test/CodeGen/LoongArch/bittest.ll
@@ -0,0 +1,3366 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc --mtriple=loongarch32 -mattr=+d < %s | FileCheck %s --check-prefixes=CHECK,LA32
+; RUN: llc --mtriple=loongarch64 -mattr=+d < %s | FileCheck %s --check-prefixes=CHECK,LA64
+
+define signext i32 @bittest_7_i32(i32 signext %a) nounwind {
+; LA32-LABEL: bittest_7_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 7
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_7_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 7, 7
+; LA64-NEXT:    ret
+  %shr = lshr i32 %a, 7
+  %not = xor i32 %shr, -1
+  %and = and i32 %not, 1
+  ret i32 %and
+}
+
+define signext i32 @bittest_10_i32(i32 signext %a) nounwind {
+; LA32-LABEL: bittest_10_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 10
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_10_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 10, 10
+; LA64-NEXT:    ret
+  %shr = lshr i32 %a, 10
+  %not = xor i32 %shr, -1
+  %and = and i32 %not, 1
+  ret i32 %and
+}
+
+define signext i32 @bittest_11_i32(i32 signext %a) nounwind {
+; LA32-LABEL: bittest_11_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 11
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_11_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 11, 11
+; LA64-NEXT:    ret
+  %shr = lshr i32 %a, 11
+  %not = xor i32 %shr, -1
+  %and = and i32 %not, 1
+  ret i32 %and
+}
+
+define signext i32 @bittest_31_i32(i32 signext %a) nounwind {
+; LA32-LABEL: bittest_31_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 31
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_31_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 31, 31
+; LA64-NEXT:    ret
+  %shr = lshr i32 %a, 31
+  %not = xor i32 %shr, -1
+  %and = and i32 %not, 1
+  ret i32 %and
+}
+
+define i64 @bittest_7_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_7_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 7
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_7_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 7, 7
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 7
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i64 @bittest_10_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_10_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 10
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_10_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 10, 10
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 10
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i64 @bittest_11_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_11_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 11
+; LA32-NEXT:    andi $a0, $a0, 1
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_11_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 11, 11
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 11
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i64 @bittest_31_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_31_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a0, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 31
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_31_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 31, 31
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 31
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i64 @bittest_32_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_32_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    ori $a0, $zero, 1
+; LA32-NEXT:    andn $a0, $a0, $a1
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_32_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    bstrpick.d $a0, $a0, 32, 32
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 32
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i64 @bittest_63_i64(i64 %a) nounwind {
+; LA32-LABEL: bittest_63_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    nor $a0, $a1, $zero
+; LA32-NEXT:    srli.w $a0, $a0, 31
+; LA32-NEXT:    move $a1, $zero
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_63_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    nor $a0, $a0, $zero
+; LA64-NEXT:    srli.d $a0, $a0, 63
+; LA64-NEXT:    ret
+  %shr = lshr i64 %a, 63
+  %not = xor i64 %shr, -1
+  %and = and i64 %not, 1
+  ret i64 %and
+}
+
+define i1 @bittest_constant_by_var_shr_i32(i32 signext %b) nounwind {
+; CHECK-LABEL: bittest_constant_by_var_shr_i32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    lu12i.w $a1, 301408
+; CHECK-NEXT:    ori $a1, $a1, 722
+; CHECK-NEXT:    srl.w $a0, $a1, $a0
+; CHECK-NEXT:    andi $a0, $a0, 1
+; CHECK-NEXT:    ret
+  %shl = lshr i32 1234567890, %b
+  %and = and i32 %shl, 1
+  %cmp = icmp ne i32 %and, 0
+  ret i1 %cmp
+}
+
+define i1 @bittest_constant_by_var_shl_i32(i32 signext %b) nounwind {
+; CHECK-LABEL: bittest_constant_by_var_shl_i32:
+; CHECK:       # %bb.0:
+; CHECK-NEXT:    ori $a1, $zero, 1
+; CHECK-NEXT:    sll.w $a0, $a1, $a0
+; CHECK-NEXT:    lu12i.w $a1, 301408
+; CHECK-NEXT:    ori $a1, $a1, 722
+; CHECK-NEXT:    and $a0, $a0, $a1
+; CHECK-NEXT:    sltu $a0, $zero, $a0
+; CHECK-NEXT:    ret
+  %shl = shl i32 1, %b
+  %and = and i32 %shl, 1234567890
+  %cmp = icmp ne i32 %and, 0
+  ret i1 %cmp
+}
+
+define i1 @bittest_constant_by_var_shr_i64(i64 %b) nounwind {
+; LA32-LABEL: bittest_constant_by_var_shr_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    lu12i.w $a1, 301408
+; LA32-NEXT:    ori $a1, $a1, 722
+; LA32-NEXT:    srl.w $a1, $a1, $a0
+; LA32-NEXT:    addi.w $a0, $a0, -32
+; LA32-NEXT:    slti $a0, $a0, 0
+; LA32-NEXT:    and $a0, $a0, $a1
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_constant_by_var_shr_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    lu12i.w $a1, 301408
+; LA64-NEXT:    ori $a1, $a1, 722
+; LA64-NEXT:    srl.d $a0, $a1, $a0
+; LA64-NEXT:    andi $a0, $a0, 1
+; LA64-NEXT:    ret
+  %shl = lshr i64 1234567890, %b
+  %and = and i64 %shl, 1
+  %cmp = icmp ne i64 %and, 0
+  ret i1 %cmp
+}
+
+define i1 @bittest_constant_by_var_shl_i64(i64 %b) nounwind {
+; LA32-LABEL: bittest_constant_by_var_shl_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    addi.w $a1, $a0, -32
+; LA32-NEXT:    slti $a1, $a1, 0
+; LA32-NEXT:    sub.w $a1, $zero, $a1
+; LA32-NEXT:    ori $a2, $zero, 1
+; LA32-NEXT:    sll.w $a0, $a2, $a0
+; LA32-NEXT:    and $a0, $a1, $a0
+; LA32-NEXT:    lu12i.w $a1, 301408
+; LA32-NEXT:    ori $a1, $a1, 722
+; LA32-NEXT:    and $a0, $a0, $a1
+; LA32-NEXT:    sltu $a0, $zero, $a0
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_constant_by_var_shl_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    ori $a1, $zero, 1
+; LA64-NEXT:    sll.d $a0, $a1, $a0
+; LA64-NEXT:    lu12i.w $a1, 301408
+; LA64-NEXT:    ori $a1, $a1, 722
+; LA64-NEXT:    and $a0, $a0, $a1
+; LA64-NEXT:    sltu $a0, $zero, $a0
+; LA64-NEXT:    ret
+  %shl = shl i64 1, %b
+  %and = and i64 %shl, 1234567890
+  %cmp = icmp ne i64 %and, 0
+  ret i1 %cmp
+}
+
+define void @bittest_switch(i32 signext %0) {
+; LA32-LABEL: bittest_switch:
+; LA32:       # %bb.0:
+; LA32-NEXT:    ori $a1, $zero, 31
+; LA32-NEXT:    bltu $a1, $a0, .LBB14_3
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    ori $a1, $zero, 1
+; LA32-NEXT:    sll.w $a0, $a1, $a0
+; LA32-NEXT:    lu12i.w $a1, -524285
+; LA32-NEXT:    ori $a1, $a1, 768
+; LA32-NEXT:    and $a0, $a0, $a1
+; LA32-NEXT:    beq $a0, $zero, .LBB14_3
+; LA32-NEXT:  # %bb.2:
+; LA32-NEXT:    b bar
+; LA32-NEXT:  .LBB14_3:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bittest_switch:
+; LA64:       # %bb.0:
+; LA64-NEXT:    ori $a1, $zero, 31
+; LA64-NEXT:    bltu $a1, $a0, .LBB14_3
+; LA64-NEXT:  # %bb.1:
+; LA64-NEXT:    ori $a1, $zero, 1
+; LA64-NEXT:    sll.d $a0, $a1, $a0
+; LA64-NEXT:    lu12i.w $a1, -524285
+; LA64-NEXT:    ori $a1, $a1, 768
+; LA64-NEXT:    lu32i.d $a1, 0
+; LA64-NEXT:    and $a0, $a0, $a1
+; LA64-NEXT:    beqz $a0, .LBB14_3
+; LA64-NEXT:  # %bb.2:
+; LA64-NEXT:    pcaddu18i $t8, %call36(bar)
+; LA64-NEXT:    jr $t8
+; LA64-NEXT:  .LBB14_3:
+; LA64-NEXT:    ret
+  switch i32 %0, label %3 [
+    i32 8, label %2
+    i32 9, label %2
+    i32 12, label %2
+    i32 13, label %2
+    i32 31, label %2
+  ]
+
+2:
+  tail call void @bar()
+  br label %3
+
+3:
+  ret void
+}
+
+declare void @bar()
+
+define signext i32 @bit_10_z_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_10_z_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    andi $a3, $a0, 1024
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    beq $a3, $zero, .LBB15_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB15_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_10_z_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    andi $a0, $a0, 1024
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 1024
+  %2 = icmp eq i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_10_nz_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_10_nz_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    andi $a0, $a0, 1024
+; LA32-NEXT:    srli.w $a3, $a0, 10
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bne $a3, $zero, .LBB16_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB16_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_10_nz_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 10, 10
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 1024
+  %2 = icmp ne i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_11_z_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_11_z_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    slli.w $a3, $a0, 20
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bgez $a3, .LBB17_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB17_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_11_z_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    andi $a0, $a0, 2048
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 2048
+  %2 = icmp eq i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_11_nz_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_11_nz_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    andi $a0, $a0, 2048
+; LA32-NEXT:    srli.w $a3, $a0, 11
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bne $a3, $zero, .LBB18_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB18_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_11_nz_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 11, 11
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 2048
+  %2 = icmp ne i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_20_z_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_20_z_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    slli.w $a3, $a0, 11
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bgez $a3, .LBB19_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB19_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_20_z_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    lu12i.w $a3, 256
+; LA64-NEXT:    and $a0, $a0, $a3
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 1048576
+  %2 = icmp eq i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_20_nz_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_20_nz_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    lu12i.w $a3, 256
+; LA32-NEXT:    and $a0, $a0, $a3
+; LA32-NEXT:    srli.w $a3, $a0, 20
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bne $a3, $zero, .LBB20_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB20_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_20_nz_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 20, 20
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 1048576
+  %2 = icmp ne i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_31_z_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_31_z_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    bgez $a0, .LBB21_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a1, $a2
+; LA32-NEXT:  .LBB21_2:
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_31_z_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrins.d $a0, $zero, 30, 0
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 2147483648
+  %2 = icmp eq i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define signext i32 @bit_31_nz_select_i32(i32 signext %a, i32 signext %b, i32 signext %c) {
+; LA32-LABEL: bit_31_nz_select_i32:
+; LA32:       # %bb.0:
+; LA32-NEXT:    srli.w $a3, $a0, 31
+; LA32-NEXT:    move $a0, $a1
+; LA32-NEXT:    bne $a3, $zero, .LBB22_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:  .LBB22_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_31_nz_select_i32:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrins.d $a0, $zero, 30, 0
+; LA64-NEXT:    sltu $a0, $zero, $a0
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i32 %a, 2147483648
+  %2 = icmp ne i32 %1, 0
+  %3 = select i1 %2, i32 %b, i32 %c
+  ret i32 %3
+}
+
+define i64 @bit_10_z_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_10_z_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    andi $a6, $a0, 1024
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    beq $a6, $zero, .LBB23_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB23_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_10_z_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    andi $a0, $a0, 1024
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 1024
+  %2 = icmp eq i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_10_nz_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_10_nz_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    srli.w $a0, $a0, 10
+; LA32-NEXT:    andi $a6, $a0, 1
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    bne $a6, $zero, .LBB24_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB24_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_10_nz_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 10, 10
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 1024
+  %2 = icmp ne i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_11_z_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_11_z_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    slli.w $a6, $a0, 20
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    bgez $a6, .LBB25_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB25_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_11_z_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    andi $a0, $a0, 2048
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 2048
+  %2 = icmp eq i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_11_nz_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_11_nz_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    srli.w $a0, $a0, 11
+; LA32-NEXT:    andi $a6, $a0, 1
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    bne $a6, $zero, .LBB26_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB26_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_11_nz_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 11, 11
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 2048
+  %2 = icmp ne i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_20_z_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_20_z_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    slli.w $a6, $a0, 11
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    bgez $a6, .LBB27_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB27_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_20_z_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    lu12i.w $a3, 256
+; LA64-NEXT:    and $a0, $a0, $a3
+; LA64-NEXT:    sltui $a0, $a0, 1
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 1048576
+  %2 = icmp eq i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_20_nz_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_20_nz_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    srli.w $a0, $a0, 20
+; LA32-NEXT:    andi $a6, $a0, 1
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    bne $a6, $zero, .LBB28_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a0, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB28_2:
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_20_nz_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    bstrpick.d $a0, $a0, 20, 20
+; LA64-NEXT:    masknez $a2, $a2, $a0
+; LA64-NEXT:    maskeqz $a0, $a1, $a0
+; LA64-NEXT:    or $a0, $a0, $a2
+; LA64-NEXT:    ret
+  %1 = and i64 %a, 1048576
+  %2 = icmp ne i64 %1, 0
+  %3 = select i1 %2, i64 %b, i64 %c
+  ret i64 %3
+}
+
+define i64 @bit_31_z_select_i64(i64 %a, i64 %b, i64 %c) {
+; LA32-LABEL: bit_31_z_select_i64:
+; LA32:       # %bb.0:
+; LA32-NEXT:    move $a1, $a3
+; LA32-NEXT:    bgez $a0, .LBB29_2
+; LA32-NEXT:  # %bb.1:
+; LA32-NEXT:    move $a2, $a4
+; LA32-NEXT:    move $a1, $a5
+; LA32-NEXT:  .LBB29_2:
+; LA32-NEXT:    move $a0, $a2
+; LA32-NEXT:    ret
+;
+; LA64-LABEL: bit_31_z_select_i64:
+; LA64:       # %bb.0:
+; LA64-NEXT:    lu12i.w $a3, ...
[truncated]

``````````

</details>


https://github.com/llvm/llvm-project/pull/151788


More information about the llvm-commits mailing list