[llvm] 28f12a0 - [RISCV] Teach ComputeNumSignBitsForTargetNode about masked atomic intrinsics
Alex Bradbury via llvm-commits
llvm-commits at lists.llvm.org
Wed Aug 3 05:45:59 PDT 2022
Author: Alex Bradbury
Date: 2022-08-03T13:41:58+01:00
New Revision: 28f12a09ae631885410d7524f7c60748a59358c5
URL: https://github.com/llvm/llvm-project/commit/28f12a09ae631885410d7524f7c60748a59358c5
DIFF: https://github.com/llvm/llvm-project/commit/28f12a09ae631885410d7524f7c60748a59358c5.diff
LOG: [RISCV] Teach ComputeNumSignBitsForTargetNode about masked atomic intrinsics
An unnecessary sext.w is generated when masking the result of the
riscv_masked_cmpxchg_i64 intrinsic. Implementing handling of the
intrinsic in ComputeNumSignBitsForTargetNode allows it to be removed.
Although this isn't a particularly important optimisation, removing the
sext.w simplifies implementation of an additional cmpxchg-related
optimisation in D130192.
Although I can't produce a test with different codegen for the other
atomics intrinsics, these are added as well for completeness.
Differential Revision: https://reviews.llvm.org/D130191
Added:
Modified:
llvm/lib/Target/RISCV/RISCVISelLowering.cpp
llvm/test/CodeGen/RISCV/atomic-cmpxchg-branch-on-result.ll
llvm/test/CodeGen/RISCV/atomic-signext.ll
Removed:
################################################################################
diff --git a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
index af5dae9f3284d..2b41fed32e76b 100644
--- a/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
+++ b/llvm/lib/Target/RISCV/RISCVISelLowering.cpp
@@ -9882,6 +9882,31 @@ unsigned RISCVTargetLowering::ComputeNumSignBitsForTargetNode(
return XLen - EltBits + 1;
break;
}
+ case ISD::INTRINSIC_W_CHAIN: {
+ unsigned IntNo = Op.getConstantOperandVal(1);
+ switch (IntNo) {
+ default:
+ break;
+ case Intrinsic::riscv_masked_atomicrmw_xchg_i64:
+ case Intrinsic::riscv_masked_atomicrmw_add_i64:
+ case Intrinsic::riscv_masked_atomicrmw_sub_i64:
+ case Intrinsic::riscv_masked_atomicrmw_nand_i64:
+ case Intrinsic::riscv_masked_atomicrmw_max_i64:
+ case Intrinsic::riscv_masked_atomicrmw_min_i64:
+ case Intrinsic::riscv_masked_atomicrmw_umax_i64:
+ case Intrinsic::riscv_masked_atomicrmw_umin_i64:
+ case Intrinsic::riscv_masked_cmpxchg_i64:
+ // riscv_masked_{atomicrmw_*,cmpxchg} intrinsics represent an emulated
+ // narrow atomic operation. These are implemented using atomic
+ // operations at the minimum supported atomicrmw/cmpxchg width whose
+ // result is then sign extended to XLEN. With +A, the minimum width is
+ // 32 for both 64 and 32.
+ assert(Subtarget.getXLen() == 64);
+ assert(getMinCmpXchgSizeInBits() == 32);
+ assert(Subtarget.hasStdExtA());
+ return 33;
+ }
+ }
}
return 1;
diff --git a/llvm/test/CodeGen/RISCV/atomic-cmpxchg-branch-on-result.ll b/llvm/test/CodeGen/RISCV/atomic-cmpxchg-branch-on-result.ll
index cfa11ac5f2b47..e2be4af75cbe3 100644
--- a/llvm/test/CodeGen/RISCV/atomic-cmpxchg-branch-on-result.ll
+++ b/llvm/test/CodeGen/RISCV/atomic-cmpxchg-branch-on-result.ll
@@ -130,7 +130,6 @@ define void @cmpxchg_masked_and_branch1(i8* %ptr, i8 signext %cmp, i8 signext %v
; RV64IA-NEXT: .LBB2_5: # %do_cmpxchg
; RV64IA-NEXT: # in Loop: Header=BB2_1 Depth=1
; RV64IA-NEXT: and a4, a4, a0
-; RV64IA-NEXT: sext.w a4, a4
; RV64IA-NEXT: bne a1, a4, .LBB2_1
; RV64IA-NEXT: # %bb.2: # %exit
; RV64IA-NEXT: ret
@@ -207,7 +206,6 @@ define void @cmpxchg_masked_and_branch2(i8* %ptr, i8 signext %cmp, i8 signext %v
; RV64IA-NEXT: .LBB3_5: # %do_cmpxchg
; RV64IA-NEXT: # in Loop: Header=BB3_1 Depth=1
; RV64IA-NEXT: and a4, a4, a0
-; RV64IA-NEXT: sext.w a4, a4
; RV64IA-NEXT: beq a1, a4, .LBB3_1
; RV64IA-NEXT: # %bb.2: # %exit
; RV64IA-NEXT: ret
diff --git a/llvm/test/CodeGen/RISCV/atomic-signext.ll b/llvm/test/CodeGen/RISCV/atomic-signext.ll
index de81b12493745..e893b38873538 100644
--- a/llvm/test/CodeGen/RISCV/atomic-signext.ll
+++ b/llvm/test/CodeGen/RISCV/atomic-signext.ll
@@ -3904,7 +3904,6 @@ define i1 @cmpxchg_i8_monotonic_monotonic_val1(i8* %ptr, i8 signext %cmp, i8 sig
; RV64IA-NEXT: bnez a5, .LBB48_1
; RV64IA-NEXT: .LBB48_3:
; RV64IA-NEXT: and a0, a2, a4
-; RV64IA-NEXT: sext.w a0, a0
; RV64IA-NEXT: xor a0, a1, a0
; RV64IA-NEXT: seqz a0, a0
; RV64IA-NEXT: ret
@@ -4077,7 +4076,6 @@ define i1 @cmpxchg_i16_monotonic_monotonic_val1(i16* %ptr, i16 signext %cmp, i16
; RV64IA-NEXT: bnez a4, .LBB50_1
; RV64IA-NEXT: .LBB50_3:
; RV64IA-NEXT: and a0, a2, a5
-; RV64IA-NEXT: sext.w a0, a0
; RV64IA-NEXT: xor a0, a1, a0
; RV64IA-NEXT: seqz a0, a0
; RV64IA-NEXT: ret
More information about the llvm-commits
mailing list