[llvm] 23c5097 - [DAGCombiner] Stop invalid sign conversion in refineIndexType.
Paul Walker via llvm-commits
llvm-commits at lists.llvm.org
Fri Apr 29 06:23:54 PDT 2022
Author: Paul Walker
Date: 2022-04-29T14:20:13+01:00
New Revision: 23c509754d4b81e5503f8da5caa3d4c00af85afb
URL: https://github.com/llvm/llvm-project/commit/23c509754d4b81e5503f8da5caa3d4c00af85afb
DIFF: https://github.com/llvm/llvm-project/commit/23c509754d4b81e5503f8da5caa3d4c00af85afb.diff
LOG: [DAGCombiner] Stop invalid sign conversion in refineIndexType.
When looking through extends of gather/scatter indices it's safe
to convert a known positive signed index to unsigned, but unsigned
indices must remain unsigned.
Depends On D123318
Differential Revision: https://reviews.llvm.org/D123326
Added:
Modified:
llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
llvm/test/CodeGen/AArch64/sve-gather-scatter-addr-opts.ll
Removed:
################################################################################
diff --git a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
index 181ff00184b3..5f964c08b28e 100644
--- a/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
+++ b/llvm/lib/CodeGen/SelectionDAG/DAGCombiner.cpp
@@ -10448,9 +10448,10 @@ bool refineUniformBase(SDValue &BasePtr, SDValue &Index, bool IndexIsScaled,
// Fold sext/zext of index into index type.
bool refineIndexType(MaskedGatherScatterSDNode *MGS, SDValue &Index,
- bool Scaled, SelectionDAG &DAG) {
+ bool Scaled, bool Signed, SelectionDAG &DAG) {
const TargetLowering &TLI = DAG.getTargetLoweringInfo();
+ // It's always safe to look through zero extends.
if (Index.getOpcode() == ISD::ZERO_EXTEND) {
SDValue Op = Index.getOperand(0);
MGS->setIndexType(Scaled ? ISD::UNSIGNED_SCALED : ISD::UNSIGNED_UNSCALED);
@@ -10460,7 +10461,8 @@ bool refineIndexType(MaskedGatherScatterSDNode *MGS, SDValue &Index,
}
}
- if (Index.getOpcode() == ISD::SIGN_EXTEND) {
+ // It's only safe to look through sign extends when Index is signed.
+ if (Index.getOpcode() == ISD::SIGN_EXTEND && Signed) {
SDValue Op = Index.getOperand(0);
MGS->setIndexType(Scaled ? ISD::SIGNED_SCALED : ISD::SIGNED_UNSCALED);
if (TLI.shouldRemoveExtendFromGSIndex(Op.getValueType())) {
@@ -10493,7 +10495,8 @@ SDValue DAGCombiner::visitMSCATTER(SDNode *N) {
MSC->getMemOperand(), MSC->getIndexType(), MSC->isTruncatingStore());
}
- if (refineIndexType(MSC, Index, MSC->isIndexScaled(), DAG)) {
+ if (refineIndexType(MSC, Index, MSC->isIndexScaled(), MSC->isIndexSigned(),
+ DAG)) {
SDValue Ops[] = {Chain, StoreVal, Mask, BasePtr, Index, Scale};
return DAG.getMaskedScatter(
DAG.getVTList(MVT::Other), MSC->getMemoryVT(), DL, Ops,
@@ -10589,7 +10592,8 @@ SDValue DAGCombiner::visitMGATHER(SDNode *N) {
MGT->getExtensionType());
}
- if (refineIndexType(MGT, Index, MGT->isIndexScaled(), DAG)) {
+ if (refineIndexType(MGT, Index, MGT->isIndexScaled(), MGT->isIndexSigned(),
+ DAG)) {
SDValue Ops[] = {Chain, PassThru, Mask, BasePtr, Index, Scale};
return DAG.getMaskedGather(DAG.getVTList(N->getValueType(0), MVT::Other),
MGT->getMemoryVT(), DL, Ops,
diff --git a/llvm/test/CodeGen/AArch64/sve-gather-scatter-addr-opts.ll b/llvm/test/CodeGen/AArch64/sve-gather-scatter-addr-opts.ll
index f4b4a033c343..82beb94dfcac 100644
--- a/llvm/test/CodeGen/AArch64/sve-gather-scatter-addr-opts.ll
+++ b/llvm/test/CodeGen/AArch64/sve-gather-scatter-addr-opts.ll
@@ -399,14 +399,12 @@ define <vscale x 4 x i32> @masked_gather_nxv4i32_u8_offsets(i32* %base, <vscale
ret <vscale x 4 x i32> %data
}
-; TODO: The generated code is wrong because we're replicating offset[31] across
-; offset[32:63] even though the IR has explicitly zero'd those bits.
define <vscale x 4 x i32> @masked_gather_nxv4i32_u32s8_offsets(i32* %base, <vscale x 4 x i8> %offsets, <vscale x 4 x i1> %mask) #0 {
; CHECK-LABEL: masked_gather_nxv4i32_u32s8_offsets:
; CHECK: // %bb.0:
; CHECK-NEXT: ptrue p1.s
; CHECK-NEXT: sxtb z0.s, p1/m, z0.s
-; CHECK-NEXT: ld1w { z0.s }, p0/z, [x0, z0.s, sxtw #2]
+; CHECK-NEXT: ld1w { z0.s }, p0/z, [x0, z0.s, uxtw #2]
; CHECK-NEXT: ret
%offsets.sext = sext <vscale x 4 x i8> %offsets to <vscale x 4 x i32>
%offsets.sext.zext = zext <vscale x 4 x i32> %offsets.sext to <vscale x 4 x i64>
@@ -482,14 +480,12 @@ define void @masked_scatter_nxv4i32_u8_offsets(i32* %base, <vscale x 4 x i8> %of
ret void
}
-; TODO: The generated code is wrong because we're replicating offset[31] across
-; offset[32:63] even though the IR has explicitly zero'd those bits.
define void @masked_scatter_nxv4i32_u32s8_offsets(i32* %base, <vscale x 4 x i8> %offsets, <vscale x 4 x i1> %mask, <vscale x 4 x i32> %data) #0 {
; CHECK-LABEL: masked_scatter_nxv4i32_u32s8_offsets:
; CHECK: // %bb.0:
; CHECK-NEXT: ptrue p1.s
; CHECK-NEXT: sxtb z0.s, p1/m, z0.s
-; CHECK-NEXT: st1w { z1.s }, p0, [x0, z0.s, sxtw #2]
+; CHECK-NEXT: st1w { z1.s }, p0, [x0, z0.s, uxtw #2]
; CHECK-NEXT: ret
%offsets.sext = sext <vscale x 4 x i8> %offsets to <vscale x 4 x i32>
%offsets.sext.zext = zext <vscale x 4 x i32> %offsets.sext to <vscale x 4 x i64>
More information about the llvm-commits
mailing list