[llvm] d8ef1d1 - AMDGPU/GlobalISel: Fix selecting broken copies for s32->s64 anyext
Matt Arsenault via llvm-commits
llvm-commits at lists.llvm.org
Mon Aug 3 05:36:47 PDT 2020
Author: Matt Arsenault
Date: 2020-08-03T08:36:41-04:00
New Revision: d8ef1d1251e3c0e11894ed82904dbab5e41c5711
URL: https://github.com/llvm/llvm-project/commit/d8ef1d1251e3c0e11894ed82904dbab5e41c5711
DIFF: https://github.com/llvm/llvm-project/commit/d8ef1d1251e3c0e11894ed82904dbab5e41c5711.diff
LOG: AMDGPU/GlobalISel: Fix selecting broken copies for s32->s64 anyext
These should probably not be legal in the first place, but that might
also be a pain.
Added:
Modified:
llvm/lib/Target/AMDGPU/AMDGPUInstructionSelector.cpp
llvm/test/CodeGen/AMDGPU/GlobalISel/inst-select-anyext.mir
Removed:
################################################################################
diff --git a/llvm/lib/Target/AMDGPU/AMDGPUInstructionSelector.cpp b/llvm/lib/Target/AMDGPU/AMDGPUInstructionSelector.cpp
index 709329b4c0c6..5aceb40933c3 100644
--- a/llvm/lib/Target/AMDGPU/AMDGPUInstructionSelector.cpp
+++ b/llvm/lib/Target/AMDGPU/AMDGPUInstructionSelector.cpp
@@ -1894,12 +1894,33 @@ bool AMDGPUInstructionSelector::selectG_SZA_EXT(MachineInstr &I) const {
if (!DstTy.isScalar())
return false;
- if (I.getOpcode() == AMDGPU::G_ANYEXT)
- return selectCOPY(I);
-
// Artifact casts should never use vcc.
const RegisterBank *SrcBank = getArtifactRegBank(SrcReg, *MRI, TRI);
+ // FIXME: This should probably be illegal and split earlier.
+ if (I.getOpcode() == AMDGPU::G_ANYEXT) {
+ if (DstSize <= 32)
+ return selectCOPY(I);
+
+ const TargetRegisterClass *SrcRC =
+ TRI.getRegClassForTypeOnBank(SrcTy, *SrcBank, *MRI);
+ const RegisterBank *DstBank = RBI.getRegBank(DstReg, *MRI, TRI);
+ const TargetRegisterClass *DstRC =
+ TRI.getRegClassForSizeOnBank(DstSize, *DstBank, *MRI);
+
+ Register UndefReg = MRI->createVirtualRegister(SrcRC);
+ BuildMI(MBB, I, DL, TII.get(AMDGPU::IMPLICIT_DEF), UndefReg);
+ BuildMI(MBB, I, DL, TII.get(AMDGPU::REG_SEQUENCE), DstReg)
+ .addReg(SrcReg)
+ .addImm(AMDGPU::sub0)
+ .addReg(UndefReg)
+ .addImm(AMDGPU::sub1);
+ I.eraseFromParent();
+
+ return RBI.constrainGenericRegister(DstReg, *DstRC, *MRI) &&
+ RBI.constrainGenericRegister(SrcReg, *SrcRC, *MRI);
+ }
+
if (SrcBank->getID() == AMDGPU::VGPRRegBankID && DstSize <= 32) {
// 64-bit should have been split up in RegBankSelect
diff --git a/llvm/test/CodeGen/AMDGPU/GlobalISel/inst-select-anyext.mir b/llvm/test/CodeGen/AMDGPU/GlobalISel/inst-select-anyext.mir
index 58d01774f745..dcad0a85e8e0 100644
--- a/llvm/test/CodeGen/AMDGPU/GlobalISel/inst-select-anyext.mir
+++ b/llvm/test/CodeGen/AMDGPU/GlobalISel/inst-select-anyext.mir
@@ -22,22 +22,88 @@ body: |
...
---
+name: anyext_sgpr_s32_to_sgpr_s64
+legalized: true
+regBankSelected: true
+tracksRegLiveness: true
+body: |
+ bb.0:
+ liveins: $sgpr0
+
+ ; GCN-LABEL: name: anyext_sgpr_s32_to_sgpr_s64
+ ; GCN: liveins: $sgpr0
+ ; GCN: [[COPY:%[0-9]+]]:sreg_32 = COPY $sgpr0
+ ; GCN: [[DEF:%[0-9]+]]:sreg_32 = IMPLICIT_DEF
+ ; GCN: [[REG_SEQUENCE:%[0-9]+]]:sreg_64 = REG_SEQUENCE [[COPY]], %subreg.sub0, [[DEF]], %subreg.sub1
+ ; GCN: S_ENDPGM 0, implicit [[REG_SEQUENCE]]
+ %0:sgpr(s32) = COPY $sgpr0
+ %1:sgpr(s64) = G_ANYEXT %0
+ S_ENDPGM 0, implicit %1
+
+...
-name: anyext_sgpr_s16_to_sgpr_s64
+---
+name: anyext_sgpr_s16_to_sgpr_s64
legalized: true
regBankSelected: true
-body: |
+tracksRegLiveness: true
+body: |
bb.0:
liveins: $sgpr0
; GCN-LABEL: name: anyext_sgpr_s16_to_sgpr_s64
+ ; GCN: liveins: $sgpr0
; GCN: [[COPY:%[0-9]+]]:sreg_32 = COPY $sgpr0
- ; GCN: [[COPY1:%[0-9]+]]:sreg_64 = COPY [[COPY]]
- ; GCN: $sgpr0_sgpr1 = COPY [[COPY1]]
+ ; GCN: [[DEF:%[0-9]+]]:sreg_32 = IMPLICIT_DEF
+ ; GCN: [[REG_SEQUENCE:%[0-9]+]]:sreg_64 = REG_SEQUENCE [[COPY]], %subreg.sub0, [[DEF]], %subreg.sub1
+ ; GCN: S_ENDPGM 0, implicit [[REG_SEQUENCE]]
%0:sgpr(s32) = COPY $sgpr0
%1:sgpr(s16) = G_TRUNC %0
%2:sgpr(s64) = G_ANYEXT %1
- $sgpr0_sgpr1 = COPY %2
+ S_ENDPGM 0, implicit %2
+
+...
+
+---
+name: anyext_vgpr_s32_to_vgpr_s64
+legalized: true
+regBankSelected: true
+tracksRegLiveness: true
+body: |
+ bb.0:
+ liveins: $vgpr0
+
+ ; GCN-LABEL: name: anyext_vgpr_s32_to_vgpr_s64
+ ; GCN: liveins: $vgpr0
+ ; GCN: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr0
+ ; GCN: [[DEF:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF
+ ; GCN: [[REG_SEQUENCE:%[0-9]+]]:vreg_64 = REG_SEQUENCE [[COPY]], %subreg.sub0, [[DEF]], %subreg.sub1
+ ; GCN: S_ENDPGM 0, implicit [[REG_SEQUENCE]]
+ %0:vgpr(s32) = COPY $vgpr0
+ %1:vgpr(s64) = G_ANYEXT %0
+ S_ENDPGM 0, implicit %1
+
+...
+
+---
+name: anyext_vgpr_s16_to_vgpr_s64
+legalized: true
+regBankSelected: true
+tracksRegLiveness: true
+body: |
+ bb.0:
+ liveins: $vgpr0
+
+ ; GCN-LABEL: name: anyext_vgpr_s16_to_vgpr_s64
+ ; GCN: liveins: $vgpr0
+ ; GCN: [[COPY:%[0-9]+]]:vgpr_32 = COPY $vgpr0
+ ; GCN: [[DEF:%[0-9]+]]:vgpr_32 = IMPLICIT_DEF
+ ; GCN: [[REG_SEQUENCE:%[0-9]+]]:vreg_64 = REG_SEQUENCE [[COPY]], %subreg.sub0, [[DEF]], %subreg.sub1
+ ; GCN: S_ENDPGM 0, implicit [[REG_SEQUENCE]]
+ %0:vgpr(s32) = COPY $vgpr0
+ %1:vgpr(s16) = G_TRUNC %0
+ %2:vgpr(s64) = G_ANYEXT %1
+ S_ENDPGM 0, implicit %2
...
More information about the llvm-commits
mailing list