[PATCH] R600/SI: Try to keep mul on SALU

Matt Arsenault arsenm2 at gmail.com
Wed Sep 3 12:32:14 PDT 2014


On Sep 3, 2014, at 2:53 PM, Tom Stellard <tom at stellard.net> wrote:

> On Sat, Aug 30, 2014 at 10:35:23PM +0000, Matt Arsenault wrote:
>> Also fix bug this exposed where when legalizing an immediate
>> operand, a v_mov_b32 would be created with a VSrc dest register.
>> 
>> http://reviews.llvm.org/D5134
>> 
>> Files:
>>  lib/Target/R600/SIInstrInfo.cpp
>>  lib/Target/R600/SIInstructions.td
>>  test/CodeGen/R600/codegen-prepare-addrmode-sext.ll
>>  test/CodeGen/R600/mul.ll
>>  test/CodeGen/R600/sign_extend.ll
> 
>> Index: lib/Target/R600/SIInstrInfo.cpp
>> ===================================================================
>> --- lib/Target/R600/SIInstrInfo.cpp
>> +++ lib/Target/R600/SIInstrInfo.cpp
>> @@ -908,6 +908,7 @@
>>   case AMDGPU::S_ADDC_U32: return AMDGPU::V_ADDC_U32_e32;
>>   case AMDGPU::S_SUB_I32: return AMDGPU::V_SUB_I32_e32;
>>   case AMDGPU::S_SUBB_U32: return AMDGPU::V_SUBB_U32_e32;
>> +  case AMDGPU::S_MUL_I32: return AMDGPU::V_MUL_LO_I32;
>>   case AMDGPU::S_AND_B32: return AMDGPU::V_AND_B32_e32;
>>   case AMDGPU::S_OR_B32: return AMDGPU::V_OR_B32_e32;
>>   case AMDGPU::S_XOR_B32: return AMDGPU::V_XOR_B32_e32;
>> @@ -981,10 +982,14 @@
>>   unsigned RCID = get(MI->getOpcode()).OpInfo[OpIdx].RegClass;
>>   const TargetRegisterClass *RC = RI.getRegClass(RCID);
>>   unsigned Opcode = AMDGPU::V_MOV_B32_e32;
>> +
>>   if (MO.isReg()) {
>>     Opcode = AMDGPU::COPY;
>>   } else if (RI.isSGPRClass(RC)) {
>>     Opcode = AMDGPU::S_MOV_B32;
>> +  } else if (MO.isImm()) {
>> +    if (RC == &AMDGPU::VSrc_32RegClass)
>> +      Opcode = AMDGPU::S_MOV_B32;
>>   }
> 
> I think I have a similar fix in a local series I'm working on.
> 
>> 
>>   const TargetRegisterClass *VRC = RI.getEquivalentVGPRClass(RC);
>> Index: lib/Target/R600/SIInstructions.td
>> ===================================================================
>> --- lib/Target/R600/SIInstructions.td
>> +++ lib/Target/R600/SIInstructions.td
>> @@ -274,11 +274,15 @@
>>   [(set i64:$dst, (sra i64:$src0, i32:$src1))]
>>> ;
>> 
>> -} // End AddedComplexity = 1
>> 
>> def S_BFM_B32 : SOP2_32 <0x00000024, "S_BFM_B32", []>;
>> def S_BFM_B64 : SOP2_64 <0x00000025, "S_BFM_B64", []>;
>> -def S_MUL_I32 : SOP2_32 <0x00000026, "S_MUL_I32", []>;
>> +def S_MUL_I32 : SOP2_32 <0x00000026, "S_MUL_I32",
>> +  [(set i32:$dst, (mul i32:$src0, i32:$src1))]
>> +>;
>> +
>> +} // End AddedComplexity = 1
>> +
>> def S_BFE_U32 : SOP2_32 <0x00000027, "S_BFE_U32", []>;
>> def S_BFE_I32 : SOP2_32 <0x00000028, "S_BFE_I32", []>;
>> def S_BFE_U64 : SOP2_64 <0x00000029, "S_BFE_U64", []>;
>> @@ -2499,11 +2503,6 @@
>> def : UMad24Pat<V_MAD_U32_U24>;
>> 
>> def : Pat <
>> -  (mul i32:$src0, i32:$src1),
>> -  (V_MUL_LO_I32 $src0, $src1)
>> ->;
>> -
>> -def : Pat <
>>   (mulhu i32:$src0, i32:$src1),
>>   (V_MUL_HI_U32 $src0, $src1)
>>> ;
>> @@ -2513,6 +2512,11 @@
>>   (V_MUL_HI_I32 $src0, $src1)
>>> ;
>> 
>> +def : Pat <
>> +  (mul i32:$src0, i32:$src1),
>> +  (V_MUL_LO_I32 $src0, $src1)
>> +>;
>> +
> 
> Did you mean to move this pattern?
> 

Not really. I deleted it at first and then re-added it to implement the branch workaround. I’ll remove this


> 
>> def : Vop3ModPat<V_MAD_F32, VOP_F32_F32_F32_F32, AMDGPUmad>;
>> 
>> 
>> Index: test/CodeGen/R600/codegen-prepare-addrmode-sext.ll
>> ===================================================================
>> --- test/CodeGen/R600/codegen-prepare-addrmode-sext.ll
>> +++ test/CodeGen/R600/codegen-prepare-addrmode-sext.ll
>> @@ -7,8 +7,8 @@
>> ; FUNC-LABEL: @test
>> ; OPT: mul nsw i32
>> ; OPT-NEXT: sext
>> -; SI-LLC: V_MUL_LO_I32
>> -; SI-LLC-NOT: V_MUL_HI
>> +; SI-LLC: S_MUL_I32
>> +; SI-LLC-NOT: MUL
>> define void @test(i8 addrspace(1)* nocapture readonly %in, i32 %a, i8 %b) {
>> entry:
>>   %0 = mul nsw i32 %a, 3
>> Index: test/CodeGen/R600/mul.ll
>> ===================================================================
>> --- test/CodeGen/R600/mul.ll
>> +++ test/CodeGen/R600/mul.ll
>> @@ -3,14 +3,14 @@
>> 
>> ; mul24 and mad24 are affected
>> 
>> -; FUNC-LABEL: @test2
>> +; FUNC-LABEL: @test_mul_v2i32
>> ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}}
>> ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}}
>> 
>> ; SI: V_MUL_LO_I32 v{{[0-9]+, v[0-9]+, v[0-9]+}}
>> ; SI: V_MUL_LO_I32 v{{[0-9]+, v[0-9]+, v[0-9]+}}
>> 
>> -define void @test2(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in) {
>> +define void @test_mul_v2i32(<2 x i32> addrspace(1)* %out, <2 x i32> addrspace(1)* %in) {
>>   %b_ptr = getelementptr <2 x i32> addrspace(1)* %in, i32 1
>>   %a = load <2 x i32> addrspace(1) * %in
>>   %b = load <2 x i32> addrspace(1) * %b_ptr
>> @@ -19,7 +19,7 @@
>>   ret void
>> }
>> 
>> -; FUNC-LABEL: @test4
>> +; FUNC-LABEL: @v_mul_v4i32
>> ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}}
>> ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}}
>> ; EG: MULLO_INT {{\*? *}}T{{[0-9]+\.[XYZW], T[0-9]+\.[XYZW], T[0-9]+\.[XYZW]}}
>> @@ -30,7 +30,7 @@
>> ; SI: V_MUL_LO_I32 v{{[0-9]+, v[0-9]+, v[0-9]+}}
>> ; SI: V_MUL_LO_I32 v{{[0-9]+, v[0-9]+, v[0-9]+}}
>> 
>> -define void @test4(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) {
>> +define void @v_mul_v4i32(<4 x i32> addrspace(1)* %out, <4 x i32> addrspace(1)* %in) {
>>   %b_ptr = getelementptr <4 x i32> addrspace(1)* %in, i32 1
>>   %a = load <4 x i32> addrspace(1) * %in
>>   %b = load <4 x i32> addrspace(1) * %b_ptr
>> @@ -39,12 +39,26 @@
>>   ret void
>> }
>> 
>> -; FUNC-LABEL: @trunc_i64_mul_to_i32
>> +; FUNC-LABEL: @s_trunc_i64_mul_to_i32
>> +; SI: S_LOAD_DWORD
>> +; SI: S_LOAD_DWORD
>> +; SI: S_MUL_I32
>> +; SI: BUFFER_STORE_DWORD
>> +define void @s_trunc_i64_mul_to_i32(i32 addrspace(1)* %out, i64 %a, i64 %b) {
>> +  %mul = mul i64 %b, %a
>> +  %trunc = trunc i64 %mul to i32
>> +  store i32 %trunc, i32 addrspace(1)* %out, align 8
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @v_trunc_i64_mul_to_i32
>> ; SI: S_LOAD_DWORD
>> ; SI: S_LOAD_DWORD
>> ; SI: V_MUL_LO_I32
>> ; SI: BUFFER_STORE_DWORD
>> -define void @trunc_i64_mul_to_i32(i32 addrspace(1)* %out, i64 %a, i64 %b) {
>> +define void @v_trunc_i64_mul_to_i32(i32 addrspace(1)* %out, i64 addrspace(1)* %aptr, i64 addrspace(1)* %bptr) nounwind {
>> +  %a = load i64 addrspace(1)* %aptr, align 8
>> +  %b = load i64 addrspace(1)* %bptr, align 8
>>   %mul = mul i64 %b, %a
>>   %trunc = trunc i64 %mul to i32
>>   store i32 %trunc, i32 addrspace(1)* %out, align 8
>> @@ -56,7 +70,7 @@
>> ; FUNC-LABEL: @mul64_sext_c
>> ; EG-DAG: MULLO_INT
>> ; EG-DAG: MULHI_INT
>> -; SI-DAG: V_MUL_LO_I32
>> +; SI-DAG: S_MUL_I32
>> ; SI-DAG: V_MUL_HI_I32
>> define void @mul64_sext_c(i64 addrspace(1)* %out, i32 %in) {
>> entry:
>> @@ -66,16 +80,120 @@
>>   ret void
>> }
>> 
>> +; FUNC-LABEL: @v_mul64_sext_c:
>> +; EG-DAG: MULLO_INT
>> +; EG-DAG: MULHI_INT
>> +; SI-DAG: V_MUL_LO_I32
>> +; SI-DAG: V_MUL_HI_I32
>> +; SI: S_ENDPGM
>> +define void @v_mul64_sext_c(i64 addrspace(1)* %out, i32 addrspace(1)* %in) {
>> +  %val = load i32 addrspace(1)* %in, align 4
>> +  %ext = sext i32 %val to i64
>> +  %mul = mul i64 %ext, 80
>> +  store i64 %mul, i64 addrspace(1)* %out, align 8
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @v_mul64_sext_inline_imm:
>> +; SI-DAG: V_MUL_LO_I32 v{{[0-9]+}}, 9, v{{[0-9]+}}
>> +; SI-DAG: V_MUL_HI_I32 v{{[0-9]+}}, 9, v{{[0-9]+}}
>> +; SI: S_ENDPGM
>> +define void @v_mul64_sext_inline_imm(i64 addrspace(1)* %out, i32 addrspace(1)* %in) {
>> +  %val = load i32 addrspace(1)* %in, align 4
>> +  %ext = sext i32 %val to i64
>> +  %mul = mul i64 %ext, 9
>> +  store i64 %mul, i64 addrspace(1)* %out, align 8
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @s_mul_i32:
>> +; SI: S_LOAD_DWORD [[SRC0:s[0-9]+]],
>> +; SI: S_LOAD_DWORD [[SRC1:s[0-9]+]],
>> +; SI: S_MUL_I32 [[SRESULT:s[0-9]+]], [[SRC0]], [[SRC1]]
>> +; SI: V_MOV_B32_e32 [[VRESULT:v[0-9]+]], [[SRESULT]]
>> +; SI: BUFFER_STORE_DWORD [[VRESULT]],
>> +; SI: S_ENDPGM
>> +define void @s_mul_i32(i32 addrspace(1)* %out, i32 %a, i32 %b) nounwind {
>> +  %mul = mul i32 %a, %b
>> +  store i32 %mul, i32 addrspace(1)* %out, align 4
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @v_mul_i32
>> +; SI: V_MUL_LO_I32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}
>> +define void @v_mul_i32(i32 addrspace(1)* %out, i32 addrspace(1)* %in) {
>> +  %b_ptr = getelementptr i32 addrspace(1)* %in, i32 1
>> +  %a = load i32 addrspace(1)* %in
>> +  %b = load i32 addrspace(1)* %b_ptr
>> +  %result = mul i32 %a, %b
>> +  store i32 %result, i32 addrspace(1)* %out
>> +  ret void
>> +}
>> +
>> ; A standard 64-bit multiply.  The expansion should be around 6 instructions.
>> ; It would be difficult to match the expansion correctly without writing
>> ; a really complicated list of FileCheck expressions.  I don't want
>> ; to confuse people who may 'break' this test with a correct optimization,
>> ; so this test just uses FUNC-LABEL to make sure the compiler does not
>> ; crash with a 'failed to select' error.
>> -; FUNC-LABEL: @mul64
>> -define void @mul64(i64 addrspace(1)* %out, i64 %a, i64 %b) {
>> +
>> +; FUNC-LABEL: @s_mul_i64:
>> +define void @s_mul_i64(i64 addrspace(1)* %out, i64 %a, i64 %b) nounwind {
>> +  %mul = mul i64 %a, %b
>> +  store i64 %mul, i64 addrspace(1)* %out, align 8
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @v_mul_i64
>> +; SI: V_MUL_LO_I32
>> +define void @v_mul_i64(i64 addrspace(1)* %out, i64 addrspace(1)* %aptr, i64 addrspace(1)* %bptr) {
>> +  %a = load i64 addrspace(1)* %aptr, align 8
>> +  %b = load i64 addrspace(1)* %bptr, align 8
>> +  %mul = mul i64 %a, %b
>> +  store i64 %mul, i64 addrspace(1)* %out, align 8
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @mul32_in_branch
>> +; SI: V_MUL_LO_I32
>> +define void @mul32_in_branch(i32 addrspace(1)* %out, i32 addrspace(1)* %in, i32 %a, i32 %b, i32 %c) {
>> entry:
>> -  %0 = mul i64 %a, %b
>> -  store i64 %0, i64 addrspace(1)* %out
>> +  %0 = icmp eq i32 %a, 0
>> +  br i1 %0, label %if, label %else
>> +
>> +if:
>> +  %1 = load i32 addrspace(1)* %in
>> +  br label %endif
>> +
>> +else:
>> +  %2 = mul i32 %a, %b
>> +  br label %endif
>> +
>> +endif:
>> +  %3 = phi i32 [%1, %if], [%2, %else]
>> +  store i32 %3, i32 addrspace(1)* %out
>> +  ret void
>> +}
>> +
>> +; FUNC-LABEL: @mul64_in_branch
>> +; SI-DAG: V_MUL_LO_I32
>> +; SI-DAG: V_MUL_HI_U32
>> +; SI: S_ENDPGM
>> +define void @mul64_in_branch(i64 addrspace(1)* %out, i64 addrspace(1)* %in, i64 %a, i64 %b, i64 %c) {
>> +entry:
>> +  %0 = icmp eq i64 %a, 0
>> +  br i1 %0, label %if, label %else
>> +
>> +if:
>> +  %1 = load i64 addrspace(1)* %in
>> +  br label %endif
>> +
>> +else:
>> +  %2 = mul i64 %a, %b
>> +  br label %endif
>> +
>> +endif:
>> +  %3 = phi i64 [%1, %if], [%2, %else]
>> +  store i64 %3, i64 addrspace(1)* %out
>>   ret void
>> }
>> Index: test/CodeGen/R600/sign_extend.ll
>> ===================================================================
>> --- test/CodeGen/R600/sign_extend.ll
>> +++ test/CodeGen/R600/sign_extend.ll
>> @@ -10,10 +10,10 @@
>>   ret void
>> }
>> 
>> -; SI-LABEL: @test:
>> -; SI: V_ASHR
>> +; SI-LABEL: @test_s_sext_i32_to_i64:
>> +; SI: S_ASHR_I32
>> ; SI: S_ENDPG
>> -define void @test(i64 addrspace(1)* %out, i32 %a, i32 %b, i32 %c) nounwind {
>> +define void @test_s_sext_i32_to_i64(i64 addrspace(1)* %out, i32 %a, i32 %b, i32 %c) nounwind {
>> entry:
>>   %mul = mul i32 %a, %b
>>   %add = add i32 %mul, %c
> 
>> _______________________________________________
>> llvm-commits mailing list
>> llvm-commits at cs.uiuc.edu
>> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
> 
> _______________________________________________
> llvm-commits mailing list
> llvm-commits at cs.uiuc.edu
> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits





More information about the llvm-commits mailing list