[llvm] r361543 - UpdateTestChecks: ppc32 triple support

Roman Lebedev via llvm-commits llvm-commits at lists.llvm.org
Thu May 23 12:54:41 PDT 2019


Author: lebedevri
Date: Thu May 23 12:54:41 2019
New Revision: 361543

URL: http://llvm.org/viewvc/llvm-project?rev=361543&view=rev
Log:
UpdateTestChecks: ppc32 triple support

Summary:
Appears identical to powerpc64{,le}.
Regenerate test that is being affected by upcoming patch.

Reviewers: RKSimon

Reviewed By: RKSimon

Subscribers: nemanjai, jsji, llvm-commits

Tags: #llvm

Differential Revision: https://reviews.llvm.org/D62339

Modified:
    llvm/trunk/test/CodeGen/PowerPC/vec_splat.ll
    llvm/trunk/utils/UpdateTestChecks/asm.py

Modified: llvm/trunk/test/CodeGen/PowerPC/vec_splat.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/PowerPC/vec_splat.ll?rev=361543&r1=361542&r2=361543&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/PowerPC/vec_splat.ll (original)
+++ llvm/trunk/test/CodeGen/PowerPC/vec_splat.ll Thu May 23 12:54:41 2019
@@ -1,71 +1,261 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -verify-machineinstrs < %s -mtriple=ppc32-- -mcpu=g3 | FileCheck %s --check-prefixes=ALL,G3
+; RUN: llc -verify-machineinstrs < %s -mtriple=ppc32-- -mcpu=g5 | FileCheck %s --check-prefixes=ALL,G5
+
 ; Test that vectors are scalarized/lowered correctly.
-; RUN: llc -verify-machineinstrs < %s -mtriple=ppc32-- -mcpu=g3 | \
-; RUN:    grep stfs | count 4
-; RUN: llc -verify-machineinstrs < %s -mtriple=ppc32-- -mcpu=g5 -o %t
-; RUN: grep vspltw %t | count 2
-; RUN: grep vsplti %t | count 3
-; RUN: grep vsplth %t | count 1
 
-        %f4 = type <4 x float>
-        %i4 = type <4 x i32>
+%f4 = type <4 x float>
+%i4 = type <4 x i32>
 
 define void @splat(%f4* %P, %f4* %Q, float %X) nounwind {
-        %tmp = insertelement %f4 undef, float %X, i32 0         ; <%f4> [#uses=1]
-        %tmp2 = insertelement %f4 %tmp, float %X, i32 1         ; <%f4> [#uses=1]
-        %tmp4 = insertelement %f4 %tmp2, float %X, i32 2                ; <%f4> [#uses=1]
-        %tmp6 = insertelement %f4 %tmp4, float %X, i32 3                ; <%f4> [#uses=1]
-        %q = load %f4, %f4* %Q               ; <%f4> [#uses=1]
-        %R = fadd %f4 %q, %tmp6          ; <%f4> [#uses=1]
-        store %f4 %R, %f4* %P
-        ret void
+; G3-LABEL: splat:
+; G3:       # %bb.0:
+; G3-NEXT:    lfs 0, 0(4)
+; G3-NEXT:    lfs 2, 8(4)
+; G3-NEXT:    lfs 3, 4(4)
+; G3-NEXT:    lfs 4, 12(4)
+; G3-NEXT:    fadds 0, 0, 1
+; G3-NEXT:    fadds 2, 2, 1
+; G3-NEXT:    fadds 3, 3, 1
+; G3-NEXT:    fadds 1, 4, 1
+; G3-NEXT:    stfs 1, 12(3)
+; G3-NEXT:    stfs 2, 8(3)
+; G3-NEXT:    stfs 3, 4(3)
+; G3-NEXT:    stfs 0, 0(3)
+; G3-NEXT:    blr
+;
+; G5-LABEL: splat:
+; G5:       # %bb.0:
+; G5-NEXT:    stwu 1, -32(1)
+; G5-NEXT:    stfs 1, 16(1)
+; G5-NEXT:    addi 5, 1, 16
+; G5-NEXT:    lvx 2, 0, 5
+; G5-NEXT:    lvx 3, 0, 4
+; G5-NEXT:    vspltw 2, 2, 0
+; G5-NEXT:    vaddfp 2, 3, 2
+; G5-NEXT:    stvx 2, 0, 3
+; G5-NEXT:    addi 1, 1, 32
+; G5-NEXT:    blr
+  %tmp = insertelement %f4 undef, float %X, i32 0   ; <%f4> [#uses=1]
+  %tmp2 = insertelement %f4 %tmp, float %X, i32 1   ; <%f4> [#uses=1]
+  %tmp4 = insertelement %f4 %tmp2, float %X, i32 2    ; <%f4> [#uses=1]
+  %tmp6 = insertelement %f4 %tmp4, float %X, i32 3    ; <%f4> [#uses=1]
+  %q = load %f4, %f4* %Q         ; <%f4> [#uses=1]
+  %R = fadd %f4 %q, %tmp6    ; <%f4> [#uses=1]
+  store %f4 %R, %f4* %P
+  ret void
 }
 
 define void @splat_i4(%i4* %P, %i4* %Q, i32 %X) nounwind {
-        %tmp = insertelement %i4 undef, i32 %X, i32 0           ; <%i4> [#uses=1]
-        %tmp2 = insertelement %i4 %tmp, i32 %X, i32 1           ; <%i4> [#uses=1]
-        %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2          ; <%i4> [#uses=1]
-        %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3          ; <%i4> [#uses=1]
-        %q = load %i4, %i4* %Q               ; <%i4> [#uses=1]
-        %R = add %i4 %q, %tmp6          ; <%i4> [#uses=1]
-        store %i4 %R, %i4* %P
-        ret void
+; G3-LABEL: splat_i4:
+; G3:       # %bb.0:
+; G3-NEXT:    lwz 6, 0(4)
+; G3-NEXT:    lwz 7, 8(4)
+; G3-NEXT:    lwz 8, 4(4)
+; G3-NEXT:    lwz 4, 12(4)
+; G3-NEXT:    add 6, 6, 5
+; G3-NEXT:    add 8, 8, 5
+; G3-NEXT:    add 7, 7, 5
+; G3-NEXT:    add 4, 4, 5
+; G3-NEXT:    stw 4, 12(3)
+; G3-NEXT:    stw 7, 8(3)
+; G3-NEXT:    stw 8, 4(3)
+; G3-NEXT:    stw 6, 0(3)
+; G3-NEXT:    blr
+;
+; G5-LABEL: splat_i4:
+; G5:       # %bb.0:
+; G5-NEXT:    stwu 1, -32(1)
+; G5-NEXT:    stw 5, 16(1)
+; G5-NEXT:    addi 5, 1, 16
+; G5-NEXT:    lvx 2, 0, 5
+; G5-NEXT:    lvx 3, 0, 4
+; G5-NEXT:    vspltw 2, 2, 0
+; G5-NEXT:    vadduwm 2, 3, 2
+; G5-NEXT:    stvx 2, 0, 3
+; G5-NEXT:    addi 1, 1, 32
+; G5-NEXT:    blr
+  %tmp = insertelement %i4 undef, i32 %X, i32 0     ; <%i4> [#uses=1]
+  %tmp2 = insertelement %i4 %tmp, i32 %X, i32 1     ; <%i4> [#uses=1]
+  %tmp4 = insertelement %i4 %tmp2, i32 %X, i32 2    ; <%i4> [#uses=1]
+  %tmp6 = insertelement %i4 %tmp4, i32 %X, i32 3    ; <%i4> [#uses=1]
+  %q = load %i4, %i4* %Q         ; <%i4> [#uses=1]
+  %R = add %i4 %q, %tmp6    ; <%i4> [#uses=1]
+  store %i4 %R, %i4* %P
+  ret void
 }
 
 define void @splat_imm_i32(%i4* %P, %i4* %Q, i32 %X) nounwind {
-        %q = load %i4, %i4* %Q               ; <%i4> [#uses=1]
-        %R = add %i4 %q, < i32 -1, i32 -1, i32 -1, i32 -1 >             ; <%i4> [#uses=1]
-        store %i4 %R, %i4* %P
-        ret void
+; G3-LABEL: splat_imm_i32:
+; G3:       # %bb.0:
+; G3-NEXT:    lwz 5, 0(4)
+; G3-NEXT:    lwz 6, 8(4)
+; G3-NEXT:    lwz 7, 4(4)
+; G3-NEXT:    lwz 4, 12(4)
+; G3-NEXT:    addi 5, 5, -1
+; G3-NEXT:    addi 7, 7, -1
+; G3-NEXT:    addi 6, 6, -1
+; G3-NEXT:    addi 4, 4, -1
+; G3-NEXT:    stw 4, 12(3)
+; G3-NEXT:    stw 6, 8(3)
+; G3-NEXT:    stw 7, 4(3)
+; G3-NEXT:    stw 5, 0(3)
+; G3-NEXT:    blr
+;
+; G5-LABEL: splat_imm_i32:
+; G5:       # %bb.0:
+; G5-NEXT:    lvx 2, 0, 4
+; G5-NEXT:    vspltisb 3, -1
+; G5-NEXT:    vadduwm 2, 2, 3
+; G5-NEXT:    stvx 2, 0, 3
+; G5-NEXT:    blr
+  %q = load %i4, %i4* %Q         ; <%i4> [#uses=1]
+  %R = add %i4 %q, < i32 -1, i32 -1, i32 -1, i32 -1 >       ; <%i4> [#uses=1]
+  store %i4 %R, %i4* %P
+  ret void
 }
 
 define void @splat_imm_i16(%i4* %P, %i4* %Q, i32 %X) nounwind {
-        %q = load %i4, %i4* %Q               ; <%i4> [#uses=1]
-        %R = add %i4 %q, < i32 65537, i32 65537, i32 65537, i32 65537 >         ; <%i4> [#uses=1]
-        store %i4 %R, %i4* %P
-        ret void
+; G3-LABEL: splat_imm_i16:
+; G3:       # %bb.0:
+; G3-NEXT:    lwz 5, 0(4)
+; G3-NEXT:    lwz 6, 8(4)
+; G3-NEXT:    lwz 7, 4(4)
+; G3-NEXT:    lwz 4, 12(4)
+; G3-NEXT:    addi 5, 5, 1
+; G3-NEXT:    addi 7, 7, 1
+; G3-NEXT:    addi 6, 6, 1
+; G3-NEXT:    addi 4, 4, 1
+; G3-NEXT:    addis 5, 5, 1
+; G3-NEXT:    addis 7, 7, 1
+; G3-NEXT:    addis 6, 6, 1
+; G3-NEXT:    addis 4, 4, 1
+; G3-NEXT:    stw 4, 12(3)
+; G3-NEXT:    stw 6, 8(3)
+; G3-NEXT:    stw 7, 4(3)
+; G3-NEXT:    stw 5, 0(3)
+; G3-NEXT:    blr
+;
+; G5-LABEL: splat_imm_i16:
+; G5:       # %bb.0:
+; G5-NEXT:    lvx 2, 0, 4
+; G5-NEXT:    vspltish 3, 1
+; G5-NEXT:    vadduwm 2, 2, 3
+; G5-NEXT:    stvx 2, 0, 3
+; G5-NEXT:    blr
+  %q = load %i4, %i4* %Q         ; <%i4> [#uses=1]
+  %R = add %i4 %q, < i32 65537, i32 65537, i32 65537, i32 65537 >   ; <%i4> [#uses=1]
+  store %i4 %R, %i4* %P
+  ret void
 }
 
 define void @splat_h(i16 %tmp, <16 x i8>* %dst) nounwind {
-        %tmp.upgrd.1 = insertelement <8 x i16> undef, i16 %tmp, i32 0           
-        %tmp72 = insertelement <8 x i16> %tmp.upgrd.1, i16 %tmp, i32 1 
-        %tmp73 = insertelement <8 x i16> %tmp72, i16 %tmp, i32 2 
-        %tmp74 = insertelement <8 x i16> %tmp73, i16 %tmp, i32 3
-        %tmp75 = insertelement <8 x i16> %tmp74, i16 %tmp, i32 4 
-        %tmp76 = insertelement <8 x i16> %tmp75, i16 %tmp, i32 5
-        %tmp77 = insertelement <8 x i16> %tmp76, i16 %tmp, i32 6 
-        %tmp78 = insertelement <8 x i16> %tmp77, i16 %tmp, i32 7 
-        %tmp78.upgrd.2 = bitcast <8 x i16> %tmp78 to <16 x i8>  
-        store <16 x i8> %tmp78.upgrd.2, <16 x i8>* %dst
-        ret void
+; G3-LABEL: splat_h:
+; G3:       # %bb.0:
+; G3-NEXT:    sth 3, 14(4)
+; G3-NEXT:    sth 3, 12(4)
+; G3-NEXT:    sth 3, 10(4)
+; G3-NEXT:    sth 3, 8(4)
+; G3-NEXT:    sth 3, 6(4)
+; G3-NEXT:    sth 3, 4(4)
+; G3-NEXT:    sth 3, 2(4)
+; G3-NEXT:    sth 3, 0(4)
+; G3-NEXT:    blr
+;
+; G5-LABEL: splat_h:
+; G5:       # %bb.0:
+; G5-NEXT:    stwu 1, -32(1)
+; G5-NEXT:    sth 3, 16(1)
+; G5-NEXT:    addi 3, 1, 16
+; G5-NEXT:    lvx 2, 0, 3
+; G5-NEXT:    vsplth 2, 2, 0
+; G5-NEXT:    stvx 2, 0, 4
+; G5-NEXT:    addi 1, 1, 32
+; G5-NEXT:    blr
+  %tmp.upgrd.1 = insertelement <8 x i16> undef, i16 %tmp, i32 0
+  %tmp72 = insertelement <8 x i16> %tmp.upgrd.1, i16 %tmp, i32 1
+  %tmp73 = insertelement <8 x i16> %tmp72, i16 %tmp, i32 2
+  %tmp74 = insertelement <8 x i16> %tmp73, i16 %tmp, i32 3
+  %tmp75 = insertelement <8 x i16> %tmp74, i16 %tmp, i32 4
+  %tmp76 = insertelement <8 x i16> %tmp75, i16 %tmp, i32 5
+  %tmp77 = insertelement <8 x i16> %tmp76, i16 %tmp, i32 6
+  %tmp78 = insertelement <8 x i16> %tmp77, i16 %tmp, i32 7
+  %tmp78.upgrd.2 = bitcast <8 x i16> %tmp78 to <16 x i8>
+  store <16 x i8> %tmp78.upgrd.2, <16 x i8>* %dst
+  ret void
 }
 
 define void @spltish(<16 x i8>* %A, <16 x i8>* %B) nounwind {
-        %tmp = load <16 x i8>, <16 x i8>* %B               ; <<16 x i8>> [#uses=1]
-        %tmp.s = bitcast <16 x i8> %tmp to <16 x i8>            ; <<16 x i8>> [#uses=1]
-        %tmp4 = sub <16 x i8> %tmp.s, bitcast (<8 x i16> < i16 15, i16 15, i16 15, i16 15, i16 15, i16
- 15, i16 15, i16 15 > to <16 x i8>)             ; <<16 x i8>> [#uses=1]
-        %tmp4.u = bitcast <16 x i8> %tmp4 to <16 x i8>          ; <<16 x i8>> [#uses=1]
-        store <16 x i8> %tmp4.u, <16 x i8>* %A
-        ret void
+; G3-LABEL: spltish:
+; G3:       # %bb.0:
+; G3-NEXT:    stwu 1, -48(1)
+; G3-NEXT:    stw 25, 20(1) # 4-byte Folded Spill
+; G3-NEXT:    stw 26, 24(1) # 4-byte Folded Spill
+; G3-NEXT:    stw 27, 28(1) # 4-byte Folded Spill
+; G3-NEXT:    stw 28, 32(1) # 4-byte Folded Spill
+; G3-NEXT:    stw 29, 36(1) # 4-byte Folded Spill
+; G3-NEXT:    stw 30, 40(1) # 4-byte Folded Spill
+; G3-NEXT:    lbz 5, 5(4)
+; G3-NEXT:    lbz 6, 3(4)
+; G3-NEXT:    lbz 7, 1(4)
+; G3-NEXT:    lbz 8, 0(4)
+; G3-NEXT:    lbz 9, 2(4)
+; G3-NEXT:    lbz 10, 4(4)
+; G3-NEXT:    lbz 11, 6(4)
+; G3-NEXT:    lbz 12, 8(4)
+; G3-NEXT:    lbz 0, 10(4)
+; G3-NEXT:    addi 7, 7, -15
+; G3-NEXT:    lbz 30, 12(4)
+; G3-NEXT:    lbz 29, 14(4)
+; G3-NEXT:    lbz 28, 15(4)
+; G3-NEXT:    lbz 27, 13(4)
+; G3-NEXT:    lbz 26, 11(4)
+; G3-NEXT:    lbz 25, 9(4)
+; G3-NEXT:    addi 6, 6, -15
+; G3-NEXT:    lbz 4, 7(4)
+; G3-NEXT:    addi 5, 5, -15
+; G3-NEXT:    addi 25, 25, -15
+; G3-NEXT:    addi 26, 26, -15
+; G3-NEXT:    addi 4, 4, -15
+; G3-NEXT:    addi 27, 27, -15
+; G3-NEXT:    addi 28, 28, -15
+; G3-NEXT:    stb 29, 14(3)
+; G3-NEXT:    stb 30, 12(3)
+; G3-NEXT:    stb 0, 10(3)
+; G3-NEXT:    stb 12, 8(3)
+; G3-NEXT:    stb 11, 6(3)
+; G3-NEXT:    stb 10, 4(3)
+; G3-NEXT:    stb 9, 2(3)
+; G3-NEXT:    stb 8, 0(3)
+; G3-NEXT:    stb 28, 15(3)
+; G3-NEXT:    stb 27, 13(3)
+; G3-NEXT:    stb 26, 11(3)
+; G3-NEXT:    stb 25, 9(3)
+; G3-NEXT:    stb 4, 7(3)
+; G3-NEXT:    stb 5, 5(3)
+; G3-NEXT:    stb 6, 3(3)
+; G3-NEXT:    stb 7, 1(3)
+; G3-NEXT:    lwz 30, 40(1) # 4-byte Folded Reload
+; G3-NEXT:    lwz 29, 36(1) # 4-byte Folded Reload
+; G3-NEXT:    lwz 28, 32(1) # 4-byte Folded Reload
+; G3-NEXT:    lwz 27, 28(1) # 4-byte Folded Reload
+; G3-NEXT:    lwz 26, 24(1) # 4-byte Folded Reload
+; G3-NEXT:    lwz 25, 20(1) # 4-byte Folded Reload
+; G3-NEXT:    addi 1, 1, 48
+; G3-NEXT:    blr
+;
+; G5-LABEL: spltish:
+; G5:       # %bb.0:
+; G5-NEXT:    lvx 2, 0, 4
+; G5-NEXT:    vspltish 3, 15
+; G5-NEXT:    vsububm 2, 2, 3
+; G5-NEXT:    stvx 2, 0, 3
+; G5-NEXT:    blr
+  %tmp = load <16 x i8>, <16 x i8>* %B         ; <<16 x i8>> [#uses=1]
+  %tmp.s = bitcast <16 x i8> %tmp to <16 x i8>      ; <<16 x i8>> [#uses=1]
+  %tmp4 = sub <16 x i8> %tmp.s, bitcast (<8 x i16> < i16 15, i16 15, i16 15, i16 15, i16 15, i16
+ 15, i16 15, i16 15 > to <16 x i8>)       ; <<16 x i8>> [#uses=1]
+  %tmp4.u = bitcast <16 x i8> %tmp4 to <16 x i8>    ; <<16 x i8>> [#uses=1]
+  store <16 x i8> %tmp4.u, <16 x i8>* %A
+  ret void
 }
-

Modified: llvm/trunk/utils/UpdateTestChecks/asm.py
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/utils/UpdateTestChecks/asm.py?rev=361543&r1=361542&r2=361543&view=diff
==============================================================================
--- llvm/trunk/utils/UpdateTestChecks/asm.py (original)
+++ llvm/trunk/utils/UpdateTestChecks/asm.py Thu May 23 12:54:41 2019
@@ -154,7 +154,7 @@ def scrub_asm_arm_eabi(asm, args):
   asm = common.SCRUB_TRAILING_WHITESPACE_RE.sub(r'', asm)
   return asm
 
-def scrub_asm_powerpc64(asm, args):
+def scrub_asm_powerpc(asm, args):
   # Scrub runs of whitespace out of the assembly, but leave the leading
   # whitespace in place.
   asm = common.SCRUB_WHITESPACE_RE.sub(r' ', asm)
@@ -261,8 +261,9 @@ def build_function_body_dictionary_for_t
       'armv7eb': (scrub_asm_arm_eabi, ASM_FUNCTION_ARM_RE),
       'armv8a': (scrub_asm_arm_eabi, ASM_FUNCTION_ARM_RE),
       'mips': (scrub_asm_mips, ASM_FUNCTION_MIPS_RE),
-      'powerpc64': (scrub_asm_powerpc64, ASM_FUNCTION_PPC_RE),
-      'powerpc64le': (scrub_asm_powerpc64, ASM_FUNCTION_PPC_RE),
+      'ppc32': (scrub_asm_powerpc, ASM_FUNCTION_PPC_RE),
+      'powerpc64': (scrub_asm_powerpc, ASM_FUNCTION_PPC_RE),
+      'powerpc64le': (scrub_asm_powerpc, ASM_FUNCTION_PPC_RE),
       'riscv32': (scrub_asm_riscv, ASM_FUNCTION_RISCV_RE),
       'riscv64': (scrub_asm_riscv, ASM_FUNCTION_RISCV_RE),
       'sparc': (scrub_asm_sparc, ASM_FUNCTION_SPARC_RE),




More information about the llvm-commits mailing list