[llvm] [NFC][AMDGPU] Automate any_extend_vector_inreg.ll check line generation (PR #145013)

Chris Jackson via llvm-commits llvm-commits at lists.llvm.org
Fri Jun 20 04:06:00 PDT 2025


https://github.com/chrisjbris updated https://github.com/llvm/llvm-project/pull/145013

>From 5d0e544f788e295444ec1adfa8ad6ba7cd471b61 Mon Sep 17 00:00:00 2001
From: Chris Jackson <chris.jackson at amd.com>
Date: Fri, 20 Jun 2025 05:42:36 -0500
Subject: [PATCH 1/2] [NFC][AMDGPU] Automate any_extend_vector_inreg.ll check
 line generation

Convert the test to use update_llc_test_checks.py.
---
 .../CodeGen/AMDGPU/any_extend_vector_inreg.ll | 165 +++++++++++++++---
 1 file changed, 142 insertions(+), 23 deletions(-)

diff --git a/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll b/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
index 8bcef24c8e23d..ce53f1e460262 100644
--- a/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
+++ b/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
@@ -1,30 +1,149 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
 ; RUN: llc -mtriple=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s
-; RUN: llc -mtriple=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s
+; RUN: llc -mtriple=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck -check-prefix=GCNF %s
 
-; GCN-LABEL: {{^}}any_extend_vector_inreg_v16i8_to_v4i32:
-; GCN: s_load_dwordx8
-; GCN-DAG: s_load_dword
 
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
-; GCN: {{buffer|flat}}_store_byte
 define amdgpu_kernel void @any_extend_vector_inreg_v16i8_to_v4i32(ptr addrspace(1) nocapture readonly %arg, ptr addrspace(1) %arg1) local_unnamed_addr #0 {
+; GCN-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
+; GCN:       ; %bb.0: ; %bb
+; GCN-NEXT:    s_load_dwordx4 s[12:15], s[4:5], 0x9
+; GCN-NEXT:    s_mov_b32 s3, 0xf000
+; GCN-NEXT:    s_mov_b32 s2, -1
+; GCN-NEXT:    v_mov_b32_e32 v0, 0
+; GCN-NEXT:    s_waitcnt lgkmcnt(0)
+; GCN-NEXT:    s_mov_b32 s0, s14
+; GCN-NEXT:    s_mov_b32 s1, s15
+; GCN-NEXT:    s_load_dwordx8 s[4:11], s[12:13], 0x0
+; GCN-NEXT:    s_waitcnt lgkmcnt(0)
+; GCN-NEXT:    s_load_dword s4, s[12:13], 0x8
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:13
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:15
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:14
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:8
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:11
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:10
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:4
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:6
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:1
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:3
+; GCN-NEXT:    s_lshr_b32 s8, s9, 16
+; GCN-NEXT:    s_waitcnt expcnt(0)
+; GCN-NEXT:    v_mov_b32_e32 v0, s6
+; GCN-NEXT:    s_waitcnt lgkmcnt(0)
+; GCN-NEXT:    s_lshl_b64 s[6:7], s[4:5], 8
+; GCN-NEXT:    v_mov_b32_e32 v1, s11
+; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:9
+; GCN-NEXT:    s_waitcnt expcnt(0)
+; GCN-NEXT:    v_mov_b32_e32 v1, s5
+; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:2
+; GCN-NEXT:    v_alignbit_b32 v0, s8, v0, 16
+; GCN-NEXT:    s_waitcnt expcnt(0)
+; GCN-NEXT:    v_mov_b32_e32 v1, s7
+; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:12
+; GCN-NEXT:    s_waitcnt expcnt(0)
+; GCN-NEXT:    v_lshrrev_b32_e32 v1, 8, v0
+; GCN-NEXT:    v_lshrrev_b32_e32 v0, 24, v0
+; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:5
+; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:7
+; GCN-NEXT:    s_endpgm
+;
+; GCNF-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
+; GCNF:       ; %bb.0: ; %bb
+; GCNF-NEXT:    s_load_dwordx4 s[8:11], s[4:5], 0x24
+; GCNF-NEXT:    v_mov_b32_e32 v2, 0
+; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
+; GCNF-NEXT:    s_load_dwordx8 s[0:7], s[8:9], 0x0
+; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
+; GCNF-NEXT:    s_load_dword s0, s[8:9], 0x20
+; GCNF-NEXT:    s_lshr_b32 s6, s5, 24
+; GCNF-NEXT:    s_lshr_b32 s8, s2, 24
+; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
+; GCNF-NEXT:    s_lshl_b64 s[2:3], s[0:1], 8
+; GCNF-NEXT:    s_add_u32 s4, s10, 13
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 15
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 14
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 8
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 11
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 10
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 4
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 6
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 1
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    v_mov_b32_e32 v0, s10
+; GCNF-NEXT:    v_mov_b32_e32 v1, s11
+; GCNF-NEXT:    s_add_u32 s4, s10, 3
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    s_add_u32 s4, s10, 9
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    v_mov_b32_e32 v2, s7
+; GCNF-NEXT:    s_add_u32 s4, s10, 2
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s5, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s4
+; GCNF-NEXT:    v_mov_b32_e32 v1, s5
+; GCNF-NEXT:    v_mov_b32_e32 v2, s1
+; GCNF-NEXT:    s_add_u32 s0, s10, 5
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s1, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s0
+; GCNF-NEXT:    v_mov_b32_e32 v1, s1
+; GCNF-NEXT:    v_mov_b32_e32 v2, s8
+; GCNF-NEXT:    s_add_u32 s0, s10, 12
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s1, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s0
+; GCNF-NEXT:    v_mov_b32_e32 v1, s1
+; GCNF-NEXT:    v_mov_b32_e32 v2, s3
+; GCNF-NEXT:    s_add_u32 s0, s10, 7
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_addc_u32 s1, s11, 0
+; GCNF-NEXT:    v_mov_b32_e32 v0, s0
+; GCNF-NEXT:    v_mov_b32_e32 v1, s1
+; GCNF-NEXT:    v_mov_b32_e32 v2, s6
+; GCNF-NEXT:    flat_store_byte v[0:1], v2
+; GCNF-NEXT:    s_endpgm
 bb:
   %tmp2 = load <16 x i8>, ptr addrspace(1) %arg, align 16
   %tmp3 = extractelement <16 x i8> %tmp2, i64 4

>From a5a87e351223ace49adf7febe84c7be786a67870 Mon Sep 17 00:00:00 2001
From: Chris Jackson <chris.jackson at amd.com>
Date: Fri, 20 Jun 2025 06:05:36 -0500
Subject: [PATCH 2/2] Modify prefixes to match AMDGPU convention.

---
 .../CodeGen/AMDGPU/any_extend_vector_inreg.ll | 282 +++++++++---------
 1 file changed, 141 insertions(+), 141 deletions(-)

diff --git a/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll b/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
index ce53f1e460262..cc9f595f9d0b6 100644
--- a/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
+++ b/llvm/test/CodeGen/AMDGPU/any_extend_vector_inreg.ll
@@ -1,149 +1,149 @@
 ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5
-; RUN: llc -mtriple=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s
-; RUN: llc -mtriple=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck -check-prefix=GCNF %s
+; RUN: llc -mtriple=amdgcn -verify-machineinstrs < %s | FileCheck -check-prefix=GFX6 %s
+; RUN: llc -mtriple=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck -check-prefix=GFX8 %s
 
 
 define amdgpu_kernel void @any_extend_vector_inreg_v16i8_to_v4i32(ptr addrspace(1) nocapture readonly %arg, ptr addrspace(1) %arg1) local_unnamed_addr #0 {
-; GCN-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
-; GCN:       ; %bb.0: ; %bb
-; GCN-NEXT:    s_load_dwordx4 s[12:15], s[4:5], 0x9
-; GCN-NEXT:    s_mov_b32 s3, 0xf000
-; GCN-NEXT:    s_mov_b32 s2, -1
-; GCN-NEXT:    v_mov_b32_e32 v0, 0
-; GCN-NEXT:    s_waitcnt lgkmcnt(0)
-; GCN-NEXT:    s_mov_b32 s0, s14
-; GCN-NEXT:    s_mov_b32 s1, s15
-; GCN-NEXT:    s_load_dwordx8 s[4:11], s[12:13], 0x0
-; GCN-NEXT:    s_waitcnt lgkmcnt(0)
-; GCN-NEXT:    s_load_dword s4, s[12:13], 0x8
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:13
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:15
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:14
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:8
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:11
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:10
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:4
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:6
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:1
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:3
-; GCN-NEXT:    s_lshr_b32 s8, s9, 16
-; GCN-NEXT:    s_waitcnt expcnt(0)
-; GCN-NEXT:    v_mov_b32_e32 v0, s6
-; GCN-NEXT:    s_waitcnt lgkmcnt(0)
-; GCN-NEXT:    s_lshl_b64 s[6:7], s[4:5], 8
-; GCN-NEXT:    v_mov_b32_e32 v1, s11
-; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:9
-; GCN-NEXT:    s_waitcnt expcnt(0)
-; GCN-NEXT:    v_mov_b32_e32 v1, s5
-; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:2
-; GCN-NEXT:    v_alignbit_b32 v0, s8, v0, 16
-; GCN-NEXT:    s_waitcnt expcnt(0)
-; GCN-NEXT:    v_mov_b32_e32 v1, s7
-; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:12
-; GCN-NEXT:    s_waitcnt expcnt(0)
-; GCN-NEXT:    v_lshrrev_b32_e32 v1, 8, v0
-; GCN-NEXT:    v_lshrrev_b32_e32 v0, 24, v0
-; GCN-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:5
-; GCN-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:7
-; GCN-NEXT:    s_endpgm
+; GFX6-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
+; GFX6:       ; %bb.0: ; %bb
+; GFX6-NEXT:    s_load_dwordx4 s[12:15], s[4:5], 0x9
+; GFX6-NEXT:    s_mov_b32 s3, 0xf000
+; GFX6-NEXT:    s_mov_b32 s2, -1
+; GFX6-NEXT:    v_mov_b32_e32 v0, 0
+; GFX6-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX6-NEXT:    s_mov_b32 s0, s14
+; GFX6-NEXT:    s_mov_b32 s1, s15
+; GFX6-NEXT:    s_load_dwordx8 s[4:11], s[12:13], 0x0
+; GFX6-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX6-NEXT:    s_load_dword s4, s[12:13], 0x8
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:13
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:15
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:14
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:8
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:11
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:10
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:4
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:6
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:1
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:3
+; GFX6-NEXT:    s_lshr_b32 s8, s9, 16
+; GFX6-NEXT:    s_waitcnt expcnt(0)
+; GFX6-NEXT:    v_mov_b32_e32 v0, s6
+; GFX6-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX6-NEXT:    s_lshl_b64 s[6:7], s[4:5], 8
+; GFX6-NEXT:    v_mov_b32_e32 v1, s11
+; GFX6-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:9
+; GFX6-NEXT:    s_waitcnt expcnt(0)
+; GFX6-NEXT:    v_mov_b32_e32 v1, s5
+; GFX6-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:2
+; GFX6-NEXT:    v_alignbit_b32 v0, s8, v0, 16
+; GFX6-NEXT:    s_waitcnt expcnt(0)
+; GFX6-NEXT:    v_mov_b32_e32 v1, s7
+; GFX6-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:12
+; GFX6-NEXT:    s_waitcnt expcnt(0)
+; GFX6-NEXT:    v_lshrrev_b32_e32 v1, 8, v0
+; GFX6-NEXT:    v_lshrrev_b32_e32 v0, 24, v0
+; GFX6-NEXT:    buffer_store_byte v1, off, s[0:3], 0 offset:5
+; GFX6-NEXT:    buffer_store_byte v0, off, s[0:3], 0 offset:7
+; GFX6-NEXT:    s_endpgm
 ;
-; GCNF-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
-; GCNF:       ; %bb.0: ; %bb
-; GCNF-NEXT:    s_load_dwordx4 s[8:11], s[4:5], 0x24
-; GCNF-NEXT:    v_mov_b32_e32 v2, 0
-; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
-; GCNF-NEXT:    s_load_dwordx8 s[0:7], s[8:9], 0x0
-; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
-; GCNF-NEXT:    s_load_dword s0, s[8:9], 0x20
-; GCNF-NEXT:    s_lshr_b32 s6, s5, 24
-; GCNF-NEXT:    s_lshr_b32 s8, s2, 24
-; GCNF-NEXT:    s_waitcnt lgkmcnt(0)
-; GCNF-NEXT:    s_lshl_b64 s[2:3], s[0:1], 8
-; GCNF-NEXT:    s_add_u32 s4, s10, 13
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 15
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 14
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 8
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 11
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 10
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 4
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 6
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 1
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    v_mov_b32_e32 v0, s10
-; GCNF-NEXT:    v_mov_b32_e32 v1, s11
-; GCNF-NEXT:    s_add_u32 s4, s10, 3
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    s_add_u32 s4, s10, 9
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    v_mov_b32_e32 v2, s7
-; GCNF-NEXT:    s_add_u32 s4, s10, 2
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s5, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s4
-; GCNF-NEXT:    v_mov_b32_e32 v1, s5
-; GCNF-NEXT:    v_mov_b32_e32 v2, s1
-; GCNF-NEXT:    s_add_u32 s0, s10, 5
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s1, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s0
-; GCNF-NEXT:    v_mov_b32_e32 v1, s1
-; GCNF-NEXT:    v_mov_b32_e32 v2, s8
-; GCNF-NEXT:    s_add_u32 s0, s10, 12
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s1, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s0
-; GCNF-NEXT:    v_mov_b32_e32 v1, s1
-; GCNF-NEXT:    v_mov_b32_e32 v2, s3
-; GCNF-NEXT:    s_add_u32 s0, s10, 7
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_addc_u32 s1, s11, 0
-; GCNF-NEXT:    v_mov_b32_e32 v0, s0
-; GCNF-NEXT:    v_mov_b32_e32 v1, s1
-; GCNF-NEXT:    v_mov_b32_e32 v2, s6
-; GCNF-NEXT:    flat_store_byte v[0:1], v2
-; GCNF-NEXT:    s_endpgm
+; GFX8-LABEL: any_extend_vector_inreg_v16i8_to_v4i32:
+; GFX8:       ; %bb.0: ; %bb
+; GFX8-NEXT:    s_load_dwordx4 s[8:11], s[4:5], 0x24
+; GFX8-NEXT:    v_mov_b32_e32 v2, 0
+; GFX8-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX8-NEXT:    s_load_dwordx8 s[0:7], s[8:9], 0x0
+; GFX8-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX8-NEXT:    s_load_dword s0, s[8:9], 0x20
+; GFX8-NEXT:    s_lshr_b32 s6, s5, 24
+; GFX8-NEXT:    s_lshr_b32 s8, s2, 24
+; GFX8-NEXT:    s_waitcnt lgkmcnt(0)
+; GFX8-NEXT:    s_lshl_b64 s[2:3], s[0:1], 8
+; GFX8-NEXT:    s_add_u32 s4, s10, 13
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 15
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 14
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 8
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 11
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 10
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 4
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 6
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 1
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    v_mov_b32_e32 v0, s10
+; GFX8-NEXT:    v_mov_b32_e32 v1, s11
+; GFX8-NEXT:    s_add_u32 s4, s10, 3
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    s_add_u32 s4, s10, 9
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    v_mov_b32_e32 v2, s7
+; GFX8-NEXT:    s_add_u32 s4, s10, 2
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s5, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s4
+; GFX8-NEXT:    v_mov_b32_e32 v1, s5
+; GFX8-NEXT:    v_mov_b32_e32 v2, s1
+; GFX8-NEXT:    s_add_u32 s0, s10, 5
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s1, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s0
+; GFX8-NEXT:    v_mov_b32_e32 v1, s1
+; GFX8-NEXT:    v_mov_b32_e32 v2, s8
+; GFX8-NEXT:    s_add_u32 s0, s10, 12
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s1, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s0
+; GFX8-NEXT:    v_mov_b32_e32 v1, s1
+; GFX8-NEXT:    v_mov_b32_e32 v2, s3
+; GFX8-NEXT:    s_add_u32 s0, s10, 7
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_addc_u32 s1, s11, 0
+; GFX8-NEXT:    v_mov_b32_e32 v0, s0
+; GFX8-NEXT:    v_mov_b32_e32 v1, s1
+; GFX8-NEXT:    v_mov_b32_e32 v2, s6
+; GFX8-NEXT:    flat_store_byte v[0:1], v2
+; GFX8-NEXT:    s_endpgm
 bb:
   %tmp2 = load <16 x i8>, ptr addrspace(1) %arg, align 16
   %tmp3 = extractelement <16 x i8> %tmp2, i64 4



More information about the llvm-commits mailing list