[llvm] 8be0da7 - [AMDGPU] Regenerate omod.ll tests

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Wed Apr 6 02:56:02 PDT 2022


Author: Simon Pilgrim
Date: 2022-04-06T10:55:54+01:00
New Revision: 8be0da7901ede4c7e25e1699efee303fa576d5c5

URL: https://github.com/llvm/llvm-project/commit/8be0da7901ede4c7e25e1699efee303fa576d5c5
DIFF: https://github.com/llvm/llvm-project/commit/8be0da7901ede4c7e25e1699efee303fa576d5c5.diff

LOG: [AMDGPU] Regenerate omod.ll tests

Added: 
    

Modified: 
    llvm/test/CodeGen/AMDGPU/omod.ll

Removed: 
    


################################################################################
diff  --git a/llvm/test/CodeGen/AMDGPU/omod.ll b/llvm/test/CodeGen/AMDGPU/omod.ll
index 94cb156c842ca..3d6fda8831695 100644
--- a/llvm/test/CodeGen/AMDGPU/omod.ll
+++ b/llvm/test/CodeGen/AMDGPU/omod.ll
@@ -1,12 +1,43 @@
-; RUN: llc -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck -check-prefix=GCN %s
-; RUN: llc -march=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck --check-prefixes=GCN,VI %s
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc -march=amdgcn -mcpu=tahiti -verify-machineinstrs < %s | FileCheck --check-prefixes=SI %s
+; RUN: llc -march=amdgcn -mcpu=fiji -verify-machineinstrs < %s | FileCheck --check-prefixes=VI %s
 
 ; IEEE bit enabled for compute kernel, so shouldn't use.
-; GCN-LABEL: {{^}}v_omod_div2_f32_enable_ieee_signed_zeros:
-; GCN: {{buffer|flat}}_load_dword [[A:v[0-9]+]]
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, [[A]]{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_kernel void @v_omod_div2_f32_enable_ieee_signed_zeros(float addrspace(1)* %out, float addrspace(1)* %aptr) #4 {
+; SI-LABEL: v_omod_div2_f32_enable_ieee_signed_zeros:
+; SI:       ; %bb.0:
+; SI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x9
+; SI-NEXT:    s_mov_b32 s7, 0xf000
+; SI-NEXT:    s_mov_b32 s6, 0
+; SI-NEXT:    v_lshlrev_b32_e32 v0, 2, v0
+; SI-NEXT:    v_mov_b32_e32 v1, 0
+; SI-NEXT:    s_waitcnt lgkmcnt(0)
+; SI-NEXT:    s_mov_b64 s[4:5], s[2:3]
+; SI-NEXT:    buffer_load_dword v2, v[0:1], s[4:7], 0 addr64
+; SI-NEXT:    s_mov_b64 s[2:3], s[6:7]
+; SI-NEXT:    s_waitcnt vmcnt(0)
+; SI-NEXT:    v_add_f32_e32 v2, 1.0, v2
+; SI-NEXT:    v_mul_f32_e32 v2, 0.5, v2
+; SI-NEXT:    buffer_store_dword v2, v[0:1], s[0:3], 0 addr64
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f32_enable_ieee_signed_zeros:
+; VI:       ; %bb.0:
+; VI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x24
+; VI-NEXT:    v_lshlrev_b32_e32 v2, 2, v0
+; VI-NEXT:    s_waitcnt lgkmcnt(0)
+; VI-NEXT:    v_mov_b32_e32 v1, s3
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s2, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    flat_load_dword v3, v[0:1]
+; VI-NEXT:    v_mov_b32_e32 v1, s1
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s0, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    s_waitcnt vmcnt(0)
+; VI-NEXT:    v_add_f32_e32 v2, 1.0, v3
+; VI-NEXT:    v_mul_f32_e32 v2, 0.5, v2
+; VI-NEXT:    flat_store_dword v[0:1], v2
+; VI-NEXT:    s_endpgm
   %tid = call i32 @llvm.amdgcn.workitem.id.x()
   %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid
   %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid
@@ -18,11 +49,41 @@ define amdgpu_kernel void @v_omod_div2_f32_enable_ieee_signed_zeros(float addrsp
 }
 
 ; IEEE bit enabled for compute kernel, so shouldn't use.
-; GCN-LABEL: {{^}}v_omod_div2_f64_enable_ieee_signed_zeros:
-; GCN: {{buffer|flat}}_load_dwordx2 [[A:v\[[0-9]+:[0-9]+\]]]
-; GCN: v_add_f64 [[ADD:v\[[0-9]+:[0-9]+\]]], [[A]], 1.0{{$}}
-; GCN: v_mul_f64 v{{\[[0-9]+:[0-9]+\]}}, [[ADD]], 0.5{{$}}
 define amdgpu_kernel void @v_omod_div2_f64_enable_ieee_signed_zeros(double addrspace(1)* %out, double addrspace(1)* %aptr) #4 {
+; SI-LABEL: v_omod_div2_f64_enable_ieee_signed_zeros:
+; SI:       ; %bb.0:
+; SI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x9
+; SI-NEXT:    s_mov_b32 s7, 0xf000
+; SI-NEXT:    s_mov_b32 s6, 0
+; SI-NEXT:    v_lshlrev_b32_e32 v0, 3, v0
+; SI-NEXT:    v_mov_b32_e32 v1, 0
+; SI-NEXT:    s_waitcnt lgkmcnt(0)
+; SI-NEXT:    s_mov_b64 s[4:5], s[2:3]
+; SI-NEXT:    buffer_load_dwordx2 v[2:3], v[0:1], s[4:7], 0 addr64
+; SI-NEXT:    s_mov_b64 s[2:3], s[6:7]
+; SI-NEXT:    s_waitcnt vmcnt(0)
+; SI-NEXT:    v_add_f64 v[2:3], v[2:3], 1.0
+; SI-NEXT:    v_mul_f64 v[2:3], v[2:3], 0.5
+; SI-NEXT:    buffer_store_dwordx2 v[2:3], v[0:1], s[0:3], 0 addr64
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f64_enable_ieee_signed_zeros:
+; VI:       ; %bb.0:
+; VI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x24
+; VI-NEXT:    v_lshlrev_b32_e32 v2, 3, v0
+; VI-NEXT:    s_waitcnt lgkmcnt(0)
+; VI-NEXT:    v_mov_b32_e32 v1, s3
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s2, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    flat_load_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    v_mov_b32_e32 v3, s1
+; VI-NEXT:    v_add_u32_e32 v2, vcc, s0, v2
+; VI-NEXT:    v_addc_u32_e32 v3, vcc, 0, v3, vcc
+; VI-NEXT:    s_waitcnt vmcnt(0)
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; VI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; VI-NEXT:    flat_store_dwordx2 v[2:3], v[0:1]
+; VI-NEXT:    s_endpgm
   %tid = call i32 @llvm.amdgcn.workitem.id.x()
   %gep0 = getelementptr double, double addrspace(1)* %aptr, i32 %tid
   %out.gep = getelementptr double, double addrspace(1)* %out, i32 %tid
@@ -34,11 +95,41 @@ define amdgpu_kernel void @v_omod_div2_f64_enable_ieee_signed_zeros(double addrs
 }
 
 ; IEEE bit enabled for compute kernel, so shouldn't use even though nsz is allowed
-; GCN-LABEL: {{^}}v_omod_div2_f32_enable_ieee_nsz:
-; GCN: {{buffer|flat}}_load_dword [[A:v[0-9]+]]
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, [[A]]{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_kernel void @v_omod_div2_f32_enable_ieee_nsz(float addrspace(1)* %out, float addrspace(1)* %aptr) #0 {
+; SI-LABEL: v_omod_div2_f32_enable_ieee_nsz:
+; SI:       ; %bb.0:
+; SI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x9
+; SI-NEXT:    s_mov_b32 s7, 0xf000
+; SI-NEXT:    s_mov_b32 s6, 0
+; SI-NEXT:    v_lshlrev_b32_e32 v0, 2, v0
+; SI-NEXT:    v_mov_b32_e32 v1, 0
+; SI-NEXT:    s_waitcnt lgkmcnt(0)
+; SI-NEXT:    s_mov_b64 s[4:5], s[2:3]
+; SI-NEXT:    buffer_load_dword v2, v[0:1], s[4:7], 0 addr64
+; SI-NEXT:    s_mov_b64 s[2:3], s[6:7]
+; SI-NEXT:    s_waitcnt vmcnt(0)
+; SI-NEXT:    v_add_f32_e32 v2, 1.0, v2
+; SI-NEXT:    v_mul_f32_e32 v2, 0.5, v2
+; SI-NEXT:    buffer_store_dword v2, v[0:1], s[0:3], 0 addr64
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f32_enable_ieee_nsz:
+; VI:       ; %bb.0:
+; VI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x24
+; VI-NEXT:    v_lshlrev_b32_e32 v2, 2, v0
+; VI-NEXT:    s_waitcnt lgkmcnt(0)
+; VI-NEXT:    v_mov_b32_e32 v1, s3
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s2, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    flat_load_dword v3, v[0:1]
+; VI-NEXT:    v_mov_b32_e32 v1, s1
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s0, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    s_waitcnt vmcnt(0)
+; VI-NEXT:    v_add_f32_e32 v2, 1.0, v3
+; VI-NEXT:    v_mul_f32_e32 v2, 0.5, v2
+; VI-NEXT:    flat_store_dword v[0:1], v2
+; VI-NEXT:    s_endpgm
   %tid = call i32 @llvm.amdgcn.workitem.id.x()
   %gep0 = getelementptr float, float addrspace(1)* %aptr, i32 %tid
   %out.gep = getelementptr float, float addrspace(1)* %out, i32 %tid
@@ -50,11 +141,41 @@ define amdgpu_kernel void @v_omod_div2_f32_enable_ieee_nsz(float addrspace(1)* %
 }
 
 ; IEEE bit enabled for compute kernel, so shouldn't use even though nsz is allowed.
-; GCN-LABEL: {{^}}v_omod_div2_f64_enable_ieee_nsz:
-; GCN: {{buffer|flat}}_load_dwordx2 [[A:v\[[0-9]+:[0-9]+\]]]
-; GCN: v_add_f64 [[ADD:v\[[0-9]+:[0-9]+\]]], [[A]], 1.0{{$}}
-; GCN: v_mul_f64 v{{\[[0-9]+:[0-9]+\]}}, [[ADD]], 0.5{{$}}
 define amdgpu_kernel void @v_omod_div2_f64_enable_ieee_nsz(double addrspace(1)* %out, double addrspace(1)* %aptr) #5 {
+; SI-LABEL: v_omod_div2_f64_enable_ieee_nsz:
+; SI:       ; %bb.0:
+; SI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x9
+; SI-NEXT:    s_mov_b32 s7, 0xf000
+; SI-NEXT:    s_mov_b32 s6, 0
+; SI-NEXT:    v_lshlrev_b32_e32 v0, 3, v0
+; SI-NEXT:    v_mov_b32_e32 v1, 0
+; SI-NEXT:    s_waitcnt lgkmcnt(0)
+; SI-NEXT:    s_mov_b64 s[4:5], s[2:3]
+; SI-NEXT:    buffer_load_dwordx2 v[2:3], v[0:1], s[4:7], 0 addr64
+; SI-NEXT:    s_mov_b64 s[2:3], s[6:7]
+; SI-NEXT:    s_waitcnt vmcnt(0)
+; SI-NEXT:    v_add_f64 v[2:3], v[2:3], 1.0
+; SI-NEXT:    v_mul_f64 v[2:3], v[2:3], 0.5
+; SI-NEXT:    buffer_store_dwordx2 v[2:3], v[0:1], s[0:3], 0 addr64
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f64_enable_ieee_nsz:
+; VI:       ; %bb.0:
+; VI-NEXT:    s_load_dwordx4 s[0:3], s[0:1], 0x24
+; VI-NEXT:    v_lshlrev_b32_e32 v2, 3, v0
+; VI-NEXT:    s_waitcnt lgkmcnt(0)
+; VI-NEXT:    v_mov_b32_e32 v1, s3
+; VI-NEXT:    v_add_u32_e32 v0, vcc, s2, v2
+; VI-NEXT:    v_addc_u32_e32 v1, vcc, 0, v1, vcc
+; VI-NEXT:    flat_load_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    v_mov_b32_e32 v3, s1
+; VI-NEXT:    v_add_u32_e32 v2, vcc, s0, v2
+; VI-NEXT:    v_addc_u32_e32 v3, vcc, 0, v3, vcc
+; VI-NEXT:    s_waitcnt vmcnt(0)
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; VI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; VI-NEXT:    flat_store_dwordx2 v[2:3], v[0:1]
+; VI-NEXT:    s_endpgm
   %tid = call i32 @llvm.amdgcn.workitem.id.x()
   %gep0 = getelementptr double, double addrspace(1)* %aptr, i32 %tid
   %out.gep = getelementptr double, double addrspace(1)* %out, i32 %tid
@@ -66,10 +187,22 @@ define amdgpu_kernel void @v_omod_div2_f64_enable_ieee_nsz(double addrspace(1)*
 }
 
 ; Only allow without IEEE bit if signed zeros are significant.
-; GCN-LABEL: {{^}}v_omod_div2_f32_signed_zeros:
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_div2_f32_signed_zeros(float %a) #4 {
+; SI-LABEL: v_omod_div2_f32_signed_zeros:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f32_signed_zeros:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 0.5
   store float %div2, float addrspace(1)* undef
@@ -77,74 +210,168 @@ define amdgpu_ps void @v_omod_div2_f32_signed_zeros(float %a) #4 {
 }
 
 ; Only allow without IEEE bit if signed zeros are significant.
-; GCN-LABEL: {{^}}v_omod_div2_f64_signed_zeros:
-; GCN: v_add_f64 [[ADD:v\[[0-9]+:[0-9]+\]]], v{{\[[0-9]+:[0-9]+\]}}, 1.0{{$}}
-; GCN: v_mul_f64 v{{\[[0-9]+:[0-9]+\]}}, [[ADD]], 0.5{{$}}
 define amdgpu_ps void @v_omod_div2_f64_signed_zeros(double %a) #4 {
+; SI-LABEL: v_omod_div2_f64_signed_zeros:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f64_signed_zeros:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; VI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd double %a, 1.0
   %div2 = fmul double %add, 0.5
   store double %div2, double addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_div2_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, 1.0 div:2{{$}}
 define amdgpu_ps void @v_omod_div2_f32(float %a) #0 {
+; SI-LABEL: v_omod_div2_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 0.5
   store float %div2, float addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_div2_f64:
-; GCN: v_add_f64  v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, 1.0 div:2{{$}}
 define amdgpu_ps void @v_omod_div2_f64(double %a) #5 {
+; SI-LABEL: v_omod_div2_f64:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 div:2
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f64:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 div:2
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd nsz double %a, 1.0
   %div2 = fmul nsz double %add, 0.5
   store double %div2, double addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul2_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, 1.0 mul:2{{$}}
 define amdgpu_ps void @v_omod_mul2_f32(float %a) #0 {
+; SI-LABEL: v_omod_mul2_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:2
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul2_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:2
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 2.0
   store float %div2, float addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul2_f64:
-; GCN: v_add_f64  v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, 1.0 mul:2{{$}}
 define amdgpu_ps void @v_omod_mul2_f64(double %a) #5 {
+; SI-LABEL: v_omod_mul2_f64:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 mul:2
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul2_f64:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 mul:2
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd nsz double %a, 1.0
   %div2 = fmul nsz double %add, 2.0
   store double %div2, double addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul4_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, 1.0 mul:4{{$}}
 define amdgpu_ps void @v_omod_mul4_f32(float %a) #0 {
+; SI-LABEL: v_omod_mul4_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:4
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul4_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:4
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 4.0
   store float %div2, float addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul4_f64:
-; GCN: v_add_f64  v{{\[[0-9]+:[0-9]+\]}}, v{{\[[0-9]+:[0-9]+\]}}, 1.0 mul:4{{$}}
 define amdgpu_ps void @v_omod_mul4_f64(double %a) #5 {
+; SI-LABEL: v_omod_mul4_f64:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 mul:4
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul4_f64:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0 mul:4
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd nsz double %a, 1.0
   %div2 = fmul nsz double %add, 4.0
   store double %div2, double addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul4_multi_use_f32:
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 4.0, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_mul4_multi_use_f32(float %a) #0 {
+; SI-LABEL: v_omod_mul4_multi_use_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_mul_f32_e32 v1, 4.0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v1, off, s[0:3], 0
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_waitcnt vmcnt(0)
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul4_multi_use_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_mul_f32_e32 v1, 4.0, v0
+; VI-NEXT:    flat_store_dword v[0:1], v1
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_waitcnt vmcnt(0)
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 4.0
   store float %div2, float addrspace(1)* undef
@@ -152,9 +379,20 @@ define amdgpu_ps void @v_omod_mul4_multi_use_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mul4_dbg_use_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, 1.0 mul:4{{$}}
 define amdgpu_ps void @v_omod_mul4_dbg_use_f32(float %a) #0 {
+; SI-LABEL: v_omod_mul4_dbg_use_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:4
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul4_dbg_use_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:4
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   call void @llvm.dbg.value(metadata float %add, i64 0, metadata !4, metadata !9), !dbg !10
   %div2 = fmul float %add, 4.0
@@ -163,9 +401,20 @@ define amdgpu_ps void @v_omod_mul4_dbg_use_f32(float %a) #0 {
 }
 
 ; Clamp is applied after omod, folding both into instruction is OK.
-; GCN-LABEL: {{^}}v_clamp_omod_div2_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, 1.0 clamp div:2{{$}}
 define amdgpu_ps void @v_clamp_omod_div2_f32(float %a) #0 {
+; SI-LABEL: v_clamp_omod_div2_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 clamp div:2
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_clamp_omod_div2_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 clamp div:2
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 0.5
 
@@ -176,10 +425,22 @@ define amdgpu_ps void @v_clamp_omod_div2_f32(float %a) #0 {
 }
 
 ; Cannot fold omod into clamp
-; GCN-LABEL: {{^}}v_omod_div2_clamp_f32:
-; GCN: v_add_f32_e64 [[ADD:v[0-9]+]], v0, 1.0 clamp{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_div2_clamp_f32(float %a) #0 {
+; SI-LABEL: v_omod_div2_clamp_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 clamp
+; SI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_clamp_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 clamp
+; VI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %max = call float @llvm.maxnum.f32(float %add, float 0.0)
   %clamp = call float @llvm.minnum.f32(float %max, float 1.0)
@@ -188,10 +449,22 @@ define amdgpu_ps void @v_omod_div2_clamp_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_div2_abs_src_f32:
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; GCN: v_mul_f32_e64 v{{[0-9]+}}, |[[ADD]]|, 0.5{{$}}
 define amdgpu_ps void @v_omod_div2_abs_src_f32(float %a) #0 {
+; SI-LABEL: v_omod_div2_abs_src_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_mul_f32_e64 v0, |v0|, 0.5
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_abs_src_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_mul_f32_e64 v0, |v0|, 0.5
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %abs.add = call float @llvm.fabs.f32(float %add)
   %div2 = fmul float %abs.add, 0.5
@@ -199,9 +472,20 @@ define amdgpu_ps void @v_omod_div2_abs_src_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_add_self_clamp_f32:
-; GCN: v_add_f32_e64 v{{[0-9]+}}, v0, v0 clamp{{$}}
 define amdgpu_ps void @v_omod_add_self_clamp_f32(float %a) #0 {
+; SI-LABEL: v_omod_add_self_clamp_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, v0 clamp
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_add_self_clamp_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, v0 clamp
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, %a
   %max = call float @llvm.maxnum.f32(float %add, float 0.0)
   %clamp = call float @llvm.minnum.f32(float %max, float 1.0)
@@ -209,10 +493,22 @@ define amdgpu_ps void @v_omod_add_self_clamp_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_add_clamp_self_f32:
-; GCN: v_max_f32_e64 [[CLAMP:v[0-9]+]], v0, v0 clamp{{$}}
-; GCN: v_add_f32_e32 v{{[0-9]+}}, [[CLAMP]], [[CLAMP]]{{$}}
 define amdgpu_ps void @v_omod_add_clamp_self_f32(float %a) #0 {
+; SI-LABEL: v_omod_add_clamp_self_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_max_f32_e64 v0, v0, v0 clamp
+; SI-NEXT:    v_add_f32_e32 v0, v0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_add_clamp_self_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_max_f32_e64 v0, v0, v0 clamp
+; VI-NEXT:    v_add_f32_e32 v0, v0, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %max = call float @llvm.maxnum.f32(float %a, float 0.0)
   %clamp = call float @llvm.minnum.f32(float %max, float 1.0)
   %add = fadd float %clamp, %clamp
@@ -220,10 +516,22 @@ define amdgpu_ps void @v_omod_add_clamp_self_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_add_abs_self_f32:
-; GCN: v_add_f32_e32 [[X:v[0-9]+]], 1.0, v0
-; GCN: v_add_f32_e64 v{{[0-9]+}}, |[[X]]|, |[[X]]|{{$}}
 define amdgpu_ps void @v_omod_add_abs_self_f32(float %a) #0 {
+; SI-LABEL: v_omod_add_abs_self_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_add_f32_e64 v0, |v0|, |v0|
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_add_abs_self_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_add_f32_e64 v0, |v0|, |v0|
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %x = fadd float %a, 1.0
   %abs.x = call float @llvm.fabs.f32(float %x)
   %add = fadd float %abs.x, %abs.x
@@ -231,11 +539,22 @@ define amdgpu_ps void @v_omod_add_abs_self_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_add_abs_x_x_f32:
-
-; GCN: v_add_f32_e32 [[X:v[0-9]+]], 1.0, v0
-; GCN: v_add_f32_e64 v{{[0-9]+}}, |[[X]]|, [[X]]{{$}}
 define amdgpu_ps void @v_omod_add_abs_x_x_f32(float %a) #0 {
+; SI-LABEL: v_omod_add_abs_x_x_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_add_f32_e64 v0, |v0|, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_add_abs_x_x_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_add_f32_e64 v0, |v0|, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %x = fadd float %a, 1.0
   %abs.x = call float @llvm.fabs.f32(float %x)
   %add = fadd float %abs.x, %x
@@ -243,10 +562,22 @@ define amdgpu_ps void @v_omod_add_abs_x_x_f32(float %a) #0 {
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_add_x_abs_x_f32:
-; GCN: v_add_f32_e32 [[X:v[0-9]+]], 1.0, v0
-; GCN: v_add_f32_e64 v{{[0-9]+}}, [[X]], |[[X]]|{{$}}
 define amdgpu_ps void @v_omod_add_x_abs_x_f32(float %a) #0 {
+; SI-LABEL: v_omod_add_x_abs_x_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_add_f32_e64 v0, v0, |v0|
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_add_x_abs_x_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_add_f32_e64 v0, v0, |v0|
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %x = fadd float %a, 1.0
   %abs.x = call float @llvm.fabs.f32(float %x)
   %add = fadd float %x, %abs.x
@@ -255,10 +586,22 @@ define amdgpu_ps void @v_omod_add_x_abs_x_f32(float %a) #0 {
 }
 
 ; Don't fold omod into omod into another omod.
-; GCN-LABEL: {{^}}v_omod_div2_omod_div2_f32:
-; GCN: v_add_f32_e64 [[ADD:v[0-9]+]], v0, 1.0 div:2{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_div2_omod_div2_f32(float %a) #0 {
+; SI-LABEL: v_omod_div2_omod_div2_f32:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; SI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_omod_div2_f32:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; VI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2.0 = fmul float %add, 0.5
   %div2.1 = fmul float %div2.0, 0.5
@@ -267,10 +610,22 @@ define amdgpu_ps void @v_omod_div2_omod_div2_f32(float %a) #0 {
 }
 
 ; Don't fold omod if denorms enabled
-; GCN-LABEL: {{^}}v_omod_div2_f32_denormals:
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; GCN: v_mul_f32_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_div2_f32_denormals(float %a) #2 {
+; SI-LABEL: v_omod_div2_f32_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f32_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_mul_f32_e32 v0, 0.5, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %div2 = fmul float %add, 0.5
   store float %div2, float addrspace(1)* undef
@@ -278,10 +633,22 @@ define amdgpu_ps void @v_omod_div2_f32_denormals(float %a) #2 {
 }
 
 ; Don't fold omod if denorms enabled.
-; GCN-LABEL: {{^}}v_omod_div2_f64_denormals:
-; GCN: v_add_f64 [[ADD:v\[[0-9]+:[0-9]+\]]], v{{\[[0-9]+:[0-9]+\]}}, 1.0{{$}}
-; GCN: v_mul_f64 v{{\[[0-9]+:[0-9]+\]}}, [[ADD]], 0.5{{$}}
 define amdgpu_ps void @v_omod_div2_f64_denormals(double %a) #6 {
+; SI-LABEL: v_omod_div2_f64_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f64_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; VI-NEXT:    v_mul_f64 v[0:1], v[0:1], 0.5
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd double %a, 1.0
   %div2 = fmul double %add, 0.5
   store double %div2, double addrspace(1)* undef
@@ -289,10 +656,22 @@ define amdgpu_ps void @v_omod_div2_f64_denormals(double %a) #6 {
 }
 
 ; Don't fold omod if denorms enabled for add form.
-; GCN-LABEL: {{^}}v_omod_mul2_f32_denormals:
-; GCN: v_add_f32_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; GCN: v_add_f32_e32 v{{[0-9]+}}, [[ADD]], [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_mul2_f32_denormals(float %a) #2 {
+; SI-LABEL: v_omod_mul2_f32_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; SI-NEXT:    v_add_f32_e32 v0, v0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul2_f32_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f32_e32 v0, 1.0, v0
+; VI-NEXT:    v_add_f32_e32 v0, v0, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd float %a, 1.0
   %mul2 = fadd float %add, %add
   store float %mul2, float addrspace(1)* undef
@@ -300,10 +679,22 @@ define amdgpu_ps void @v_omod_mul2_f32_denormals(float %a) #2 {
 }
 
 ; Don't fold omod if denorms enabled for add form.
-; GCN-LABEL: {{^}}v_omod_mul2_f64_denormals:
-; GCN: v_add_f64 [[ADD:v\[[0-9]+:[0-9]+\]]], v{{\[[0-9]+:[0-9]+\]}}, 1.0{{$}}
-; GCN: v_add_f64 v{{\[[0-9]+:[0-9]+\]}}, [[ADD]], [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_mul2_f64_denormals(double %a) #2 {
+; SI-LABEL: v_omod_mul2_f64_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    v_add_f64 v[0:1], v[0:1], v[0:1]
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dwordx2 v[0:1], off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul2_f64_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], 1.0
+; VI-NEXT:    v_add_f64 v[0:1], v[0:1], v[0:1]
+; VI-NEXT:    flat_store_dwordx2 v[0:1], v[0:1]
+; VI-NEXT:    s_endpgm
   %add = fadd double %a, 1.0
   %mul2 = fadd double %add, %add
   store double %mul2, double addrspace(1)* undef
@@ -311,10 +702,24 @@ define amdgpu_ps void @v_omod_mul2_f64_denormals(double %a) #2 {
 }
 
 ; Don't fold omod if denorms enabled
-; GCN-LABEL: {{^}}v_omod_div2_f16_denormals:
-; VI: v_add_f16_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; VI: v_mul_f16_e32 v{{[0-9]+}}, 0.5, [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_div2_f16_denormals(half %a) #0 {
+; SI-LABEL: v_omod_div2_f16_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    v_cvt_f32_f16_e32 v0, v0
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    buffer_store_short v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f16_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f16_e32 v0, 1.0, v0
+; VI-NEXT:    v_mul_f16_e32 v0, 0.5, v0
+; VI-NEXT:    flat_store_short v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd half %a, 1.0
   %div2 = fmul half %add, 0.5
   store half %div2, half addrspace(1)* undef
@@ -322,29 +727,69 @@ define amdgpu_ps void @v_omod_div2_f16_denormals(half %a) #0 {
 }
 
 ; Don't fold omod if denorms enabled for add form.
-; GCN-LABEL: {{^}}v_omod_mul2_f16_denormals:
-; VI: v_add_f16_e32 [[ADD:v[0-9]+]], 1.0, v0{{$}}
-; VI: v_add_f16_e32 v{{[0-9]+}}, [[ADD]], [[ADD]]{{$}}
 define amdgpu_ps void @v_omod_mul2_f16_denormals(half %a) #0 {
+; SI-LABEL: v_omod_mul2_f16_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    v_cvt_f32_f16_e32 v0, v0
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 mul:2
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    buffer_store_short v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mul2_f16_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f16_e32 v0, 1.0, v0
+; VI-NEXT:    v_add_f16_e32 v0, v0, v0
+; VI-NEXT:    flat_store_short v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd half %a, 1.0
   %mul2 = fadd half %add, %add
   store half %mul2, half addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_div2_f16_no_denormals:
-; VI-NOT: v0
-; VI: v_add_f16_e64 [[ADD:v[0-9]+]], v0, 1.0 div:2{{$}}
 define amdgpu_ps void @v_omod_div2_f16_no_denormals(half %a) #3 {
+; SI-LABEL: v_omod_div2_f16_no_denormals:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    v_cvt_f32_f16_e32 v0, v0
+; SI-NEXT:    v_add_f32_e64 v0, v0, 1.0 div:2
+; SI-NEXT:    v_cvt_f16_f32_e32 v0, v0
+; SI-NEXT:    buffer_store_short v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_div2_f16_no_denormals:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_add_f16_e64 v0, v0, 1.0 div:2
+; VI-NEXT:    flat_store_short v[0:1], v0
+; VI-NEXT:    s_endpgm
   %add = fadd half %a, 1.0
   %div2 = fmul half %add, 0.5
   store half %div2, half addrspace(1)* undef
   ret void
 }
 
-; GCN-LABEL: {{^}}v_omod_mac_to_mad:
-; GCN: v_mad_f32 v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]+}}, v{{[0-9]}} mul:2{{$}}
 define amdgpu_ps void @v_omod_mac_to_mad(float %b, float %a) #0 {
+; SI-LABEL: v_omod_mac_to_mad:
+; SI:       ; %bb.0:
+; SI-NEXT:    v_mad_f32 v1, v1, v1, v0 mul:2
+; SI-NEXT:    v_mul_f32_e32 v0, v1, v0
+; SI-NEXT:    s_mov_b32 s3, 0xf000
+; SI-NEXT:    s_mov_b32 s2, -1
+; SI-NEXT:    buffer_store_dword v0, off, s[0:3], 0
+; SI-NEXT:    s_endpgm
+;
+; VI-LABEL: v_omod_mac_to_mad:
+; VI:       ; %bb.0:
+; VI-NEXT:    v_mad_f32 v1, v1, v1, v0 mul:2
+; VI-NEXT:    v_mul_f32_e32 v0, v1, v0
+; VI-NEXT:    flat_store_dword v[0:1], v0
+; VI-NEXT:    s_endpgm
   %mul = fmul float %a, %a
   %add = fadd float %mul, %b
   %mad = fmul float %add, 2.0


        


More information about the llvm-commits mailing list