[llvm] r363315 - Revert "[NFC][CodeGen] Add unary fneg tests to X86/fma_patterns_wide.ll"

Cameron McInally via llvm-commits llvm-commits at lists.llvm.org
Thu Jun 13 12:25:09 PDT 2019


Author: mcinally
Date: Thu Jun 13 12:25:09 2019
New Revision: 363315

URL: http://llvm.org/viewvc/llvm-project?rev=363315&view=rev
Log:
Revert "[NFC][CodeGen] Add unary fneg tests to X86/fma_patterns_wide.ll"

This reverts commit f1b8c6ac4f9d31899a2bc128f8a37b5a1c3e1f77.

Modified:
    llvm/trunk/test/CodeGen/X86/fma_patterns_wide.ll

Modified: llvm/trunk/test/CodeGen/X86/fma_patterns_wide.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/fma_patterns_wide.ll?rev=363315&r1=363314&r2=363315&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/fma_patterns_wide.ll (original)
+++ llvm/trunk/test/CodeGen/X86/fma_patterns_wide.ll Thu Jun 13 12:25:09 2019
@@ -179,29 +179,6 @@ define <16 x float> @test_16f32_fnmsub(<
   ret <16 x float> %res
 }
 
-define <16 x float> @test_16f32_fnmsub_unary_fneg(<16 x float> %a0, <16 x float> %a1, <16 x float> %a2) {
-; FMA-LABEL: test_16f32_fnmsub_unary_fneg:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm0 = -(ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm1 = -(ymm3 * ymm1) - ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_16f32_fnmsub_unary_fneg:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfnmsubps %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmsubps %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_16f32_fnmsub_unary_fneg:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfnmsub213ps {{.*#+}} zmm0 = -(zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %x = fmul <16 x float> %a0, %a1
-  %y = fneg <16 x float> %x
-  %res = fsub <16 x float> %y, %a2
-  ret <16 x float> %res
-}
-
 define <8 x double> @test_8f64_fnmsub(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) {
 ; FMA-LABEL: test_8f64_fnmsub:
 ; FMA:       # %bb.0:
@@ -225,29 +202,6 @@ define <8 x double> @test_8f64_fnmsub(<8
   ret <8 x double> %res
 }
 
-define <8 x double> @test_8f64_fnmsub_unary_fneg(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) {
-; FMA-LABEL: test_8f64_fnmsub_unary_fneg:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfnmsub213pd {{.*#+}} ymm0 = -(ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfnmsub213pd {{.*#+}} ymm1 = -(ymm3 * ymm1) - ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_8f64_fnmsub_unary_fneg:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfnmsubpd %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmsubpd %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_8f64_fnmsub_unary_fneg:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfnmsub213pd {{.*#+}} zmm0 = -(zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %x = fmul <8 x double> %a0, %a1
-  %y = fneg <8 x double> %x
-  %res = fsub <8 x double> %y, %a2
-  ret <8 x double> %res
-}
-
 ;
 ; Load Folding Patterns
 ;
@@ -1007,29 +961,6 @@ define <16 x float> @test_v16f32_fneg_fm
   ret <16 x float> %neg
 }
 
-define <16 x float> @test_v16f32_unary_fneg_fmadd(<16 x float> %a0, <16 x float> %a1, <16 x float> %a2) #0 {
-; FMA-LABEL: test_v16f32_unary_fneg_fmadd:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm0 = -(ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm1 = -(ymm3 * ymm1) - ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v16f32_unary_fneg_fmadd:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfnmsubps %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmsubps %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v16f32_unary_fneg_fmadd:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfnmsub213ps {{.*#+}} zmm0 = -(zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %mul = fmul <16 x float> %a0, %a1
-  %add = fadd <16 x float> %mul, %a2
-  %neg = fneg <16 x float> %add
-  ret <16 x float> %neg
-}
-
 define <8 x double> @test_v8f64_fneg_fmsub(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) #0 {
 ; FMA-LABEL: test_v8f64_fneg_fmsub:
 ; FMA:       # %bb.0:
@@ -1053,29 +984,6 @@ define <8 x double> @test_v8f64_fneg_fms
   ret <8 x double> %neg
 }
 
-define <8 x double> @test_v8f64_unary_fneg_fmsub(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) #0 {
-; FMA-LABEL: test_v8f64_unary_fneg_fmsub:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfnmadd213pd {{.*#+}} ymm0 = -(ymm2 * ymm0) + ymm4
-; FMA-NEXT:    vfnmadd213pd {{.*#+}} ymm1 = -(ymm3 * ymm1) + ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v8f64_unary_fneg_fmsub:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfnmaddpd %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmaddpd %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v8f64_unary_fneg_fmsub:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfnmadd213pd {{.*#+}} zmm0 = -(zmm1 * zmm0) + zmm2
-; AVX512-NEXT:    retq
-  %mul = fmul <8 x double> %a0, %a1
-  %sub = fsub <8 x double> %mul, %a2
-  %neg = fneg <8 x double> %sub
-  ret <8 x double> %neg
-}
-
 define <16 x float> @test_v16f32_fneg_fnmadd(<16 x float> %a0, <16 x float> %a1, <16 x float> %a2) #0 {
 ; FMA-LABEL: test_v16f32_fneg_fnmadd:
 ; FMA:       # %bb.0:
@@ -1100,30 +1008,6 @@ define <16 x float> @test_v16f32_fneg_fn
   ret <16 x float> %neg1
 }
 
-define <16 x float> @test_v16f32_unary_fneg_fnmadd(<16 x float> %a0, <16 x float> %a1, <16 x float> %a2) #0 {
-; FMA-LABEL: test_v16f32_unary_fneg_fnmadd:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfmsub213ps {{.*#+}} ymm0 = (ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfmsub213ps {{.*#+}} ymm1 = (ymm3 * ymm1) - ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v16f32_unary_fneg_fnmadd:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfmsubps %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfmsubps %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v16f32_unary_fneg_fnmadd:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfmsub213ps {{.*#+}} zmm0 = (zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %mul = fmul <16 x float> %a0, %a1
-  %neg0 = fneg <16 x float> %mul
-  %add = fadd <16 x float> %neg0, %a2
-  %neg1 = fneg <16 x float> %add
-  ret <16 x float> %neg1
-}
-
 define <8 x double> @test_v8f64_fneg_fnmsub(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) #0 {
 ; FMA-LABEL: test_v8f64_fneg_fnmsub:
 ; FMA:       # %bb.0:
@@ -1148,30 +1032,6 @@ define <8 x double> @test_v8f64_fneg_fnm
   ret <8 x double> %neg1
 }
 
-define <8 x double> @test_v8f64_unary_fneg_fnmsub(<8 x double> %a0, <8 x double> %a1, <8 x double> %a2) #0 {
-; FMA-LABEL: test_v8f64_unary_fneg_fnmsub:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vfmadd213pd {{.*#+}} ymm0 = (ymm2 * ymm0) + ymm4
-; FMA-NEXT:    vfmadd213pd {{.*#+}} ymm1 = (ymm3 * ymm1) + ymm5
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v8f64_unary_fneg_fnmsub:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vfmaddpd %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfmaddpd %ymm5, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v8f64_unary_fneg_fnmsub:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vfmadd213pd {{.*#+}} zmm0 = (zmm1 * zmm0) + zmm2
-; AVX512-NEXT:    retq
-  %mul = fmul <8 x double> %a0, %a1
-  %neg0 = fneg <8 x double> %mul
-  %sub = fsub <8 x double> %neg0, %a2
-  %neg1 = fneg <8 x double> %sub
-  ret <8 x double> %neg1
-}
-
 ;
 ; Pattern: (fma x, c1, (fmul x, c2)) -> (fmul x, c1+c2)
 ;
@@ -1253,31 +1113,6 @@ define <16 x float> @test_v16f32_fneg_fm
   ret <16 x float> %n
 }
 
-define <16 x float> @test_v16f32_unary_fneg_fmul(<16 x float> %x, <16 x float> %y) #0 {
-; FMA-LABEL: test_v16f32_unary_fneg_fmul:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vxorps %xmm4, %xmm4, %xmm4
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm0 = -(ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfnmsub213ps {{.*#+}} ymm1 = -(ymm3 * ymm1) - ymm4
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v16f32_unary_fneg_fmul:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vxorps %xmm4, %xmm4, %xmm4
-; FMA4-NEXT:    vfnmsubps %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmsubps %ymm4, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v16f32_unary_fneg_fmul:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vxorps %xmm2, %xmm2, %xmm2
-; AVX512-NEXT:    vfnmsub213ps {{.*#+}} zmm0 = -(zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %m = fmul nsz <16 x float> %x, %y
-  %n = fneg <16 x float> %m
-  ret <16 x float> %n
-}
-
 define <8 x double> @test_v8f64_fneg_fmul(<8 x double> %x, <8 x double> %y) #0 {
 ; FMA-LABEL: test_v8f64_fneg_fmul:
 ; FMA:       # %bb.0:
@@ -1303,31 +1138,6 @@ define <8 x double> @test_v8f64_fneg_fmu
   ret <8 x double> %n
 }
 
-define <8 x double> @test_v8f64_unary_fneg_fmul(<8 x double> %x, <8 x double> %y) #0 {
-; FMA-LABEL: test_v8f64_unary_fneg_fmul:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vxorpd %xmm4, %xmm4, %xmm4
-; FMA-NEXT:    vfnmsub213pd {{.*#+}} ymm0 = -(ymm2 * ymm0) - ymm4
-; FMA-NEXT:    vfnmsub213pd {{.*#+}} ymm1 = -(ymm3 * ymm1) - ymm4
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v8f64_unary_fneg_fmul:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vxorpd %xmm4, %xmm4, %xmm4
-; FMA4-NEXT:    vfnmsubpd %ymm4, %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vfnmsubpd %ymm4, %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v8f64_unary_fneg_fmul:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vxorpd %xmm2, %xmm2, %xmm2
-; AVX512-NEXT:    vfnmsub213pd {{.*#+}} zmm0 = -(zmm1 * zmm0) - zmm2
-; AVX512-NEXT:    retq
-  %m = fmul nsz <8 x double> %x, %y
-  %n = fneg <8 x double> %m
-  ret <8 x double> %n
-}
-
 define <8 x double> @test_v8f64_fneg_fmul_no_nsz(<8 x double> %x, <8 x double> %y) #0 {
 ; FMA-LABEL: test_v8f64_fneg_fmul_no_nsz:
 ; FMA:       # %bb.0:
@@ -1357,33 +1167,4 @@ define <8 x double> @test_v8f64_fneg_fmu
   ret <8 x double> %n
 }
 
-define <8 x double> @test_v8f64_unary_fneg_fmul_no_nsz(<8 x double> %x, <8 x double> %y) #0 {
-; FMA-LABEL: test_v8f64_unary_fneg_fmul_no_nsz:
-; FMA:       # %bb.0:
-; FMA-NEXT:    vmulpd %ymm3, %ymm1, %ymm1
-; FMA-NEXT:    vmulpd %ymm2, %ymm0, %ymm0
-; FMA-NEXT:    vmovapd {{.*#+}} ymm2 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0]
-; FMA-NEXT:    vxorpd %ymm2, %ymm0, %ymm0
-; FMA-NEXT:    vxorpd %ymm2, %ymm1, %ymm1
-; FMA-NEXT:    retq
-;
-; FMA4-LABEL: test_v8f64_unary_fneg_fmul_no_nsz:
-; FMA4:       # %bb.0:
-; FMA4-NEXT:    vmulpd %ymm3, %ymm1, %ymm1
-; FMA4-NEXT:    vmulpd %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vmovapd {{.*#+}} ymm2 = [-0.0E+0,-0.0E+0,-0.0E+0,-0.0E+0]
-; FMA4-NEXT:    vxorpd %ymm2, %ymm0, %ymm0
-; FMA4-NEXT:    vxorpd %ymm2, %ymm1, %ymm1
-; FMA4-NEXT:    retq
-;
-; AVX512-LABEL: test_v8f64_unary_fneg_fmul_no_nsz:
-; AVX512:       # %bb.0:
-; AVX512-NEXT:    vmulpd %zmm1, %zmm0, %zmm0
-; AVX512-NEXT:    vxorpd {{.*}}(%rip){1to8}, %zmm0, %zmm0
-; AVX512-NEXT:    retq
-  %m = fmul <8 x double> %x, %y
-  %n = fneg <8 x double> %m
-  ret <8 x double> %n
-}
-
 attributes #0 = { "unsafe-fp-math"="true" }




More information about the llvm-commits mailing list