[llvm] r276565 - [X86][SSE] Regenerate SSE copysign tests
Simon Pilgrim via llvm-commits
llvm-commits at lists.llvm.org
Sun Jul 24 08:17:50 PDT 2016
Author: rksimon
Date: Sun Jul 24 10:17:50 2016
New Revision: 276565
URL: http://llvm.org/viewvc/llvm-project?rev=276565&view=rev
Log:
[X86][SSE] Regenerate SSE copysign tests
Modified:
llvm/trunk/test/CodeGen/X86/sse-fcopysign.ll
Modified: llvm/trunk/test/CodeGen/X86/sse-fcopysign.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/sse-fcopysign.ll?rev=276565&r1=276564&r2=276565&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/sse-fcopysign.ll (original)
+++ llvm/trunk/test/CodeGen/X86/sse-fcopysign.ll Sun Jul 24 10:17:50 2016
@@ -1,3 +1,4 @@
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
; RUN: llc < %s -mtriple=i686-unknown -mattr=+sse2 | FileCheck %s --check-prefix=ALL --check-prefix=X32
; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+sse2 | FileCheck %s --check-prefix=ALL --check-prefix=X64
@@ -5,41 +6,47 @@
; Library Functions
;
-define float @tst1(float %a, float %b) {
-; X32-LABEL: @tst1
-; X32: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
-; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
-; X32-NEXT: movss %xmm1, 4(%esp)
-; X32-NEXT: movss %xmm0, (%esp)
-; X32-NEXT: calll copysignf
-; X32-NEXT: addl $8, %esp
-; X32-NEXT: retl
-;
-; X64-LABEL: @tst1
-; X64: movaps %xmm0, %xmm2
-; X64-NEXT: movaps %xmm1, %xmm0
-; X64-NEXT: movaps %xmm2, %xmm1
-; X64-NEXT: jmp copysignf
+define float @tst1(float %a, float %b) nounwind {
+; X32-LABEL: tst1:
+; X32: # BB#0:
+; X32-NEXT: subl $8, %esp
+; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; X32-NEXT: movss %xmm1, {{[0-9]+}}(%esp)
+; X32-NEXT: movss %xmm0, (%esp)
+; X32-NEXT: calll copysignf
+; X32-NEXT: addl $8, %esp
+; X32-NEXT: retl
+;
+; X64-LABEL: tst1:
+; X64: # BB#0:
+; X64-NEXT: movaps %xmm0, %xmm2
+; X64-NEXT: movaps %xmm1, %xmm0
+; X64-NEXT: movaps %xmm2, %xmm1
+; X64-NEXT: jmp copysignf # TAILCALL
%tmp = tail call float @copysignf( float %b, float %a )
ret float %tmp
}
-define double @tst2(double %a, float %b, float %c) {
-; X32-LABEL: @tst2
-; X32: movsd {{.*#+}} xmm0 = mem[0],zero
-; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
-; X32-NEXT: addss 32(%esp), %xmm1
-; X32-NEXT: cvtss2sd %xmm1, %xmm1
-; X32-NEXT: movsd %xmm0, (%esp)
-; X32-NEXT: movsd %xmm1, 8(%esp)
-; X32-NEXT: calll copysign
-; X32-NEXT: addl $16, %esp
-; X32-NEXT: retl
-;
-; X64-LABEL: @tst2
-; X64: addss %xmm2, %xmm1
-; X64-NEXT: cvtss2sd %xmm1, %xmm1
-; X64-NEXT: jmp copysign
+define double @tst2(double %a, float %b, float %c) nounwind {
+; X32-LABEL: tst2:
+; X32: # BB#0:
+; X32-NEXT: subl $16, %esp
+; X32-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero
+; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; X32-NEXT: addss {{[0-9]+}}(%esp), %xmm1
+; X32-NEXT: cvtss2sd %xmm1, %xmm1
+; X32-NEXT: movsd %xmm0, (%esp)
+; X32-NEXT: movsd %xmm1, {{[0-9]+}}(%esp)
+; X32-NEXT: calll copysign
+; X32-NEXT: addl $16, %esp
+; X32-NEXT: retl
+;
+; X64-LABEL: tst2:
+; X64: # BB#0:
+; X64-NEXT: addss %xmm2, %xmm1
+; X64-NEXT: cvtss2sd %xmm1, %xmm1
+; X64-NEXT: jmp copysign # TAILCALL
%tmp1 = fadd float %b, %c
%tmp2 = fpext float %tmp1 to double
%tmp = tail call double @copysign( double %a, double %tmp2 )
@@ -53,77 +60,90 @@ declare double @copysign(double, double)
; LLVM Intrinsic
;
-define float @int1(float %a, float %b) {
-; X32-LABEL: @int1
-; X32: movss 8(%esp), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero
-; X32-NEXT: andps .LCPI2_0, %xmm0
-; X32-NEXT: movss 12(%esp), %xmm1 {{.*#+}} xmm1 = mem[0],zero,zero,zero
-; X32-NEXT: andps .LCPI2_1, %xmm1
-; X32-NEXT: orps %xmm0, %xmm1
-; X32-NEXT: movss %xmm1, (%esp)
-; X32-NEXT: flds (%esp)
-; X32-NEXT: popl %eax
-; X32-NEXT: retl
-;
-; X64-LABEL: @int1
-; X64: andps .LCPI2_0(%rip), %xmm0
-; X64-NEXT: andps .LCPI2_1(%rip), %xmm1
-; X64-NEXT: orps %xmm1, %xmm0
-; X64-NEXT: retq
+define float @int1(float %a, float %b) nounwind {
+; X32-LABEL: int1:
+; X32: # BB#0:
+; X32-NEXT: pushl %eax
+; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; X32-NEXT: andps {{\.LCPI.*}}, %xmm0
+; X32-NEXT: movss {{.*#+}} xmm1 = mem[0],zero,zero,zero
+; X32-NEXT: andps {{\.LCPI.*}}, %xmm1
+; X32-NEXT: orps %xmm0, %xmm1
+; X32-NEXT: movss %xmm1, (%esp)
+; X32-NEXT: flds (%esp)
+; X32-NEXT: popl %eax
+; X32-NEXT: retl
+;
+; X64-LABEL: int1:
+; X64: # BB#0:
+; X64-NEXT: andps {{.*}}(%rip), %xmm0
+; X64-NEXT: andps {{.*}}(%rip), %xmm1
+; X64-NEXT: orps %xmm1, %xmm0
+; X64-NEXT: retq
%tmp = tail call float @llvm.copysign.f32( float %b, float %a )
ret float %tmp
}
-define double @int2(double %a, float %b, float %c) {
-; X32-LABEL: @int2
-; X32: movss 16(%ebp), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero
-; X32-NEXT: addss 20(%ebp), %xmm0
-; X32-NEXT: movsd 8(%ebp), %xmm1 {{.*#+}} xmm1 = mem[0],zero
-; X32-NEXT: andpd .LCPI3_0, %xmm1
-; X32-NEXT: cvtss2sd %xmm0, %xmm0
-; X32-NEXT: andpd .LCPI3_1, %xmm0
-; X32-NEXT: orpd %xmm1, %xmm0
-; X32-NEXT: movlpd %xmm0, (%esp)
-; X32-NEXT: fldl (%esp)
-; X32-NEXT: movl %ebp, %esp
-; X32-NEXT: popl %ebp
-; X32-NEXT: retl
-;
-; X64-LABEL: @int2
-; X64: addss %xmm2, %xmm1
-; X64-NEXT: cvtss2sd %xmm1, %xmm1
-; X64-NEXT: andpd .LCPI3_0(%rip), %xmm1
-; X64-NEXT: andpd .LCPI3_1(%rip), %xmm0
-; X64-NEXT: orpd %xmm1, %xmm0
-; X64-NEXT: retq
+define double @int2(double %a, float %b, float %c) nounwind {
+; X32-LABEL: int2:
+; X32: # BB#0:
+; X32-NEXT: pushl %ebp
+; X32-NEXT: movl %esp, %ebp
+; X32-NEXT: andl $-8, %esp
+; X32-NEXT: subl $8, %esp
+; X32-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; X32-NEXT: addss 20(%ebp), %xmm0
+; X32-NEXT: movsd {{.*#+}} xmm1 = mem[0],zero
+; X32-NEXT: andpd {{\.LCPI.*}}, %xmm1
+; X32-NEXT: cvtss2sd %xmm0, %xmm0
+; X32-NEXT: andpd {{\.LCPI.*}}, %xmm0
+; X32-NEXT: orpd %xmm1, %xmm0
+; X32-NEXT: movlpd %xmm0, (%esp)
+; X32-NEXT: fldl (%esp)
+; X32-NEXT: movl %ebp, %esp
+; X32-NEXT: popl %ebp
+; X32-NEXT: retl
+;
+; X64-LABEL: int2:
+; X64: # BB#0:
+; X64-NEXT: addss %xmm2, %xmm1
+; X64-NEXT: cvtss2sd %xmm1, %xmm1
+; X64-NEXT: andpd {{.*}}(%rip), %xmm1
+; X64-NEXT: andpd {{.*}}(%rip), %xmm0
+; X64-NEXT: orpd %xmm1, %xmm0
+; X64-NEXT: retq
%tmp1 = fadd float %b, %c
%tmp2 = fpext float %tmp1 to double
%tmp = tail call double @llvm.copysign.f64( double %a, double %tmp2 )
ret double %tmp
}
-define float @cst1() {
-; X32-LABEL: @cst1
-; X32: fld1
-; X32-NEXT: fchs
-; X32-NEXT: retl
-;
-; X64-LABEL: @cst1
-; X64: movss .LCPI4_0(%rip), %xmm0 {{.*#+}} xmm0 = mem[0],zero,zero,zero
-; X64-NEXT: retq
+define float @cst1() nounwind {
+; X32-LABEL: cst1:
+; X32: # BB#0:
+; X32-NEXT: fld1
+; X32-NEXT: fchs
+; X32-NEXT: retl
+;
+; X64-LABEL: cst1:
+; X64: # BB#0:
+; X64-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; X64-NEXT: retq
%tmp = tail call float @llvm.copysign.f32( float 1.0, float -2.0 )
ret float %tmp
}
-define double @cst2() {
-; X32-LABEL: @cst2
-; X32: fldz
-; X32-NEXT: fchs
-; X32-NEXT: retl
-;
-; X64-LABEL: @cst2
-; X64: movsd .LCPI5_0(%rip), %xmm0 {{.*#+}} xmm0 = mem[0],zero
-; X64-NEXT: retq
+define double @cst2() nounwind {
+; X32-LABEL: cst2:
+; X32: # BB#0:
+; X32-NEXT: fldz
+; X32-NEXT: fchs
+; X32-NEXT: retl
+;
+; X64-LABEL: cst2:
+; X64: # BB#0:
+; X64-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero
+; X64-NEXT: retq
%tmp1 = fadd float -1.0, -1.0
%tmp2 = fpext float %tmp1 to double
%tmp = tail call double @llvm.copysign.f64( double 0.0, double %tmp2 )
More information about the llvm-commits
mailing list