[llvm] r283883 - [X86][SSE] Regenerate scalar i64 uitofp test

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Tue Oct 11 07:01:38 PDT 2016


Author: rksimon
Date: Tue Oct 11 09:01:38 2016
New Revision: 283883

URL: http://llvm.org/viewvc/llvm-project?rev=283883&view=rev
Log:
[X86][SSE] Regenerate scalar i64 uitofp test

Added 32-bit target test

Modified:
    llvm/trunk/test/CodeGen/X86/uint64-to-float.ll

Modified: llvm/trunk/test/CodeGen/X86/uint64-to-float.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/uint64-to-float.ll?rev=283883&r1=283882&r2=283883&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/uint64-to-float.ll (original)
+++ llvm/trunk/test/CodeGen/X86/uint64-to-float.ll Tue Oct 11 09:01:38 2016
@@ -1,23 +1,48 @@
-; RUN: llc < %s -mcpu=generic -march=x86-64 | FileCheck %s
+; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
+; RUN: llc < %s -mtriple=i686-apple-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X86
+; RUN: llc < %s -mtriple=x86_64-apple-unknown -mattr=+sse2 | FileCheck %s --check-prefix=X64
+
 ; Verify that we are using the efficient uitofp --> sitofp lowering illustrated
 ; by the compiler_rt implementation of __floatundisf.
 ; <rdar://problem/8493982>
 
-target datalayout = "e-p:64:64:64-i1:8:8-i8:8:8-i16:16:16-i32:32:32-i64:64:64-f32:32:32-f64:64:64-v64:64:64-v128:128:128-a0:0:64-s0:64:64-f80:128:128-n8:16:32:64"
-target triple = "x86_64-apple-darwin10.0.0"
-
-; CHECK: %entry
-; CHECK-NEXT: testq %rdi, %rdi
-; CHECK-NEXT: js LBB0_1
-; CHECK: cvtsi2ss
-; CHECK-NEXT: ret
-; CHECK: LBB0_1
-; CHECK: movq
-; CHECK-NEXT: shrq
-; CHECK-NEXT: andl
-; CHECK-NEXT: orq
-; CHECK-NEXT: cvtsi2ss
-define float @test(i64 %a) {
+define float @test(i64 %a) nounwind {
+; X86-LABEL: test:
+; X86:       # BB#0: # %entry
+; X86-NEXT:    pushl %ebp
+; X86-NEXT:    movl %esp, %ebp
+; X86-NEXT:    andl $-8, %esp
+; X86-NEXT:    subl $16, %esp
+; X86-NEXT:    movq {{.*#+}} xmm0 = mem[0],zero
+; X86-NEXT:    movq %xmm0, {{[0-9]+}}(%esp)
+; X86-NEXT:    xorl %eax, %eax
+; X86-NEXT:    cmpl $0, 12(%ebp)
+; X86-NEXT:    setns %al
+; X86-NEXT:    fildll {{[0-9]+}}(%esp)
+; X86-NEXT:    fadds {{\.LCPI.*}}(,%eax,4)
+; X86-NEXT:    fstps {{[0-9]+}}(%esp)
+; X86-NEXT:    movss {{.*#+}} xmm0 = mem[0],zero,zero,zero
+; X86-NEXT:    movss %xmm0, (%esp)
+; X86-NEXT:    flds (%esp)
+; X86-NEXT:    movl %ebp, %esp
+; X86-NEXT:    popl %ebp
+; X86-NEXT:    retl
+;
+; X64-LABEL: test:
+; X64:       # BB#0: # %entry
+; X64-NEXT:    testq %rdi, %rdi
+; X64-NEXT:    js .LBB0_1
+; X64-NEXT:  # BB#2: # %entry
+; X64-NEXT:    cvtsi2ssq %rdi, %xmm0
+; X64-NEXT:    retq
+; X64-NEXT:  .LBB0_1:
+; X64-NEXT:    movq %rdi, %rax
+; X64-NEXT:    shrq %rax
+; X64-NEXT:    andl $1, %edi
+; X64-NEXT:    orq %rax, %rdi
+; X64-NEXT:    cvtsi2ssq %rdi, %xmm0
+; X64-NEXT:    addss %xmm0, %xmm0
+; X64-NEXT:    retq
 entry:
   %b = uitofp i64 %a to float
   ret float %b




More information about the llvm-commits mailing list