[llvm] r297475 - [X86][MMX] Updated bad stack spill shift value test to actually show the problem

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Fri Mar 10 08:18:50 PST 2017


Author: rksimon
Date: Fri Mar 10 10:18:50 2017
New Revision: 297475

URL: http://llvm.org/viewvc/llvm-project?rev=297475&view=rev
Log:
[X86][MMX] Updated bad stack spill shift value test to actually show the problem

Cleaning up the ir had stopped showing the issue.

Modified:
    llvm/trunk/test/CodeGen/X86/mmx-fold-load.ll

Modified: llvm/trunk/test/CodeGen/X86/mmx-fold-load.ll
URL: http://llvm.org/viewvc/llvm-project/llvm/trunk/test/CodeGen/X86/mmx-fold-load.ll?rev=297475&r1=297474&r2=297475&view=diff
==============================================================================
--- llvm/trunk/test/CodeGen/X86/mmx-fold-load.ll (original)
+++ llvm/trunk/test/CodeGen/X86/mmx-fold-load.ll Fri Mar 10 10:18:50 2017
@@ -565,19 +565,20 @@ declare x86_mmx @llvm.x86.mmx.psrl.q(x86
 
 ; FIXME: Show issue with storing i32 to stack and then reloading as x86_mmx
 ; which will lead to garbage in the other 32-bits.
-define i64 @test_psrlq_by_volatile_shift_amount(x86_mmx %t) nounwind {
+define void @test_psrlq_by_volatile_shift_amount(x86_mmx* %t) nounwind {
 ; X86-LABEL: test_psrlq_by_volatile_shift_amount:
 ; X86:       # BB#0: # %entry
 ; X86-NEXT:    pushl %ebp
 ; X86-NEXT:    movl %esp, %ebp
 ; X86-NEXT:    andl $-8, %esp
 ; X86-NEXT:    subl $16, %esp
+; X86-NEXT:    movl 8(%ebp), %eax
 ; X86-NEXT:    movl $1, {{[0-9]+}}(%esp)
-; X86-NEXT:    movd {{[0-9]+}}(%esp), %mm1
-; X86-NEXT:    psrlq %mm1, %mm0
-; X86-NEXT:    movq %mm0, {{[0-9]+}}(%esp)
-; X86-NEXT:    movl {{[0-9]+}}(%esp), %eax
-; X86-NEXT:    movl {{[0-9]+}}(%esp), %edx
+; X86-NEXT:    movl $0, {{[0-9]+}}(%esp)
+; X86-NEXT:    movl $255, {{[0-9]+}}(%esp)
+; X86-NEXT:    movq {{[0-9]+}}(%esp), %mm0
+; X86-NEXT:    psrlq {{[0-9]+}}(%esp), %mm0
+; X86-NEXT:    movq %mm0, (%eax)
 ; X86-NEXT:    movl %ebp, %esp
 ; X86-NEXT:    popl %ebp
 ; X86-NEXT:    retl
@@ -585,9 +586,10 @@ define i64 @test_psrlq_by_volatile_shift
 ; X64-LABEL: test_psrlq_by_volatile_shift_amount:
 ; X64:       # BB#0: # %entry
 ; X64-NEXT:    movl $1, -{{[0-9]+}}(%rsp)
-; X64-NEXT:    movd -{{[0-9]+}}(%rsp), %mm1
-; X64-NEXT:    psrlq %mm1, %mm0
-; X64-NEXT:    movd %mm0, %rax
+; X64-NEXT:    movl $255, %eax
+; X64-NEXT:    movd %rax, %mm0
+; X64-NEXT:    psrlq -{{[0-9]+}}(%rsp), %mm0
+; X64-NEXT:    movq %mm0, (%rdi)
 ; X64-NEXT:    retq
 entry:
   %0 = alloca i32, align 4
@@ -595,10 +597,10 @@ entry:
   call void @llvm.lifetime.start(i64 4, i8* nonnull %1)
   store volatile i32 1, i32* %0, align 4
   %2 = load volatile i32, i32* %0, align 4
-  %3 = tail call x86_mmx @llvm.x86.mmx.psrli.q(x86_mmx %t, i32 %2)
-  %4 = bitcast x86_mmx %3 to i64
+  %3 = tail call x86_mmx @llvm.x86.mmx.psrli.q(x86_mmx bitcast (<1 x i64> <i64 255> to x86_mmx), i32 %2)
+  store x86_mmx %3, x86_mmx* %t, align 8
   call void @llvm.lifetime.end(i64 4, i8* nonnull %1)
-  ret i64 %4
+  ret void
 }
 
 declare void @llvm.lifetime.start(i64, i8* nocapture)




More information about the llvm-commits mailing list