[llvm] d035bc3 - [X86] emutls.ll - replace X32 check prefix with X86. NFC.

Simon Pilgrim via llvm-commits llvm-commits at lists.llvm.org
Tue Nov 17 10:34:17 PST 2020


Author: Simon Pilgrim
Date: 2020-11-17T18:32:43Z
New Revision: d035bc3e6b87db281da610f64392517036bf5411

URL: https://github.com/llvm/llvm-project/commit/d035bc3e6b87db281da610f64392517036bf5411
DIFF: https://github.com/llvm/llvm-project/commit/d035bc3e6b87db281da610f64392517036bf5411.diff

LOG: [X86] emutls.ll - replace X32 check prefix with X86. NFC.

We typically use X32 for gnux32 triples

Added: 
    

Modified: 
    llvm/test/CodeGen/X86/emutls.ll

Removed: 
    


################################################################################
diff  --git a/llvm/test/CodeGen/X86/emutls.ll b/llvm/test/CodeGen/X86/emutls.ll
index 1e706c1267d1..329a74733c55 100644
--- a/llvm/test/CodeGen/X86/emutls.ll
+++ b/llvm/test/CodeGen/X86/emutls.ll
@@ -1,11 +1,11 @@
-; RUN: llc < %s -emulated-tls -mtriple=i386-linux-gnu | FileCheck -check-prefix=X32 %s
+; RUN: llc < %s -emulated-tls -mtriple=i386-linux-gnu | FileCheck -check-prefix=X86 %s
 ; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=X64 %s
-; RUN: llc < %s -emulated-tls -mtriple=i386-linux-android | FileCheck -check-prefix=X32 %s
+; RUN: llc < %s -emulated-tls -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s
 ; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s
 
 ; RUN: llc < %s -mtriple=i386-linux-gnu | FileCheck -check-prefix=NoEMU %s
 ; RUN: llc < %s -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=NoEMU %s
-; RUN: llc < %s -mtriple=i386-linux-android | FileCheck -check-prefix=X32 %s
+; RUN: llc < %s -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s
 ; RUN: llc < %s -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s
 
 ; Copied from tls.ll; emulated TLS model is not implemented
@@ -18,13 +18,13 @@
 declare i8* @my_emutls_get_address(i8*)
 
 define i32 @my_get_xyz() {
-; X32-LABEL: my_get_xyz:
-; X32:         movl $my_emutls_v_xyz, (%esp)
-; X32-NEXT:    calll my_emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: my_get_xyz:
+; X86:         movl $my_emutls_v_xyz, (%esp)
+; X86-NEXT:    calll my_emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 ;
 ; X64-LABEL: my_get_xyz:
 ; X64:         movl $my_emutls_v_xyz, %edi
@@ -49,13 +49,13 @@ entry:
 @b1 = thread_local global i8 0
 
 define i32 @f1() {
-; X32-LABEL: f1:
-; X32:         movl $__emutls_v.i1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f1:
+; X86:         movl $__emutls_v.i1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 ;
 ; X64-LABEL: f1:
 ; X64:         movl $__emutls_v.i1, %edi
@@ -70,12 +70,12 @@ entry:
 }
 
 define i32* @f2() {
-; X32-LABEL: f2:
-; X32:         movl $__emutls_v.i1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f2:
+; X86:         movl $__emutls_v.i1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 ;
 ; X64-LABEL: f2:
 ; X64:         movl $__emutls_v.i1, %edi
@@ -88,120 +88,120 @@ entry:
 }
 
 define i32 @f3() nounwind {
-; X32-LABEL: f3:
-; X32:         movl $__emutls_v.i2, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    retl
+; X86-LABEL: f3:
+; X86:         movl $__emutls_v.i2, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i32, i32* @i2
   ret i32 %tmp1
 }
 
 define i32* @f4() {
-; X32-LABEL: f4:
-; X32:         movl $__emutls_v.i2, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f4:
+; X86:         movl $__emutls_v.i2, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   ret i32* @i2
 }
 
 define i32 @f5() nounwind {
-; X32-LABEL: f5:
-; X32:         movl $__emutls_v.i3, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    retl
+; X86-LABEL: f5:
+; X86:         movl $__emutls_v.i3, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i32, i32* @i3
   ret i32 %tmp1
 }
 
 define i32* @f6() {
-; X32-LABEL: f6:
-; X32:         movl $__emutls_v.i3, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f6:
+; X86:         movl $__emutls_v.i3, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   ret i32* @i3
 }
 
 define i32 @f7() {
-; X32-LABEL: f7:
-; X32:         movl $__emutls_v.i4, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f7:
+; X86:         movl $__emutls_v.i4, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i32, i32* @i4
   ret i32 %tmp1
 }
 
 define i32* @f8() {
-; X32-LABEL: f8:
-; X32:         movl $__emutls_v.i4, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f8:
+; X86:         movl $__emutls_v.i4, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   ret i32* @i4
 }
 
 define i32 @f9() {
-; X32-LABEL: f9:
-; X32:         movl $__emutls_v.i5, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f9:
+; X86:         movl $__emutls_v.i5, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i32, i32* @i5
   ret i32 %tmp1
 }
 
 define i32* @f10() {
-; X32-LABEL: f10:
-; X32:         movl $__emutls_v.i5, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f10:
+; X86:         movl $__emutls_v.i5, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   ret i32* @i5
 }
 
 define i16 @f11() {
-; X32-LABEL: f11:
-; X32:         movl $__emutls_v.s1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movzwl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f11:
+; X86:         movl $__emutls_v.s1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movzwl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i16, i16* @s1
   ret i16 %tmp1
 }
 
 define i32 @f12() {
-; X32-LABEL: f12:
-; X32:         movl $__emutls_v.s1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movswl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f12:
+; X86:         movl $__emutls_v.s1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movswl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i16, i16* @s1
   %tmp2 = sext i16 %tmp1 to i32
@@ -209,26 +209,26 @@ entry:
 }
 
 define i8 @f13() {
-; X32-LABEL: f13:
-; X32:         movl $__emutls_v.b1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movb (%eax), %al
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f13:
+; X86:         movl $__emutls_v.b1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movb (%eax), %al
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i8, i8* @b1
   ret i8 %tmp1
 }
 
 define i32 @f14() {
-; X32-LABEL: f14:
-; X32:         movl $__emutls_v.b1, (%esp)
-; X32-NEXT:    calll __emutls_get_address
-; X32-NEXT:    movsbl (%eax), %eax
-; X32-NEXT:    addl $12, %esp
-; X32-NEXT:    .cfi_def_cfa_offset 4
-; X32-NEXT:    retl
+; X86-LABEL: f14:
+; X86:         movl $__emutls_v.b1, (%esp)
+; X86-NEXT:    calll __emutls_get_address
+; X86-NEXT:    movsbl (%eax), %eax
+; X86-NEXT:    addl $12, %esp
+; X86-NEXT:    .cfi_def_cfa_offset 4
+; X86-NEXT:    retl
 entry:
   %tmp1 = load i8, i8* @b1
   %tmp2 = sext i8 %tmp1 to i32
@@ -237,55 +237,55 @@ entry:
 
 ;;;;;;;;;;;;;; 32-bit __emutls_v. and __emutls_t.
 
-; X32-LABEL: __emutls_v.i1:
-; X32-NEXT: .long 4
-; X32-NEXT: .long 4
-; X32-NEXT: .long 0
-; X32-NEXT: .long __emutls_t.i1
+; X86-LABEL: __emutls_v.i1:
+; X86-NEXT: .long 4
+; X86-NEXT: .long 4
+; X86-NEXT: .long 0
+; X86-NEXT: .long __emutls_t.i1
 
-; X32-LABEL: __emutls_t.i1:
-; X32-NEXT: .long 15
+; X86-LABEL: __emutls_t.i1:
+; X86-NEXT: .long 15
 
-; X32-NOT:   __emutls_v.i2
+; X86-NOT:   __emutls_v.i2
 
-; X32-LABEL: __emutls_v.i3:
-; X32-NEXT: .long 4
-; X32-NEXT: .long 4
-; X32-NEXT: .long 0
-; X32-NEXT: .long __emutls_t.i3
+; X86-LABEL: __emutls_v.i3:
+; X86-NEXT: .long 4
+; X86-NEXT: .long 4
+; X86-NEXT: .long 0
+; X86-NEXT: .long __emutls_t.i3
 
-; X32-LABEL: __emutls_t.i3:
-; X32-NEXT: .long 15
+; X86-LABEL: __emutls_t.i3:
+; X86-NEXT: .long 15
 
-; X32-LABEL: __emutls_v.i4:
-; X32-NEXT: .long 4
-; X32-NEXT: .long 4
-; X32-NEXT: .long 0
-; X32-NEXT: .long __emutls_t.i4
+; X86-LABEL: __emutls_v.i4:
+; X86-NEXT: .long 4
+; X86-NEXT: .long 4
+; X86-NEXT: .long 0
+; X86-NEXT: .long __emutls_t.i4
 
-; X32-LABEL: __emutls_t.i4:
-; X32-NEXT: .long 15
+; X86-LABEL: __emutls_t.i4:
+; X86-NEXT: .long 15
 
-; X32-NOT:   __emutls_v.i5:
-; X32:      .hidden __emutls_v.i5
-; X32-NOT:   __emutls_v.i5:
+; X86-NOT:   __emutls_v.i5:
+; X86:      .hidden __emutls_v.i5
+; X86-NOT:   __emutls_v.i5:
 
-; X32-LABEL: __emutls_v.s1:
-; X32-NEXT: .long 2
-; X32-NEXT: .long 2
-; X32-NEXT: .long 0
-; X32-NEXT: .long __emutls_t.s1
+; X86-LABEL: __emutls_v.s1:
+; X86-NEXT: .long 2
+; X86-NEXT: .long 2
+; X86-NEXT: .long 0
+; X86-NEXT: .long __emutls_t.s1
 
-; X32-LABEL: __emutls_t.s1:
-; X32-NEXT: .short 15
+; X86-LABEL: __emutls_t.s1:
+; X86-NEXT: .short 15
 
-; X32-LABEL: __emutls_v.b1:
-; X32-NEXT: .long 1
-; X32-NEXT: .long 1
-; X32-NEXT: .long 0
-; X32-NEXT: .long 0
+; X86-LABEL: __emutls_v.b1:
+; X86-NEXT: .long 1
+; X86-NEXT: .long 1
+; X86-NEXT: .long 0
+; X86-NEXT: .long 0
 
-; X32-NOT:   __emutls_t.b1
+; X86-NOT:   __emutls_t.b1
 
 ;;;;;;;;;;;;;; 64-bit __emutls_v. and __emutls_t.
 


        


More information about the llvm-commits mailing list