[X86] Cleanup emulated TLS test checks. NFC
diff --git a/llvm/test/CodeGen/X86/emutls.ll b/llvm/test/CodeGen/X86/emutls.ll
index 67e80a3..d11f65a 100644
--- a/llvm/test/CodeGen/X86/emutls.ll
+++ b/llvm/test/CodeGen/X86/emutls.ll
@@ -25,6 +25,7 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
+;
; X64-LABEL: my_get_xyz:
; X64: movl $my_emutls_v_xyz, %edi
; X64-NEXT: callq my_emutls_get_address
@@ -32,7 +33,6 @@
; X64-NEXT: popq %rcx
; X64-NEXT: .cfi_def_cfa_offset 8
; X64-NEXT: retq
-
entry:
%call = call i8* @my_emutls_get_address(i8* bitcast (i8** @my_emutls_v_xyz to i8*))
%0 = bitcast i8* %call to i32*
@@ -56,6 +56,7 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
+;
; X64-LABEL: f1:
; X64: movl $__emutls_v.i1, %edi
; X64-NEXT: callq __emutls_get_address
@@ -63,7 +64,6 @@
; X64-NEXT: popq %rcx
; X64-NEXT: .cfi_def_cfa_offset 8
; X64-NEXT: retq
-
entry:
%tmp1 = load i32, i32* @i1
ret i32 %tmp1
@@ -76,13 +76,13 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
+;
; X64-LABEL: f2:
; X64: movl $__emutls_v.i1, %edi
; X64-NEXT: callq __emutls_get_address
; X64-NEXT: popq %rcx
; X64-NEXT: .cfi_def_cfa_offset 8
; X64-NEXT: retq
-
entry:
ret i32* @i1
}
@@ -94,7 +94,6 @@
; X32-NEXT: movl (%eax), %eax
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
-
entry:
%tmp1 = load i32, i32* @i2
ret i32 %tmp1
@@ -107,7 +106,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
ret i32* @i2
}
@@ -119,7 +117,6 @@
; X32-NEXT: movl (%eax), %eax
; X32-NEXT: addl $12, %esp
; X32-NEXT: retl
-
entry:
%tmp1 = load i32, i32* @i3
ret i32 %tmp1
@@ -132,7 +129,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
ret i32* @i3
}
@@ -145,7 +141,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i32, i32* @i4
ret i32 %tmp1
@@ -158,7 +153,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
ret i32* @i4
}
@@ -171,7 +165,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i32, i32* @i5
ret i32 %tmp1
@@ -184,7 +177,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
ret i32* @i5
}
@@ -197,7 +189,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i16, i16* @s1
ret i16 %tmp1
@@ -211,7 +202,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i16, i16* @s1
%tmp2 = sext i16 %tmp1 to i32
@@ -226,7 +216,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i8, i8* @b1
ret i8 %tmp1
@@ -240,7 +229,6 @@
; X32-NEXT: addl $12, %esp
; X32-NEXT: .cfi_def_cfa_offset 4
; X32-NEXT: retl
-
entry:
%tmp1 = load i8, i8* @b1
%tmp2 = sext i8 %tmp1 to i32