; RUN: llc < %s -emulated-tls -mtriple=i386-linux-gnu | FileCheck -check-prefix=X86 %s ; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=X64 %s ; RUN: llc < %s -emulated-tls -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s ; RUN: llc < %s -emulated-tls -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s ; RUN: llc < %s -mtriple=i386-linux-gnu | FileCheck -check-prefix=NoEMU %s ; RUN: llc < %s -mtriple=x86_64-linux-gnu | FileCheck -check-prefix=NoEMU %s ; RUN: llc < %s -mtriple=i386-linux-android | FileCheck -check-prefix=X86 %s ; RUN: llc < %s -mtriple=x86_64-linux-android | FileCheck -check-prefix=X64 %s ; Copied from tls.ll; emulated TLS model is not implemented ; for *-pc-win32 and *-pc-windows targets yet. ; NoEMU-NOT: __emutls ; Use my_emutls_get_address like __emutls_get_address. @my_emutls_v_xyz = external global ptr, align 4 declare ptr @my_emutls_get_address(ptr) define dso_local i32 @my_get_xyz() { ; X86-LABEL: my_get_xyz: ; X86: movl $my_emutls_v_xyz, (%esp) ; X86-NEXT: calll my_emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: my_get_xyz: ; X64: movq my_emutls_v_xyz@GOTPCREL(%rip), %rdi ; X64-NEXT: callq my_emutls_get_address ; X64-NEXT: movl (%rax), %eax ; X64-NEXT: popq %rcx ; X64-NEXT: .cfi_def_cfa_offset 8 ; X64-NEXT: retq entry: %call = call ptr @my_emutls_get_address(ptr @my_emutls_v_xyz) %0 = load i32, ptr %call, align 4 ret i32 %0 } @i1 = dso_local thread_local global i32 15 @i2 = external thread_local global i32 @i3 = internal thread_local global i32 15 @i4 = hidden thread_local global i32 15 @i5 = external hidden thread_local global i32 @s1 = dso_local thread_local global i16 15 @b1 = dso_local thread_local global i8 0 define dso_local i32 @f1() { ; X86-LABEL: f1: ; X86: movl $__emutls_v.i1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: f1: ; X64: movl $__emutls_v.i1, %edi ; X64-NEXT: callq __emutls_get_address ; X64-NEXT: movl (%rax), %eax ; X64-NEXT: popq %rcx ; X64-NEXT: .cfi_def_cfa_offset 8 ; X64-NEXT: retq entry: %tmp1 = load i32, ptr @i1 ret i32 %tmp1 } define dso_local ptr @f2() { ; X86-LABEL: f2: ; X86: movl $__emutls_v.i1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl ; ; X64-LABEL: f2: ; X64: movl $__emutls_v.i1, %edi ; X64-NEXT: callq __emutls_get_address ; X64-NEXT: popq %rcx ; X64-NEXT: .cfi_def_cfa_offset 8 ; X64-NEXT: retq entry: ret ptr @i1 } define dso_local i32 @f3() nounwind { ; X86-LABEL: f3: ; X86: movl $__emutls_v.i2, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: retl entry: %tmp1 = load i32, ptr @i2 ret i32 %tmp1 } define dso_local ptr @f4() { ; X86-LABEL: f4: ; X86: movl $__emutls_v.i2, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: ret ptr @i2 } define dso_local i32 @f5() nounwind { ; X86-LABEL: f5: ; X86: movl $__emutls_v.i3, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: retl entry: %tmp1 = load i32, ptr @i3 ret i32 %tmp1 } define dso_local ptr @f6() { ; X86-LABEL: f6: ; X86: movl $__emutls_v.i3, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: ret ptr @i3 } define dso_local i32 @f7() { ; X86-LABEL: f7: ; X86: movl $__emutls_v.i4, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i32, ptr @i4 ret i32 %tmp1 } define dso_local ptr @f8() { ; X86-LABEL: f8: ; X86: movl $__emutls_v.i4, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: ret ptr @i4 } define dso_local i32 @f9() { ; X86-LABEL: f9: ; X86: movl $__emutls_v.i5, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i32, ptr @i5 ret i32 %tmp1 } define dso_local ptr @f10() { ; X86-LABEL: f10: ; X86: movl $__emutls_v.i5, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: ret ptr @i5 } define dso_local i16 @f11() { ; X86-LABEL: f11: ; X86: movl $__emutls_v.s1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movzwl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i16, ptr @s1 ret i16 %tmp1 } define dso_local i32 @f12() { ; X86-LABEL: f12: ; X86: movl $__emutls_v.s1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movswl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i16, ptr @s1 %tmp2 = sext i16 %tmp1 to i32 ret i32 %tmp2 } define dso_local i8 @f13() { ; X86-LABEL: f13: ; X86: movl $__emutls_v.b1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movzbl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i8, ptr @b1 ret i8 %tmp1 } define dso_local i32 @f14() { ; X86-LABEL: f14: ; X86: movl $__emutls_v.b1, (%esp) ; X86-NEXT: calll __emutls_get_address ; X86-NEXT: movsbl (%eax), %eax ; X86-NEXT: addl $12, %esp ; X86-NEXT: .cfi_def_cfa_offset 4 ; X86-NEXT: retl entry: %tmp1 = load i8, ptr @b1 %tmp2 = sext i8 %tmp1 to i32 ret i32 %tmp2 } ;;;;;;;;;;;;;; 32-bit __emutls_v. and __emutls_t. ; X86-LABEL: __emutls_v.i1: ; X86-NEXT: .long 4 ; X86-NEXT: .long 4 ; X86-NEXT: .long 0 ; X86-NEXT: .long __emutls_t.i1 ; X86-LABEL: __emutls_t.i1: ; X86-NEXT: .long 15 ; X86-NOT: __emutls_v.i2 ; X86-LABEL: __emutls_v.i3: ; X86-NEXT: .long 4 ; X86-NEXT: .long 4 ; X86-NEXT: .long 0 ; X86-NEXT: .long __emutls_t.i3 ; X86-LABEL: __emutls_t.i3: ; X86-NEXT: .long 15 ; X86-LABEL: __emutls_v.i4: ; X86-NEXT: .long 4 ; X86-NEXT: .long 4 ; X86-NEXT: .long 0 ; X86-NEXT: .long __emutls_t.i4 ; X86-LABEL: __emutls_t.i4: ; X86-NEXT: .long 15 ; X86-NOT: __emutls_v.i5: ; X86: .hidden __emutls_v.i5 ; X86-NOT: __emutls_v.i5: ; X86-LABEL: __emutls_v.s1: ; X86-NEXT: .long 2 ; X86-NEXT: .long 2 ; X86-NEXT: .long 0 ; X86-NEXT: .long __emutls_t.s1 ; X86-LABEL: __emutls_t.s1: ; X86-NEXT: .short 15 ; X86-LABEL: __emutls_v.b1: ; X86-NEXT: .long 1 ; X86-NEXT: .long 1 ; X86-NEXT: .long 0 ; X86-NEXT: .long 0 ; X86-NOT: __emutls_t.b1 ;;;;;;;;;;;;;; 64-bit __emutls_v. and __emutls_t. ; X64-LABEL: __emutls_v.i1: ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 0 ; X64-NEXT: .quad __emutls_t.i1 ; X64-LABEL: __emutls_t.i1: ; X64-NEXT: .long 15 ; X64-NOT: __emutls_v.i2 ; X64-LABEL: __emutls_v.i3: ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 0 ; X64-NEXT: .quad __emutls_t.i3 ; X64-LABEL: __emutls_t.i3: ; X64-NEXT: .long 15 ; X64-LABEL: __emutls_v.i4: ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 4 ; X64-NEXT: .quad 0 ; X64-NEXT: .quad __emutls_t.i4 ; X64-LABEL: __emutls_t.i4: ; X64-NEXT: .long 15 ; X64-NOT: __emutls_v.i5: ; X64: .hidden __emutls_v.i5 ; X64-NOT: __emutls_v.i5: ; X64-LABEL: __emutls_v.s1: ; X64-NEXT: .quad 2 ; X64-NEXT: .quad 2 ; X64-NEXT: .quad 0 ; X64-NEXT: .quad __emutls_t.s1 ; X64-LABEL: __emutls_t.s1: ; X64-NEXT: .short 15 ; X64-LABEL: __emutls_v.b1: ; X64-NEXT: .quad 1 ; X64-NEXT: .quad 1 ; X64-NEXT: .quad 0 ; X64-NEXT: .quad 0 ; X64-NOT: __emutls_t.b1