; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=tahiti -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,LOOP %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=hawaii -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,LOOP %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=fiji -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,LOOP %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx900 -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,NOLOOP,NOLOOP-SDAG %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx1010 -asm-verbose=0 -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,NOLOOP,NOLOOP-SDAG,GFX10 %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx1100 -asm-verbose=0 -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefixes=GCN,NOLOOP,NOLOOP-SDAG,GFX10 %s ; Make sure the op is emitted bundled with a waitcnt with and without the retry loop, and the bundle is not removed by ExpandPostRAPseudos. ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=tahiti -stop-after=postrapseudos -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefix=MIR %s ; RUN: llc -mtriple=amdgcn-mesa-mesa3d -mcpu=gfx900 -stop-after=postrapseudos -o - -verify-machineinstrs < %s | FileCheck -enable-var-scope -check-prefix=MIR %s ; Minimum offset ; GCN-LABEL: {{^}}gws_barrier_offset0: ; NOLOOP-DAG: s_load_{{dword|b32}} [[BAR_NUM:s[0-9]+]] ; NOLOOP-DAG: s_mov_b32 m0, 0{{$}} ; NOLOOP: v_mov_b32_e32 v0, [[BAR_NUM]] ; NOLOOP: ds_gws_barrier v0 gds{{$}} ; LOOP: s_mov_b32 m0, 0{{$}} ; LOOP: [[LOOP:.LBB[0-9]+_[0-9]+]]: ; LOOP-NEXT: s_setreg_imm32_b32 hwreg(HW_REG_TRAPSTS, 8, 1), 0 ; LOOP-NEXT: ds_gws_barrier v0 gds ; LOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; LOOP-NEXT: s_getreg_b32 [[GETREG:s[0-9]+]], hwreg(HW_REG_TRAPSTS, 8, 1) ; LOOP-NEXT: s_cmp_lg_u32 [[GETREG]], 0 ; LOOP-NEXT: s_cbranch_scc1 [[LOOP]] ; MIR-LABEL: name: gws_barrier_offset0{{$}} ; MIR: BUNDLE implicit{{( killed)?( renamable)?}} $vgpr0, implicit $m0, implicit $exec { ; MIR-NEXT: DS_GWS_BARRIER renamable $vgpr0, 0, implicit $m0, implicit $exec :: (load (s32) from custom "GWSResource") ; MIR-NEXT: S_WAITCNT 0 ; MIR-NEXT: } define amdgpu_kernel void @gws_barrier_offset0(i32 %val) #0 { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 0) ret void } ; MIR-LABEL: name: gws_barrier_offset63{{$}} ; Maximum offset ; GCN-LABEL: {{^}}gws_barrier_offset63: ; NOLOOP-DAG: s_load_{{dword|b32}} [[BAR_NUM:s[0-9]+]] ; NOLOOP-DAG: s_mov_b32 m0, 0{{$}} ; NOLOOP-DAG: v_mov_b32_e32 v0, [[BAR_NUM]] ; NOLOOP: ds_gws_barrier v0 offset:63 gds{{$}} define amdgpu_kernel void @gws_barrier_offset63(i32 %val) #0 { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 63) ret void } ; FIXME: Should be able to shift directly into m0 ; GCN-LABEL: {{^}}gws_barrier_sgpr_offset: ; NOLOOP-DAG: s_load_{{dwordx2|b64}} s[[[BAR_NUM:[0-9]+]]:[[OFFSET:[0-9]+]]] ; NOLOOP-SDAG-DAG: s_lshl_b32 [[SHL:s[0-9]+]], s[[OFFSET]], 16 ; NOLOOP-SDAG-DAG: s_mov_b32 m0, [[SHL]]{{$}} ; NOLOOP-GISEL-DAG: s_lshl_b32 m0, s[[OFFSET]], 16 ; NOLOOP-DAG: v_mov_b32_e32 [[GWS_VAL:v[0-9]+]], s[[BAR_NUM]] ; NOLOOP: ds_gws_barrier [[GWS_VAL]] gds{{$}} define amdgpu_kernel void @gws_barrier_sgpr_offset(i32 %val, i32 %offset) #0 { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 %offset) ret void } ; Variable offset in SGPR with constant add ; GCN-LABEL: {{^}}gws_barrier_sgpr_offset_add1: ; NOLOOP-DAG: s_load_{{dwordx2|b64}} s[[[BAR_NUM:[0-9]+]]:[[OFFSET:[0-9]+]]] ; NOLOOP-SDAG-DAG: s_lshl_b32 [[SHL:s[0-9]+]], s[[OFFSET]], 16 ; NOLOOP-SDAG-DAG: s_mov_b32 m0, [[SHL]]{{$}} ; NOLOOP-GISEL-DAG: s_lshl_b32 m0, s[[OFFSET]], 16 ; NOLOOP-DAG: v_mov_b32_e32 [[GWS_VAL:v[0-9]+]], s[[BAR_NUM]] ; NOLOOP: ds_gws_barrier [[GWS_VAL]] offset:1 gds{{$}} define amdgpu_kernel void @gws_barrier_sgpr_offset_add1(i32 %val, i32 %offset.base) #0 { %offset = add i32 %offset.base, 1 call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 %offset) ret void } ; GCN-LABEL: {{^}}gws_barrier_vgpr_offset: ; NOLOOP-DAG: s_load_{{dword|b32}} [[BAR_NUM:s[0-9]+]] ; NOLOOP-DAG: v_readfirstlane_b32 [[READLANE:s[0-9]+]], v0 ; NOLOOP-SDAG-DAG: s_lshl_b32 [[SHL:s[0-9]+]], [[READLANE]], 16 ; NOLOOP-SDAG-DAG: s_mov_b32 m0, [[SHL]]{{$}} ; NOLOOP-GISEL-DAG: s_lshl_b32 m0, [[READLANE]], 16 ; NOLOOP-DAG: v_mov_b32_e32 [[GWS_VAL:v[0-9]+]], [[BAR_NUM]] ; NOLOOP: ds_gws_barrier [[GWS_VAL]] gds{{$}} define amdgpu_kernel void @gws_barrier_vgpr_offset(i32 %val) #0 { %vgpr.offset = call i32 @llvm.amdgcn.workitem.id.x() call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 %vgpr.offset) ret void } ; Variable offset in VGPR with constant add ; GCN-LABEL: {{^}}gws_barrier_vgpr_offset_add: ; NOLOOP-DAG: s_load_{{dword|b32}} [[BAR_NUM:s[0-9]+]] ; NOLOOP-DAG: v_readfirstlane_b32 [[READLANE:s[0-9]+]], v0 ; NOLOOP-SDAG-DAG: s_lshl_b32 [[SHL:s[0-9]+]], [[READLANE]], 16 ; NOLOOP-SDAG-DAG: s_mov_b32 m0, [[SHL]]{{$}} ; NOLOOP-GISEL-DAG: s_lshl_b32 m0, [[READLANE]], 16 ; NOLOOP-DAG: v_mov_b32_e32 [[GWS_VAL:v[0-9]+]], [[BAR_NUM]] ; NOLOOP: ds_gws_barrier [[GWS_VAL]] offset:3 gds{{$}} define amdgpu_kernel void @gws_barrier_vgpr_offset_add(i32 %val) #0 { %vgpr.offset.base = call i32 @llvm.amdgcn.workitem.id.x() %vgpr.offset = add i32 %vgpr.offset.base, 3 call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 %vgpr.offset) ret void } @lds = internal unnamed_addr addrspace(3) global i32 undef ; Check if m0 initialization is shared ; GCN-LABEL: {{^}}gws_barrier_save_m0_barrier_constant_offset: ; NOLOOP: s_mov_b32 m0, 0 ; NOLOOP: ds_gws_barrier v{{[0-9]+}} offset:10 gds ; LOOP: s_mov_b32 m0, -1 ; LOOP: ds_write_b32 ; LOOP: s_mov_b32 m0, 0 ; LOOP: s_setreg_imm32_b32 ; LOOP: ds_gws_barrier v{{[0-9]+}} offset:10 gds ; LOOP: s_cbranch_scc1 ; LOOP: s_mov_b32 m0, -1 ; LOOP: ds_write_b32 define amdgpu_kernel void @gws_barrier_save_m0_barrier_constant_offset(i32 %val) #0 { store i32 1, i32 addrspace(3)* @lds call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 10) store i32 2, i32 addrspace(3)* @lds ret void } ; Make sure this increments lgkmcnt ; GCN-LABEL: {{^}}gws_barrier_lgkmcnt: ; NOLOOP: s_mov_b32 m0, 0{{$}} ; NOLOOP: ds_gws_barrier v0 gds{{$}} ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; NOLOOP-NEXT: s_setpc_b64 define void @gws_barrier_lgkmcnt(i32 %val) { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 0) ret void } ; Does not imply memory fence on its own ; GCN-LABEL: {{^}}gws_barrier_wait_before: ; NOLOOP: s_waitcnt ; NOLOOP-NOT: s_waitcnt{{$}} define amdgpu_kernel void @gws_barrier_wait_before(i32 %val, i32 addrspace(1)* %ptr) #0 { store i32 0, i32 addrspace(1)* %ptr call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) ret void } ; GCN-LABEL: {{^}}gws_barrier_wait_after: ; NOLOOP: s_mov_b32 m0, 0{{$}} ; NOLOOP: ds_gws_barrier v{{[0-9]+}} offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; NOLOOP: load_{{dword|b32}} define amdgpu_kernel void @gws_barrier_wait_after(i32 %val, i32 addrspace(1)* %ptr) #0 { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) %load = load volatile i32, i32 addrspace(1)* %ptr ret void } ; Does not imply memory fence on its own ; GCN-LABEL: {{^}}gws_barrier_fence_before: ; NOLOOP: s_mov_b32 m0, 0{{$}} ; NOLOOP: store_{{dword|b32}} ; NOLOOP: s_waitcnt vmcnt(0) lgkmcnt(0) ; NOLOOP: ds_gws_barrier v{{[0-9]+}} offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) define amdgpu_kernel void @gws_barrier_fence_before(i32 %val, i32 addrspace(1)* %ptr) #0 { store i32 0, i32 addrspace(1)* %ptr fence release call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) ret void } ; FIXME: Extra waitcnt ; GCN-LABEL: {{^}}gws_barrier_fence_after: ; NOLOOP: s_mov_b32 m0, 0{{$}} ; NOLOOP: ds_gws_barrier v{{[0-9]+}} offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 ; NOLOOP-NEXT: load_{{dword|b32}} define amdgpu_kernel void @gws_barrier_fence_after(i32 %val, i32 addrspace(1)* %ptr) #0 { call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) fence release %load = load volatile i32, i32 addrspace(1)* %ptr ret void } ; FIXME: Should a wait be inserted here, or is an explicit fence needed? ; GCN-LABEL: {{^}}gws_init_barrier: ; NOLOOP: s_mov_b32 m0, 0 ; NOLOOP: ds_gws_init v0 offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; NOLOOP-NEXT: ds_gws_barrier v0 offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) define amdgpu_kernel void @gws_init_barrier(i32 %val) #0 { call void @llvm.amdgcn.ds.gws.init(i32 %val, i32 7) call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) ret void } ; FIXME: Why vmcnt, not expcnt? ; GCN-LABEL: {{^}}gws_init_fence_barrier: ; NOLOOP: s_mov_b32 m0, 0 ; NOLOOP: ds_gws_init v0 offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) ; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 ; NOLOOP-NEXT: ds_gws_barrier v0 offset:7 gds ; NOLOOP-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) define amdgpu_kernel void @gws_init_fence_barrier(i32 %val) #0 { call void @llvm.amdgcn.ds.gws.init(i32 %val, i32 7) fence release call void @llvm.amdgcn.ds.gws.barrier(i32 %val, i32 7) ret void } declare void @llvm.amdgcn.ds.gws.barrier(i32, i32) #1 declare void @llvm.amdgcn.ds.gws.init(i32, i32) #2 declare i32 @llvm.amdgcn.workitem.id.x() #3 attributes #0 = { nounwind } attributes #1 = { convergent inaccessiblememonly nounwind } attributes #2 = { convergent inaccessiblememonly nounwind writeonly } attributes #3 = { nounwind readnone speculatable }