diff --git a/llvm/test/CodeGen/AMDGPU/waitcnt-vscnt.ll b/llvm/test/CodeGen/AMDGPU/waitcnt-vscnt.ll --- a/llvm/test/CodeGen/AMDGPU/waitcnt-vscnt.ll +++ b/llvm/test/CodeGen/AMDGPU/waitcnt-vscnt.ll @@ -1,12 +1,13 @@ ; RUN: llc -march=amdgcn -mcpu=gfx802 -asm-verbose=0 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX8,GFX8_9 %s -; RUN: llc -march=amdgcn -mcpu=gfx900 -asm-verbose=0 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX9,GFX9_10,GFX8_9 %s +; RUN: llc -march=amdgcn -mcpu=gfx900 -asm-verbose=0 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX9_10,GFX8_9 %s ; RUN: llc -march=amdgcn -mcpu=gfx1010 -asm-verbose=0 -verify-machineinstrs < %s | FileCheck -check-prefixes=GCN,GFX10,GFX9_10 %s ; GCN-LABEL: barrier_vmcnt_global: ; GFX8: flat_load_dword ; GFX9_10: global_load_dword ; GFX8: s_waitcnt vmcnt(0){{$}} -; GFX9_10: s_waitcnt vmcnt(0){{$}} +; GFX9_10: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} +; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier define amdgpu_kernel void @barrier_vmcnt_global(i32 addrspace(1)* %arg) { bb: @@ -15,9 +16,9 @@ %tmp2 = shl nuw nsw i64 %tmp1, 32 %tmp3 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp1 %tmp4 = load i32, i32 addrspace(1)* %tmp3, align 4 - fence syncscope("singlethread") release + fence syncscope("workgroup") release tail call void @llvm.amdgcn.s.barrier() - fence syncscope("singlethread") acquire + fence syncscope("workgroup") acquire %tmp5 = add nuw nsw i64 %tmp2, 4294967296 %tmp6 = lshr exact i64 %tmp5, 32 %tmp7 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp6 @@ -28,8 +29,7 @@ ; GCN-LABEL: barrier_vscnt_global: ; GFX8: flat_store_dword ; GFX9_10: global_store_dword -; GFX8: s_waitcnt vmcnt(0){{$}} -; GFX9: s_waitcnt vmcnt(0){{$}} +; GCN: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} ; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier define amdgpu_kernel void @barrier_vscnt_global(i32 addrspace(1)* %arg) { @@ -41,9 +41,9 @@ %tmp4 = lshr exact i64 %tmp3, 32 %tmp5 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp4 store i32 0, i32 addrspace(1)* %tmp5, align 4 - fence syncscope("singlethread") release + fence syncscope("workgroup") release tail call void @llvm.amdgcn.s.barrier() #3 - fence syncscope("singlethread") acquire + fence syncscope("workgroup") acquire %tmp6 = add nuw nsw i64 %tmp2, 4294967296 %tmp7 = lshr exact i64 %tmp6, 32 %tmp8 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp7 @@ -54,8 +54,8 @@ ; GCN-LABEL: barrier_vmcnt_vscnt_global: ; GFX8: flat_load_dword ; GFX9_10: global_load_dword -; GFX8: s_waitcnt vmcnt(0){{$}} -; GFX9_10: s_waitcnt vmcnt(0){{$}} +; GFX8_9: s_waitcnt vmcnt(0){{$}} +; GFX10: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} ; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier define amdgpu_kernel void @barrier_vmcnt_vscnt_global(i32 addrspace(1)* %arg) { @@ -69,9 +69,9 @@ store i32 0, i32 addrspace(1)* %tmp5, align 4 %tmp6 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp1 %tmp7 = load i32, i32 addrspace(1)* %tmp6, align 4 - fence syncscope("singlethread") release + fence syncscope("workgroup") release tail call void @llvm.amdgcn.s.barrier() - fence syncscope("singlethread") acquire + fence syncscope("workgroup") acquire %tmp8 = add nuw nsw i64 %tmp2, 4294967296 %tmp9 = lshr exact i64 %tmp8, 32 %tmp10 = getelementptr inbounds i32, i32 addrspace(1)* %arg, i64 %tmp9 @@ -81,7 +81,9 @@ ; GCN-LABEL: barrier_vmcnt_flat: ; GCN: flat_load_dword -; GCN: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} +; GFX8_9: s_waitcnt vmcnt(0){{$}} +; GFX10: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} +; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier define amdgpu_kernel void @barrier_vmcnt_flat(i32* %arg) { bb: @@ -90,9 +92,9 @@ %tmp2 = shl nuw nsw i64 %tmp1, 32 %tmp3 = getelementptr inbounds i32, i32* %arg, i64 %tmp1 %tmp4 = load i32, i32* %tmp3, align 4 - fence syncscope("singlethread") release + fence syncscope("workgroup") release tail call void @llvm.amdgcn.s.barrier() - fence syncscope("singlethread") acquire + fence syncscope("workgroup") acquire %tmp5 = add nuw nsw i64 %tmp2, 4294967296 %tmp6 = lshr exact i64 %tmp5, 32 %tmp7 = getelementptr inbounds i32, i32* %arg, i64 %tmp6 @@ -102,8 +104,7 @@ ; GCN-LABEL: barrier_vscnt_flat: ; GCN: flat_store_dword -; GFX8_9: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} -; GFX10: s_waitcnt lgkmcnt(0){{$}} +; GCN: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} ; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier define amdgpu_kernel void @barrier_vscnt_flat(i32* %arg) { @@ -115,9 +116,9 @@ %tmp4 = lshr exact i64 %tmp3, 32 %tmp5 = getelementptr inbounds i32, i32* %arg, i64 %tmp4 store i32 0, i32* %tmp5, align 4 - fence syncscope("singlethread") release + fence syncscope("workgroup") release tail call void @llvm.amdgcn.s.barrier() #3 - fence syncscope("singlethread") acquire + fence syncscope("workgroup") acquire %tmp6 = add nuw nsw i64 %tmp2, 4294967296 %tmp7 = lshr exact i64 %tmp6, 32 %tmp8 = getelementptr inbounds i32, i32* %arg, i64 %tmp7 @@ -127,38 +128,11 @@ ; GCN-LABEL: barrier_vmcnt_vscnt_flat: ; GCN: flat_load_dword -; GCN: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} -; GFX10: s_waitcnt_vscnt null, 0x0 -; GCN-NEXT: s_barrier -define amdgpu_kernel void @barrier_vmcnt_vscnt_flat(i32* %arg) { -bb: - %tmp = tail call i32 @llvm.amdgcn.workitem.id.x() - %tmp1 = zext i32 %tmp to i64 - %tmp2 = shl nuw nsw i64 %tmp1, 32 - %tmp3 = add nuw nsw i64 %tmp2, 8589934592 - %tmp4 = lshr exact i64 %tmp3, 32 - %tmp5 = getelementptr inbounds i32, i32* %arg, i64 %tmp4 - store i32 0, i32* %tmp5, align 4 - %tmp6 = getelementptr inbounds i32, i32* %arg, i64 %tmp1 - %tmp7 = load i32, i32* %tmp6, align 4 - fence syncscope("singlethread") release - tail call void @llvm.amdgcn.s.barrier() - fence syncscope("singlethread") acquire - %tmp8 = add nuw nsw i64 %tmp2, 4294967296 - %tmp9 = lshr exact i64 %tmp8, 32 - %tmp10 = getelementptr inbounds i32, i32* %arg, i64 %tmp9 - store i32 %tmp7, i32* %tmp10, align 4 - ret void -} - -; GCN-LABEL: barrier_vmcnt_vscnt_flat_workgroup: -; GCN: flat_load_dword -; GFX8_9: s_waitcnt lgkmcnt(0){{$}} ; GFX8_9: s_waitcnt vmcnt(0){{$}} ; GFX10: s_waitcnt vmcnt(0) lgkmcnt(0){{$}} ; GFX10: s_waitcnt_vscnt null, 0x0 ; GCN-NEXT: s_barrier -define amdgpu_kernel void @barrier_vmcnt_vscnt_flat_workgroup(i32* %arg) { +define amdgpu_kernel void @barrier_vmcnt_vscnt_flat(i32* %arg) { bb: %tmp = tail call i32 @llvm.amdgcn.workitem.id.x() %tmp1 = zext i32 %tmp to i64 @@ -182,8 +156,7 @@ ; GCN-LABEL: load_vmcnt_global: ; GFX8: flat_load_dword ; GFX9_10: global_load_dword -; GFX8: s_waitcnt vmcnt(0){{$}} -; GFX9_10: s_waitcnt vmcnt(0){{$}} +; GCN: s_waitcnt vmcnt(0){{$}} ; GCN-NEXT: {{global|flat}}_store_dword define amdgpu_kernel void @load_vmcnt_global(i32 addrspace(1)* %arg) { bb: