; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 5 ; RUN: llc -mtriple=amdgcn-amd-amdhsa -mcpu=gfx1100 -o - %s | FileCheck %s define amdgpu_kernel void @readfirstlane_with_readfirstlane(ptr addrspace(1) %out) { ; CHECK-LABEL: readfirstlane_with_readfirstlane: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, 5 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %v1 = call i32 @llvm.amdgcn.readfirstlane(i32 5) %v2 = call i32 @llvm.amdgcn.readfirstlane(i32 %v1) store i32 %v2, ptr addrspace(1) %out ret void } define amdgpu_kernel void @readfirstlane_with_readlane(ptr addrspace(1) %out) { ; CHECK-LABEL: readfirstlane_with_readlane: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_bfe_u32 v1, v0, 10, 10 ; CHECK-NEXT: v_and_b32_e32 v0, 0x3ff, v0 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_2) | instskip(NEXT) | instid1(VALU_DEP_1) ; CHECK-NEXT: v_readfirstlane_b32 s2, v1 ; CHECK-NEXT: v_readlane_b32 s2, v0, s2 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_1) ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, s2 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %tidx = call i32 @llvm.amdgcn.workitem.id.x() %tidy = call i32 @llvm.amdgcn.workitem.id.y() %v1 = call i32 @llvm.amdgcn.readlane(i32 %tidx, i32 %tidy) %v2 = call i32 @llvm.amdgcn.readfirstlane(i32 %v1) store i32 %v2, ptr addrspace(1) %out ret void } define amdgpu_kernel void @readlane_with_firstlane(ptr addrspace(1) %out) { ; CHECK-LABEL: readlane_with_firstlane: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_and_b32_e32 v0, 0x3ff, v0 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_1) | instskip(NEXT) | instid1(VALU_DEP_1) ; CHECK-NEXT: v_readfirstlane_b32 s2, v0 ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, s2 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %tidx = call i32 @llvm.amdgcn.workitem.id.x() %v1 = call i32 @llvm.amdgcn.readfirstlane(i32 %tidx) %v2 = call i32 @llvm.amdgcn.readlane(i32 %v1, i32 3) store i32 %v2, ptr addrspace(1) %out ret void } define amdgpu_kernel void @readlane_readlane(ptr addrspace(1) %out) { ; CHECK-LABEL: readlane_readlane: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_bfe_u32 v1, v0, 10, 10 ; CHECK-NEXT: v_and_b32_e32 v0, 0x3ff, v0 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_2) | instskip(NEXT) | instid1(VALU_DEP_1) ; CHECK-NEXT: v_readfirstlane_b32 s2, v1 ; CHECK-NEXT: v_readlane_b32 s2, v0, s2 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_1) ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, s2 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %tidx = call i32 @llvm.amdgcn.workitem.id.x() %tidy = call i32 @llvm.amdgcn.workitem.id.y() %v1 = call i32 @llvm.amdgcn.readlane(i32 %tidx, i32 %tidy) %v2 = call i32 @llvm.amdgcn.readlane(i32 %v1, i32 2) store i32 %v2, ptr addrspace(1) %out ret void } define amdgpu_kernel void @permlane64_uniform(ptr addrspace(1) %out, i32 %src) { ; CHECK-LABEL: permlane64_uniform: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_clause 0x1 ; CHECK-NEXT: s_load_b32 s2, s[4:5], 0x8 ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, s2 ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %v = call i32 @llvm.amdgcn.permlane64(i32 %src) store i32 %v, ptr addrspace(1) %out ret void } define amdgpu_kernel void @permlane64_nonuniform(i32 addrspace(1)* %out) { ; CHECK-LABEL: permlane64_nonuniform: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_and_b32_e32 v0, 0x3ff, v0 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_1) ; CHECK-NEXT: v_permlane64_b32 v1, v0 ; CHECK-NEXT: v_lshlrev_b32_e32 v0, 2, v0 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %tid = call i32 @llvm.amdgcn.workitem.id.x() %v = call i32 @llvm.amdgcn.permlane64(i32 %tid) %out_ptr = getelementptr i32, i32 addrspace(1)* %out, i32 %tid store i32 %v, i32 addrspace(1)* %out_ptr ret void } define amdgpu_kernel void @permlane64_nonuniform_expression(i32 addrspace(1)* %out) { ; CHECK-LABEL: permlane64_nonuniform_expression: ; CHECK: ; %bb.0: ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_and_b32_e32 v0, 0x3ff, v0 ; CHECK-NEXT: s_delay_alu instid0(VALU_DEP_1) | instskip(SKIP_1) | instid1(VALU_DEP_2) ; CHECK-NEXT: v_add_nc_u32_e32 v1, 1, v0 ; CHECK-NEXT: v_lshlrev_b32_e32 v0, 2, v0 ; CHECK-NEXT: v_permlane64_b32 v1, v1 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_endpgm %tid = call i32 @llvm.amdgcn.workitem.id.x() %tid2 = add i32 %tid, 1 %v = call i32 @llvm.amdgcn.permlane64(i32 %tid2) %out_ptr = getelementptr i32, i32 addrspace(1)* %out, i32 %tid store i32 %v, i32 addrspace(1)* %out_ptr ret void } define protected amdgpu_kernel void @trivial_waterfall_eq_zero(ptr addrspace(1) %out) { ; CHECK-LABEL: trivial_waterfall_eq_zero: ; CHECK: ; %bb.0: ; %entry ; CHECK-NEXT: s_load_b64 s[0:1], s[4:5], 0x0 ; CHECK-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_mov_b32 v1, 5 ; CHECK-NEXT: s_mov_b32 s2, 0 ; CHECK-NEXT: s_branch .LBB7_2 ; CHECK-NEXT: .LBB7_1: ; %Flow ; CHECK-NEXT: ; in Loop: Header=BB7_2 Depth=1 ; CHECK-NEXT: s_and_not1_b32 vcc_lo, exec_lo, s2 ; CHECK-NEXT: s_mov_b32 s2, -1 ; CHECK-NEXT: s_cbranch_vccz .LBB7_4 ; CHECK-NEXT: .LBB7_2: ; %while ; CHECK-NEXT: ; =>This Inner Loop Header: Depth=1 ; CHECK-NEXT: s_and_b32 vcc_lo, exec_lo, s2 ; CHECK-NEXT: s_mov_b32 s2, -1 ; CHECK-NEXT: s_cbranch_vccnz .LBB7_1 ; CHECK-NEXT: ; %bb.3: ; %if ; CHECK-NEXT: ; in Loop: Header=BB7_2 Depth=1 ; CHECK-NEXT: s_mov_b32 s2, 0 ; CHECK-NEXT: s_waitcnt lgkmcnt(0) ; CHECK-NEXT: global_store_b32 v0, v1, s[0:1] ; CHECK-NEXT: s_branch .LBB7_1 ; CHECK-NEXT: .LBB7_4: ; %exit ; CHECK-NEXT: s_endpgm entry: br label %while while: %done = phi i1 [ 0, %entry ], [ 1, %if ] %not_done = xor i1 %done, true %ballot = tail call i64 @llvm.amdgcn.ballot.i64(i1 %not_done) %is_done = icmp eq i64 %ballot, 0 ; in this case is_done = !not_done br i1 %is_done, label %exit, label %if if: store i32 5, ptr addrspace(1) %out br label %while exit: ret void }