; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py UTC_ARGS: --version 3 ; RUN: llc -global-isel=0 -mtriple=amdgcn -mcpu=gfx1250 < %s | FileCheck --check-prefixes=GCN,GCN-SDAG %s ; RUN: llc -global-isel=1 -mtriple=amdgcn -mcpu=gfx1250 < %s | FileCheck --check-prefixes=GCN,GCN-GISEL %s define amdgpu_ps i64 @s_add_u64(i64 inreg %a) { ; GCN-LABEL: s_add_u64: ; GCN: ; %bb.0: ; GCN-NEXT: s_add_nc_u64 s[0:1], s[0:1], lit64(0xf12345678) ; GCN-NEXT: ; return to shader part epilog %result = add i64 %a, 64729929336 ret i64 %result } define amdgpu_ps void @v_add_u64(i64 %a, ptr addrspace(1) %out) { ; GCN-SDAG-LABEL: v_add_u64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: s_mov_b64 s[0:1], lit64(0xf12345678) ; GCN-SDAG-NEXT: s_delay_alu instid0(SALU_CYCLE_1) ; GCN-SDAG-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 0, s[0:1] ; GCN-SDAG-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-SDAG-NEXT: s_endpgm ; ; GCN-GISEL-LABEL: v_add_u64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_mov_b64_e32 v[4:5], lit64(0xf12345678) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-GISEL-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 0, v[4:5] ; GCN-GISEL-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-GISEL-NEXT: s_endpgm %result = add i64 %a, 64729929336 store i64 %result, ptr addrspace(1) %out, align 8 ret void } define amdgpu_ps i64 @s_add_neg_u64(i64 inreg %a) { ; GCN-LABEL: s_add_neg_u64: ; GCN: ; %bb.0: ; GCN-NEXT: s_add_nc_u64 s[0:1], s[0:1], lit64(0xfffffff0edcba988) ; GCN-NEXT: ; return to shader part epilog %result = sub i64 %a, 64729929336 ret i64 %result } define amdgpu_ps void @v_add_neg_u64(i64 %a, ptr addrspace(1) %out) { ; GCN-SDAG-LABEL: v_add_neg_u64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: s_mov_b64 s[0:1], lit64(0xfffffff0edcba988) ; GCN-SDAG-NEXT: s_delay_alu instid0(SALU_CYCLE_1) ; GCN-SDAG-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 0, s[0:1] ; GCN-SDAG-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-SDAG-NEXT: s_endpgm ; ; GCN-GISEL-LABEL: v_add_neg_u64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_mov_b64_e32 v[4:5], lit64(0xfffffff0edcba988) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-GISEL-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 0, v[4:5] ; GCN-GISEL-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-GISEL-NEXT: s_endpgm %result = sub i64 %a, 64729929336 store i64 %result, ptr addrspace(1) %out, align 8 ret void } define amdgpu_ps i64 @s_sub_u64(i64 inreg %a) { ; GCN-LABEL: s_sub_u64: ; GCN: ; %bb.0: ; GCN-NEXT: s_sub_nc_u64 s[0:1], lit64(0xf12345678), s[0:1] ; GCN-NEXT: ; return to shader part epilog %result = sub i64 64729929336, %a ret i64 %result } define amdgpu_ps void @v_sub_u64(i64 %a, ptr addrspace(1) %out) { ; GCN-LABEL: v_sub_u64: ; GCN: ; %bb.0: ; GCN-NEXT: v_sub_co_u32 v0, vcc_lo, 0x12345678, v0 ; GCN-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-NEXT: v_sub_co_ci_u32_e64 v1, null, 15, v1, vcc_lo ; GCN-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-NEXT: s_endpgm %result = sub i64 64729929336, %a store i64 %result, ptr addrspace(1) %out, align 8 ret void } define void @v_mov_b64_double(ptr addrspace(1) %ptr) { ; GCN-LABEL: v_mov_b64_double: ; GCN: ; %bb.0: ; GCN-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-NEXT: s_wait_kmcnt 0x0 ; GCN-NEXT: global_load_b64 v[4:5], v[0:1], off ; GCN-NEXT: s_mov_b32 s0, 0 ; GCN-NEXT: .LBB6_1: ; %atomicrmw.start ; GCN-NEXT: ; =>This Inner Loop Header: Depth=1 ; GCN-NEXT: s_wait_loadcnt 0x0 ; GCN-NEXT: v_add_f64_e32 v[2:3], lit64(0x4063233333333333), v[4:5] ; GCN-NEXT: global_atomic_cmpswap_b64 v[2:3], v[0:1], v[2:5], off th:TH_ATOMIC_RETURN scope:SCOPE_SYS ; GCN-NEXT: s_wait_loadcnt 0x0 ; GCN-NEXT: v_cmp_eq_u64_e32 vcc_lo, v[2:3], v[4:5] ; GCN-NEXT: s_wait_xcnt 0x0 ; GCN-NEXT: v_mov_b64_e32 v[4:5], v[2:3] ; GCN-NEXT: s_wait_alu 0xfffe ; GCN-NEXT: s_or_b32 s0, vcc_lo, s0 ; GCN-NEXT: s_wait_alu 0xfffe ; GCN-NEXT: s_and_not1_b32 exec_lo, exec_lo, s0 ; GCN-NEXT: s_cbranch_execnz .LBB6_1 ; GCN-NEXT: ; %bb.2: ; %atomicrmw.end ; GCN-NEXT: s_or_b32 exec_lo, exec_lo, s0 ; GCN-NEXT: s_set_pc_i64 s[30:31] %result = atomicrmw fadd ptr addrspace(1) %ptr, double 153.1 monotonic ret void } define void @v_mov_b64_int(ptr addrspace(1) %ptr) { ; GCN-LABEL: v_mov_b64_int: ; GCN: ; %bb.0: ; GCN-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-NEXT: s_wait_kmcnt 0x0 ; GCN-NEXT: v_mov_b64_e32 v[2:3], lit64(0xf12345678) ; GCN-NEXT: global_atomic_add_u64 v[0:1], v[2:3], off scope:SCOPE_SYS ; GCN-NEXT: s_set_pc_i64 s[30:31] %result = atomicrmw add ptr addrspace(1) %ptr, i64 64729929336 monotonic ret void } define void @store_double(ptr addrspace(1) %ptr) { ; GCN-LABEL: store_double: ; GCN: ; %bb.0: ; GCN-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-NEXT: s_wait_kmcnt 0x0 ; GCN-NEXT: v_mov_b64_e32 v[2:3], lit64(0x4063233333333333) ; GCN-NEXT: global_store_b64 v[0:1], v[2:3], off ; GCN-NEXT: s_set_pc_i64 s[30:31] store double 153.1, ptr addrspace(1) %ptr ret void } define i1 @class_f64() noinline optnone { ; GCN-SDAG-LABEL: class_f64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-SDAG-NEXT: s_wait_kmcnt 0x0 ; GCN-SDAG-NEXT: s_mov_b32 s2, 1 ; GCN-SDAG-NEXT: s_mov_b64 s[0:1], lit64(0x4063233333333333) ; GCN-SDAG-NEXT: s_wait_alu 0xfffe ; GCN-SDAG-NEXT: v_cmp_class_f64_e64 s0, s[0:1], s2 ; GCN-SDAG-NEXT: s_wait_alu 0xf1ff ; GCN-SDAG-NEXT: v_cndmask_b32_e64 v0, 0, 1, s0 ; GCN-SDAG-NEXT: s_set_pc_i64 s[30:31] ; ; GCN-GISEL-LABEL: class_f64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-GISEL-NEXT: s_wait_kmcnt 0x0 ; GCN-GISEL-NEXT: s_mov_b32 s2, 1 ; GCN-GISEL-NEXT: s_mov_b64 s[0:1], lit64(0x4063233333333333) ; GCN-GISEL-NEXT: s_wait_alu 0xfffe ; GCN-GISEL-NEXT: v_mov_b64_e32 v[0:1], s[0:1] ; GCN-GISEL-NEXT: v_mov_b32_e32 v2, s2 ; GCN-GISEL-NEXT: v_cmp_class_f64_e64 s0, v[0:1], v2 ; GCN-GISEL-NEXT: v_mov_b32_e32 v0, 1 ; GCN-GISEL-NEXT: v_mov_b32_e32 v1, 0 ; GCN-GISEL-NEXT: s_wait_alu 0xf1ff ; GCN-GISEL-NEXT: v_cndmask_b32_e64 v0, v1, v0, s0 ; GCN-GISEL-NEXT: s_set_pc_i64 s[30:31] %result = call i1 @llvm.amdgcn.class.f64(double 153.1, i32 1) nounwind readnone ret i1 %result } define double @rsq_f64() { ; GCN-LABEL: rsq_f64: ; GCN: ; %bb.0: ; GCN-NEXT: s_wait_loadcnt_dscnt 0x0 ; GCN-NEXT: s_wait_kmcnt 0x0 ; GCN-NEXT: v_rsq_f64_e32 v[0:1], lit64(0x4063233333333333) ; GCN-NEXT: s_set_pc_i64 s[30:31] %result = call double @llvm.amdgcn.rsq.f64(double 153.1) nounwind readnone ret double %result } define amdgpu_ps i64 @s_and_b64(i64 inreg %a) { ; GCN-LABEL: s_and_b64: ; GCN: ; %bb.0: ; GCN-NEXT: s_and_b64 s[0:1], s[0:1], lit64(0xf12345678) ; GCN-NEXT: ; return to shader part epilog %result = and i64 %a, 64729929336 ret i64 %result } ; No V_AND_B64 instruction, it has to be split define amdgpu_ps void @v_and_b64(i64 %a, ptr addrspace(1) %out) { ; GCN-SDAG-LABEL: v_and_b64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: v_and_b32_e32 v1, 15, v1 ; GCN-SDAG-NEXT: v_and_b32_e32 v0, 0x12345678, v0 ; GCN-SDAG-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-SDAG-NEXT: s_endpgm ; ; GCN-GISEL-LABEL: v_and_b64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_and_b32_e32 v0, 0x12345678, v0 ; GCN-GISEL-NEXT: v_and_b32_e32 v1, 15, v1 ; GCN-GISEL-NEXT: global_store_b64 v[2:3], v[0:1], off ; GCN-GISEL-NEXT: s_endpgm %result = and i64 %a, 64729929336 store i64 %result, ptr addrspace(1) %out, align 8 ret void } define amdgpu_ps <2 x float> @v_add_f64_200.1(double %a) { ; GCN-LABEL: v_add_f64_200.1: ; GCN: ; %bb.0: ; GCN-NEXT: v_add_f64_e32 v[0:1], lit64(0x4069033333333333), v[0:1] ; GCN-NEXT: ; return to shader part epilog %add = fadd double %a, 200.1 %ret = bitcast double %add to <2 x float> ret <2 x float> %ret } ; 200.0 can be encoded as 32-bit literal define amdgpu_ps <2 x float> @v_add_f64_200.0(double %a) { ; GCN-LABEL: v_add_f64_200.0: ; GCN: ; %bb.0: ; GCN-NEXT: v_add_f64_e32 v[0:1], 0x40690000, v[0:1] ; GCN-NEXT: ; return to shader part epilog %add = fadd double %a, 200.0 %ret = bitcast double %add to <2 x float> ret <2 x float> %ret } ; No folding into VOP3 define amdgpu_ps <2 x float> @v_lshl_add_u64(i64 %a) { ; GCN-SDAG-LABEL: v_lshl_add_u64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: s_mov_b64 s[0:1], lit64(0xf12345678) ; GCN-SDAG-NEXT: s_delay_alu instid0(SALU_CYCLE_1) ; GCN-SDAG-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 1, s[0:1] ; GCN-SDAG-NEXT: ; return to shader part epilog ; ; GCN-GISEL-LABEL: v_lshl_add_u64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_mov_b64_e32 v[2:3], lit64(0xf12345678) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-GISEL-NEXT: v_lshl_add_u64 v[0:1], v[0:1], 1, v[2:3] ; GCN-GISEL-NEXT: ; return to shader part epilog %shl = shl i64 %a, 1 %add = add i64 %shl, 64729929336 %ret = bitcast i64 %add to <2 x float> ret <2 x float> %ret } ; No folding into VOP2 promoted to VOP3 define amdgpu_ps <2 x float> @v_fma_f64(double %a, double %b) { ; GCN-SDAG-LABEL: v_fma_f64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: v_fmaak_f64 v[4:5], v[0:1], v[2:3], lit64(0x4063233333333333) ; GCN-SDAG-NEXT: v_mov_b64_e32 v[2:3], lit64(0x4069033333333333) ; GCN-SDAG-NEXT: s_delay_alu instid0(VALU_DEP_2) | instskip(NEXT) | instid1(VALU_DEP_1) ; GCN-SDAG-NEXT: v_fmaak_f64 v[0:1], v[0:1], v[4:5], lit64(0x4069033333333333) ; GCN-SDAG-NEXT: v_fmac_f64_e32 v[2:3], v[0:1], v[4:5] ; GCN-SDAG-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-SDAG-NEXT: v_dual_mov_b32 v0, v2 :: v_dual_mov_b32 v1, v3 ; GCN-SDAG-NEXT: ; return to shader part epilog ; ; GCN-GISEL-LABEL: v_fma_f64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_mov_b64_e32 v[4:5], lit64(0x4063233333333333) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) | instskip(SKIP_1) | instid1(VALU_DEP_2) ; GCN-GISEL-NEXT: v_fmac_f64_e32 v[4:5], v[0:1], v[2:3] ; GCN-GISEL-NEXT: v_mov_b64_e32 v[2:3], lit64(0x4069033333333333) ; GCN-GISEL-NEXT: v_fmaak_f64 v[0:1], v[0:1], v[4:5], lit64(0x4069033333333333) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) | instskip(NEXT) | instid1(VALU_DEP_1) ; GCN-GISEL-NEXT: v_fmac_f64_e32 v[2:3], v[0:1], v[4:5] ; GCN-GISEL-NEXT: v_dual_mov_b32 v0, v2 :: v_dual_mov_b32 v1, v3 ; GCN-GISEL-NEXT: ; return to shader part epilog %r1 = call double @llvm.fma.f64(double %a, double %b, double 153.1) nounwind readnone %r2 = call double @llvm.fma.f64(double %a, double %r1, double 200.1) nounwind readnone %r3 = call double @llvm.fma.f64(double %r2, double %r1, double 200.1) nounwind readnone %ret = bitcast double %r3 to <2 x float> ret <2 x float> %ret } define amdgpu_ps <2 x float> @v_add_neg_f64(double %a) { ; GCN-SDAG-LABEL: v_add_neg_f64: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: s_mov_b64 s[0:1], lit64(0x4069033333333333) ; GCN-SDAG-NEXT: s_delay_alu instid0(SALU_CYCLE_1) ; GCN-SDAG-NEXT: v_add_f64_e64 v[0:1], -v[0:1], s[0:1] ; GCN-SDAG-NEXT: ; return to shader part epilog ; ; GCN-GISEL-LABEL: v_add_neg_f64: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_max_num_f64_e32 v[0:1], v[0:1], v[0:1] ; GCN-GISEL-NEXT: v_mov_b64_e32 v[2:3], lit64(0x4069033333333333) ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_1) ; GCN-GISEL-NEXT: v_add_f64_e64 v[0:1], -v[0:1], v[2:3] ; GCN-GISEL-NEXT: ; return to shader part epilog %fneg = fsub double -0.0, %a %add = fadd double %fneg, 200.1 %ret = bitcast double %add to <2 x float> ret <2 x float> %ret } define amdgpu_ps <2 x float> @v_cndmask(double %a) { ; GCN-SDAG-LABEL: v_cndmask: ; GCN-SDAG: ; %bb.0: ; GCN-SDAG-NEXT: v_cmp_eq_f64_e32 vcc_lo, 0, v[0:1] ; GCN-SDAG-NEXT: v_mov_b32_e32 v1, 0x40632000 ; GCN-SDAG-NEXT: v_cndmask_b32_e64 v0, 0x33333333, 0, vcc_lo ; GCN-SDAG-NEXT: s_delay_alu instid0(VALU_DEP_2) ; GCN-SDAG-NEXT: v_cndmask_b32_e32 v1, 0x40690333, v1, vcc_lo ; GCN-SDAG-NEXT: ; return to shader part epilog ; ; GCN-GISEL-LABEL: v_cndmask: ; GCN-GISEL: ; %bb.0: ; GCN-GISEL-NEXT: v_cmp_eq_f64_e32 vcc_lo, 0, v[0:1] ; GCN-GISEL-NEXT: v_mov_b32_e32 v1, 0x40690333 ; GCN-GISEL-NEXT: v_cndmask_b32_e64 v0, 0x33333333, 0, vcc_lo ; GCN-GISEL-NEXT: s_delay_alu instid0(VALU_DEP_2) ; GCN-GISEL-NEXT: v_cndmask_b32_e64 v1, v1, 0x40632000, vcc_lo ; GCN-GISEL-NEXT: ; return to shader part epilog %cmp = fcmp oeq double %a, 0.0 %sel = select i1 %cmp, double 153.0, double 200.1 %ret = bitcast double %sel to <2 x float> ret <2 x float> %ret } declare i1 @llvm.amdgcn.class.f64(double, i32) nounwind readnone declare double @llvm.amdgcn.rsq.f64(double) nounwind readnone declare double @llvm.fma.f64(double, double, double) nounwind readnone