diff --git a/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp b/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp --- a/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp +++ b/llvm/lib/Target/AMDGPU/AMDGPUISelLowering.cpp @@ -304,6 +304,8 @@ setOperationAction(ISD::FNEARBYINT, {MVT::f16, MVT::f32, MVT::f64}, Custom); + setOperationAction(ISD::FROUNDEVEN, {MVT::f16, MVT::f32, MVT::f64}, Custom); + setOperationAction(ISD::FREM, {MVT::f16, MVT::f32, MVT::f64}, Custom); // Expand to fneg + fadd. @@ -1189,6 +1191,7 @@ case ISD::FTRUNC: return LowerFTRUNC(Op, DAG); case ISD::FRINT: return LowerFRINT(Op, DAG); case ISD::FNEARBYINT: return LowerFNEARBYINT(Op, DAG); + case ISD::FROUNDEVEN: return LowerFROUND(Op, DAG); case ISD::FROUND: return LowerFROUND(Op, DAG); case ISD::FFLOOR: return LowerFFLOOR(Op, DAG); case ISD::FLOG: diff --git a/llvm/test/CodeGen/AMDGPU/roundeven.ll b/llvm/test/CodeGen/AMDGPU/roundeven.ll new file mode 100644 --- /dev/null +++ b/llvm/test/CodeGen/AMDGPU/roundeven.ll @@ -0,0 +1,1217 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -march=amdgcn -mcpu=tahiti < %s | FileCheck -check-prefix=GFX6 %s +; RUN: llc -march=amdgcn -mcpu=hawaii < %s | FileCheck -check-prefix=GFX7 %s +; RUN: llc -march=amdgcn -mcpu=fiji < %s | FileCheck -check-prefix=GFX8 %s +; RUN: llc -march=amdgcn -mcpu=gfx900 < %s | FileCheck -check-prefix=GFX9 %s +; RUN: llc -march=amdgcn -mcpu=gfx1010 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX10 %s +; RUN: llc -march=amdgcn -mcpu=gfx1100 -amdgpu-enable-delay-alu=0 < %s | FileCheck -check-prefixes=GFX10PLUS,GFX11 %s + +define half @roundeven_f16(half %x) { +; GFX6-LABEL: roundeven_f16: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX6-NEXT: v_trunc_f32_e32 v2, v0 +; GFX6-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_f16: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX7-NEXT: v_trunc_f32_e32 v2, v0 +; GFX7-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_f16: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: s_movk_i32 s4, 0x7fff +; GFX8-NEXT: v_mov_b32_e32 v1, 0x3c00 +; GFX8-NEXT: v_trunc_f16_e32 v2, v0 +; GFX8-NEXT: v_bfi_b32 v1, s4, v1, v0 +; GFX8-NEXT: v_sub_f16_e32 v0, v0, v2 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX8-NEXT: v_add_f16_e32 v0, v2, v0 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: s_movk_i32 s4, 0x7fff +; GFX9-NEXT: v_mov_b32_e32 v1, 0x3c00 +; GFX9-NEXT: v_trunc_f16_e32 v2, v0 +; GFX9-NEXT: v_bfi_b32 v1, s4, v1, v0 +; GFX9-NEXT: v_sub_f16_e32 v0, v0, v2 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v0|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX9-NEXT: v_add_f16_e32 v0, v2, v0 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_f16: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f16_e32 v1, v0 +; GFX10-NEXT: s_movk_i32 s4, 0x3c00 +; GFX10-NEXT: v_sub_f16_e32 v2, v0, v1 +; GFX10-NEXT: v_bfi_b32 v0, 0x7fff, s4, v0 +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v2|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX10-NEXT: v_add_f16_e32 v0, v1, v0 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_f16: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f16_e32 v1, v0 +; GFX11-NEXT: s_movk_i32 s0, 0x3c00 +; GFX11-NEXT: v_sub_f16_e32 v2, v0, v1 +; GFX11-NEXT: v_bfi_b32 v0, 0x7fff, s0, v0 +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v2|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX11-NEXT: v_add_f16_e32 v0, v1, v0 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call half @llvm.roundeven.f16(half %x) + ret half %roundeven +} + +define <2 x half> @roundeven_v2f16(<2 x half> %x) { +; GFX6-LABEL: roundeven_v2f16: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1 +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1 +; GFX6-NEXT: v_trunc_f32_e32 v3, v0 +; GFX6-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX6-NEXT: v_trunc_f32_e32 v3, v1 +; GFX6-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX6-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX6-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v2f16: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1 +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1 +; GFX7-NEXT: v_trunc_f32_e32 v3, v0 +; GFX7-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX7-NEXT: v_trunc_f32_e32 v3, v1 +; GFX7-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX7-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX7-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v2f16: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: v_lshrrev_b32_e32 v1, 16, v0 +; GFX8-NEXT: v_trunc_f16_e32 v2, v1 +; GFX8-NEXT: v_sub_f16_e32 v3, v1, v2 +; GFX8-NEXT: s_movk_i32 s4, 0x7fff +; GFX8-NEXT: v_mov_b32_e32 v4, 0x3c00 +; GFX8-NEXT: v_bfi_b32 v1, s4, v4, v1 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v3|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX8-NEXT: v_trunc_f16_e32 v3, v0 +; GFX8-NEXT: v_add_f16_sdwa v1, v2, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD src1_sel:DWORD +; GFX8-NEXT: v_bfi_b32 v2, s4, v4, v0 +; GFX8-NEXT: v_sub_f16_e32 v0, v0, v3 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX8-NEXT: v_add_f16_e32 v0, v3, v0 +; GFX8-NEXT: v_or_b32_e32 v0, v0, v1 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v2f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_trunc_f16_e32 v3, v0 +; GFX9-NEXT: s_movk_i32 s4, 0x7fff +; GFX9-NEXT: v_mov_b32_e32 v1, 0x3c00 +; GFX9-NEXT: v_sub_f16_e32 v4, v0, v3 +; GFX9-NEXT: v_bfi_b32 v2, s4, v1, v0 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v4|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v2, 0, v2, vcc +; GFX9-NEXT: v_lshrrev_b32_e32 v0, 16, v0 +; GFX9-NEXT: v_add_f16_e32 v2, v3, v2 +; GFX9-NEXT: v_trunc_f16_e32 v3, v0 +; GFX9-NEXT: v_sub_f16_e32 v4, v0, v3 +; GFX9-NEXT: v_bfi_b32 v0, s4, v1, v0 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v4|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc +; GFX9-NEXT: v_add_f16_e32 v0, v3, v0 +; GFX9-NEXT: v_pack_b32_f16 v0, v2, v0 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v2f16: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_lshrrev_b32_e32 v1, 16, v0 +; GFX10-NEXT: v_trunc_f16_e32 v2, v0 +; GFX10-NEXT: s_movk_i32 s4, 0x3c00 +; GFX10-NEXT: v_trunc_f16_e32 v3, v1 +; GFX10-NEXT: v_sub_f16_e32 v4, v0, v2 +; GFX10-NEXT: v_bfi_b32 v0, 0x7fff, s4, v0 +; GFX10-NEXT: v_sub_f16_e32 v5, v1, v3 +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v4|, 0.5 +; GFX10-NEXT: v_bfi_b32 v1, 0x7fff, s4, v1 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v5|, 0.5 +; GFX10-NEXT: v_add_f16_e32 v0, v2, v0 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc_lo +; GFX10-NEXT: v_add_f16_e32 v1, v3, v1 +; GFX10-NEXT: v_pack_b32_f16 v0, v0, v1 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v2f16: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_lshrrev_b32_e32 v1, 16, v0 +; GFX11-NEXT: v_trunc_f16_e32 v2, v0 +; GFX11-NEXT: s_movk_i32 s0, 0x3c00 +; GFX11-NEXT: v_trunc_f16_e32 v3, v1 +; GFX11-NEXT: v_sub_f16_e32 v4, v0, v2 +; GFX11-NEXT: v_bfi_b32 v0, 0x7fff, s0, v0 +; GFX11-NEXT: v_sub_f16_e32 v5, v1, v3 +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v4|, 0.5 +; GFX11-NEXT: v_bfi_b32 v1, 0x7fff, s0, v1 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v5|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc_lo +; GFX11-NEXT: v_add_f16_e32 v0, v2, v0 +; GFX11-NEXT: v_add_f16_e32 v1, v3, v1 +; GFX11-NEXT: v_pack_b32_f16 v0, v0, v1 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <2 x half> @llvm.roundeven.v2f16(<2 x half> %x) + ret <2 x half> %roundeven +} + +define <4 x half> @roundeven_v4f16(<4 x half> %x) { +; GFX6-LABEL: roundeven_v4f16: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX6-NEXT: v_cvt_f16_f32_e32 v1, v1 +; GFX6-NEXT: v_cvt_f16_f32_e32 v2, v2 +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX6-NEXT: v_cvt_f32_f16_e32 v1, v1 +; GFX6-NEXT: v_cvt_f16_f32_e32 v3, v3 +; GFX6-NEXT: v_cvt_f32_f16_e32 v2, v2 +; GFX6-NEXT: v_trunc_f32_e32 v5, v0 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX6-NEXT: v_trunc_f32_e32 v5, v1 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX6-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX6-NEXT: v_cvt_f32_f16_e32 v3, v3 +; GFX6-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX6-NEXT: v_trunc_f32_e32 v5, v2 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX6-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX6-NEXT: v_trunc_f32_e32 v5, v3 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX6-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v4f16: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: v_cvt_f16_f32_e32 v0, v0 +; GFX7-NEXT: v_cvt_f16_f32_e32 v1, v1 +; GFX7-NEXT: v_cvt_f16_f32_e32 v2, v2 +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_cvt_f32_f16_e32 v0, v0 +; GFX7-NEXT: v_cvt_f32_f16_e32 v1, v1 +; GFX7-NEXT: v_cvt_f16_f32_e32 v3, v3 +; GFX7-NEXT: v_cvt_f32_f16_e32 v2, v2 +; GFX7-NEXT: v_trunc_f32_e32 v5, v0 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX7-NEXT: v_trunc_f32_e32 v5, v1 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX7-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX7-NEXT: v_cvt_f32_f16_e32 v3, v3 +; GFX7-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX7-NEXT: v_trunc_f32_e32 v5, v2 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX7-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX7-NEXT: v_trunc_f32_e32 v5, v3 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX7-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v4f16: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: v_trunc_f16_e32 v4, v1 +; GFX8-NEXT: s_movk_i32 s4, 0x7fff +; GFX8-NEXT: v_mov_b32_e32 v2, 0x3c00 +; GFX8-NEXT: v_sub_f16_e32 v5, v1, v4 +; GFX8-NEXT: v_bfi_b32 v3, s4, v2, v1 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v5|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v3, 0, v3, vcc +; GFX8-NEXT: v_lshrrev_b32_e32 v1, 16, v1 +; GFX8-NEXT: v_add_f16_e32 v3, v4, v3 +; GFX8-NEXT: v_trunc_f16_e32 v4, v1 +; GFX8-NEXT: v_sub_f16_e32 v5, v1, v4 +; GFX8-NEXT: v_bfi_b32 v1, s4, v2, v1 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v5|, 0.5 +; GFX8-NEXT: v_trunc_f16_e32 v5, v0 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX8-NEXT: v_sub_f16_e32 v6, v0, v5 +; GFX8-NEXT: v_add_f16_sdwa v1, v4, v1 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD src1_sel:DWORD +; GFX8-NEXT: v_bfi_b32 v4, s4, v2, v0 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v6|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v4, 0, v4, vcc +; GFX8-NEXT: v_lshrrev_b32_e32 v0, 16, v0 +; GFX8-NEXT: v_add_f16_e32 v4, v5, v4 +; GFX8-NEXT: v_trunc_f16_e32 v5, v0 +; GFX8-NEXT: v_sub_f16_e32 v6, v0, v5 +; GFX8-NEXT: v_bfi_b32 v0, s4, v2, v0 +; GFX8-NEXT: v_cmp_ge_f16_e64 vcc, |v6|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc +; GFX8-NEXT: v_add_f16_sdwa v0, v5, v0 dst_sel:WORD_1 dst_unused:UNUSED_PAD src0_sel:DWORD src1_sel:DWORD +; GFX8-NEXT: v_or_b32_e32 v0, v4, v0 +; GFX8-NEXT: v_or_b32_e32 v1, v3, v1 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v4f16: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_trunc_f16_e32 v4, v1 +; GFX9-NEXT: s_movk_i32 s4, 0x7fff +; GFX9-NEXT: v_mov_b32_e32 v2, 0x3c00 +; GFX9-NEXT: v_sub_f16_e32 v5, v1, v4 +; GFX9-NEXT: v_bfi_b32 v3, s4, v2, v1 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v5|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v3, 0, v3, vcc +; GFX9-NEXT: v_lshrrev_b32_e32 v1, 16, v1 +; GFX9-NEXT: v_add_f16_e32 v3, v4, v3 +; GFX9-NEXT: v_trunc_f16_e32 v4, v1 +; GFX9-NEXT: v_sub_f16_e32 v5, v1, v4 +; GFX9-NEXT: v_bfi_b32 v1, s4, v2, v1 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v5|, 0.5 +; GFX9-NEXT: v_trunc_f16_e32 v5, v0 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX9-NEXT: v_sub_f16_e32 v6, v0, v5 +; GFX9-NEXT: v_add_f16_e32 v1, v4, v1 +; GFX9-NEXT: v_bfi_b32 v4, s4, v2, v0 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v6|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v4, 0, v4, vcc +; GFX9-NEXT: v_lshrrev_b32_e32 v0, 16, v0 +; GFX9-NEXT: v_add_f16_e32 v4, v5, v4 +; GFX9-NEXT: v_trunc_f16_e32 v5, v0 +; GFX9-NEXT: v_sub_f16_e32 v6, v0, v5 +; GFX9-NEXT: v_bfi_b32 v0, s4, v2, v0 +; GFX9-NEXT: v_cmp_ge_f16_e64 vcc, |v6|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc +; GFX9-NEXT: v_add_f16_e32 v0, v5, v0 +; GFX9-NEXT: v_pack_b32_f16 v0, v4, v0 +; GFX9-NEXT: v_pack_b32_f16 v1, v3, v1 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v4f16: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f16_e32 v2, v1 +; GFX10-NEXT: s_movk_i32 s4, 0x3c00 +; GFX10-NEXT: v_lshrrev_b32_e32 v4, 16, v1 +; GFX10-NEXT: v_bfi_b32 v3, 0x7fff, s4, v1 +; GFX10-NEXT: v_lshrrev_b32_e32 v5, 16, v0 +; GFX10-NEXT: v_trunc_f16_e32 v6, v0 +; GFX10-NEXT: v_sub_f16_e32 v1, v1, v2 +; GFX10-NEXT: v_bfi_b32 v7, 0x7fff, s4, v0 +; GFX10-NEXT: v_trunc_f16_e32 v8, v4 +; GFX10-NEXT: v_trunc_f16_e32 v9, v5 +; GFX10-NEXT: v_sub_f16_e32 v0, v0, v6 +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v1|, 0.5 +; GFX10-NEXT: v_sub_f16_e32 v10, v5, v9 +; GFX10-NEXT: v_bfi_b32 v5, 0x7fff, s4, v5 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc_lo +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v0|, 0.5 +; GFX10-NEXT: v_sub_f16_e32 v3, v4, v8 +; GFX10-NEXT: v_bfi_b32 v4, 0x7fff, s4, v4 +; GFX10-NEXT: v_add_f16_e32 v1, v2, v1 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v7, vcc_lo +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v10|, 0.5 +; GFX10-NEXT: v_add_f16_e32 v0, v6, v0 +; GFX10-NEXT: v_cndmask_b32_e32 v5, 0, v5, vcc_lo +; GFX10-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v3|, 0.5 +; GFX10-NEXT: v_add_f16_e32 v2, v9, v5 +; GFX10-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc_lo +; GFX10-NEXT: v_pack_b32_f16 v0, v0, v2 +; GFX10-NEXT: v_add_f16_e32 v3, v8, v3 +; GFX10-NEXT: v_pack_b32_f16 v1, v1, v3 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v4f16: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f16_e32 v2, v1 +; GFX11-NEXT: s_movk_i32 s0, 0x3c00 +; GFX11-NEXT: v_lshrrev_b32_e32 v4, 16, v1 +; GFX11-NEXT: v_bfi_b32 v3, 0x7fff, s0, v1 +; GFX11-NEXT: v_lshrrev_b32_e32 v5, 16, v0 +; GFX11-NEXT: v_trunc_f16_e32 v6, v0 +; GFX11-NEXT: v_sub_f16_e32 v1, v1, v2 +; GFX11-NEXT: v_bfi_b32 v7, 0x7fff, s0, v0 +; GFX11-NEXT: v_trunc_f16_e32 v8, v4 +; GFX11-NEXT: v_trunc_f16_e32 v9, v5 +; GFX11-NEXT: v_sub_f16_e32 v0, v0, v6 +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v1|, 0.5 +; GFX11-NEXT: v_sub_f16_e32 v10, v5, v9 +; GFX11-NEXT: v_bfi_b32 v5, 0x7fff, s0, v5 +; GFX11-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc_lo +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v0|, 0.5 +; GFX11-NEXT: v_sub_f16_e32 v3, v4, v8 +; GFX11-NEXT: v_bfi_b32 v4, 0x7fff, s0, v4 +; GFX11-NEXT: v_add_f16_e32 v1, v2, v1 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v7, vcc_lo +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v10|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v5, 0, v5, vcc_lo +; GFX11-NEXT: v_cmp_ge_f16_e64 vcc_lo, |v3|, 0.5 +; GFX11-NEXT: v_add_f16_e32 v0, v6, v0 +; GFX11-NEXT: v_add_f16_e32 v2, v9, v5 +; GFX11-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc_lo +; GFX11-NEXT: v_pack_b32_f16 v0, v0, v2 +; GFX11-NEXT: v_add_f16_e32 v3, v8, v3 +; GFX11-NEXT: v_pack_b32_f16 v1, v1, v3 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <4 x half> @llvm.roundeven.v4f16(<4 x half> %x) + ret <4 x half> %roundeven +} + +define float @roundeven_f32(float %x) { +; GFX6-LABEL: roundeven_f32: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_trunc_f32_e32 v2, v0 +; GFX6-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_f32: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_trunc_f32_e32 v2, v0 +; GFX7-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_f32: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_trunc_f32_e32 v2, v0 +; GFX8-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX8-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX8-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_trunc_f32_e32 v2, v0 +; GFX9-NEXT: v_bfi_b32 v1, s4, 1.0, v0 +; GFX9-NEXT: v_sub_f32_e32 v0, v0, v2 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v1, vcc +; GFX9-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10PLUS-LABEL: roundeven_f32: +; GFX10PLUS: ; %bb.0: +; GFX10PLUS-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10PLUS-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10PLUS-NEXT: v_trunc_f32_e32 v1, v0 +; GFX10PLUS-NEXT: v_sub_f32_e32 v2, v0, v1 +; GFX10PLUS-NEXT: v_bfi_b32 v0, 0x7fffffff, 1.0, v0 +; GFX10PLUS-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v2|, 0.5 +; GFX10PLUS-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX10PLUS-NEXT: v_add_f32_e32 v0, v1, v0 +; GFX10PLUS-NEXT: s_setpc_b64 s[30:31] + %roundeven = call float @llvm.roundeven.f32(float %x) + ret float %roundeven +} + +define <2 x float> @roundeven_v2f32(<2 x float> %x) { +; GFX6-LABEL: roundeven_v2f32: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_trunc_f32_e32 v3, v0 +; GFX6-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX6-NEXT: v_trunc_f32_e32 v3, v1 +; GFX6-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX6-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX6-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v2f32: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_trunc_f32_e32 v3, v0 +; GFX7-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX7-NEXT: v_trunc_f32_e32 v3, v1 +; GFX7-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX7-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX7-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v2f32: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_trunc_f32_e32 v3, v0 +; GFX8-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX8-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX8-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX8-NEXT: v_trunc_f32_e32 v3, v1 +; GFX8-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX8-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX8-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v2f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_trunc_f32_e32 v3, v0 +; GFX9-NEXT: v_bfi_b32 v2, s4, 1.0, v0 +; GFX9-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v2, vcc +; GFX9-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX9-NEXT: v_trunc_f32_e32 v3, v1 +; GFX9-NEXT: v_bfi_b32 v2, s4, 1.0, v1 +; GFX9-NEXT: v_sub_f32_e32 v1, v1, v3 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v2, vcc +; GFX9-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v2f32: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f32_e32 v2, v0 +; GFX10-NEXT: v_trunc_f32_e32 v3, v1 +; GFX10-NEXT: v_sub_f32_e32 v4, v0, v2 +; GFX10-NEXT: v_bfi_b32 v0, 0x7fffffff, 1.0, v0 +; GFX10-NEXT: v_sub_f32_e32 v5, v1, v3 +; GFX10-NEXT: v_bfi_b32 v1, 0x7fffffff, 1.0, v1 +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v4|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v5|, 0.5 +; GFX10-NEXT: v_add_f32_e32 v0, v2, v0 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc_lo +; GFX10-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v2f32: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f32_e32 v2, v0 +; GFX11-NEXT: v_trunc_f32_e32 v3, v1 +; GFX11-NEXT: v_dual_sub_f32 v4, v0, v2 :: v_dual_sub_f32 v5, v1, v3 +; GFX11-NEXT: v_bfi_b32 v0, 0x7fffffff, 1.0, v0 +; GFX11-NEXT: v_bfi_b32 v1, 0x7fffffff, 1.0, v1 +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v4|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v0, vcc_lo +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v5|, 0.5 +; GFX11-NEXT: v_dual_cndmask_b32 v1, 0, v1 :: v_dual_add_f32 v0, v2, v0 +; GFX11-NEXT: v_add_f32_e32 v1, v3, v1 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <2 x float> @llvm.roundeven.v2f32(<2 x float> %x) + ret <2 x float> %roundeven +} + +define <3 x float> @roundeven_v3f32(<3 x float> %x) { +; GFX6-LABEL: roundeven_v3f32: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_trunc_f32_e32 v4, v0 +; GFX6-NEXT: v_bfi_b32 v3, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v3, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX6-NEXT: v_trunc_f32_e32 v4, v1 +; GFX6-NEXT: v_bfi_b32 v3, s4, 1.0, v1 +; GFX6-NEXT: v_sub_f32_e32 v1, v1, v4 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc +; GFX6-NEXT: v_add_f32_e32 v1, v4, v1 +; GFX6-NEXT: v_trunc_f32_e32 v4, v2 +; GFX6-NEXT: v_bfi_b32 v3, s4, 1.0, v2 +; GFX6-NEXT: v_sub_f32_e32 v2, v2, v4 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v2, 0, v3, vcc +; GFX6-NEXT: v_add_f32_e32 v2, v4, v2 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v3f32: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_trunc_f32_e32 v4, v0 +; GFX7-NEXT: v_bfi_b32 v3, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v3, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX7-NEXT: v_trunc_f32_e32 v4, v1 +; GFX7-NEXT: v_bfi_b32 v3, s4, 1.0, v1 +; GFX7-NEXT: v_sub_f32_e32 v1, v1, v4 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc +; GFX7-NEXT: v_add_f32_e32 v1, v4, v1 +; GFX7-NEXT: v_trunc_f32_e32 v4, v2 +; GFX7-NEXT: v_bfi_b32 v3, s4, 1.0, v2 +; GFX7-NEXT: v_sub_f32_e32 v2, v2, v4 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v2, 0, v3, vcc +; GFX7-NEXT: v_add_f32_e32 v2, v4, v2 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v3f32: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_trunc_f32_e32 v4, v0 +; GFX8-NEXT: v_bfi_b32 v3, s4, 1.0, v0 +; GFX8-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v3, vcc +; GFX8-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX8-NEXT: v_trunc_f32_e32 v4, v1 +; GFX8-NEXT: v_bfi_b32 v3, s4, 1.0, v1 +; GFX8-NEXT: v_sub_f32_e32 v1, v1, v4 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc +; GFX8-NEXT: v_add_f32_e32 v1, v4, v1 +; GFX8-NEXT: v_trunc_f32_e32 v4, v2 +; GFX8-NEXT: v_bfi_b32 v3, s4, 1.0, v2 +; GFX8-NEXT: v_sub_f32_e32 v2, v2, v4 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v2, 0, v3, vcc +; GFX8-NEXT: v_add_f32_e32 v2, v4, v2 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v3f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_trunc_f32_e32 v4, v0 +; GFX9-NEXT: v_bfi_b32 v3, s4, 1.0, v0 +; GFX9-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v3, vcc +; GFX9-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX9-NEXT: v_trunc_f32_e32 v4, v1 +; GFX9-NEXT: v_bfi_b32 v3, s4, 1.0, v1 +; GFX9-NEXT: v_sub_f32_e32 v1, v1, v4 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v3, vcc +; GFX9-NEXT: v_add_f32_e32 v1, v4, v1 +; GFX9-NEXT: v_trunc_f32_e32 v4, v2 +; GFX9-NEXT: v_bfi_b32 v3, s4, 1.0, v2 +; GFX9-NEXT: v_sub_f32_e32 v2, v2, v4 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v2, 0, v3, vcc +; GFX9-NEXT: v_add_f32_e32 v2, v4, v2 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v3f32: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f32_e32 v3, v0 +; GFX10-NEXT: v_trunc_f32_e32 v4, v1 +; GFX10-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v0 +; GFX10-NEXT: v_trunc_f32_e32 v6, v2 +; GFX10-NEXT: v_bfi_b32 v7, 0x7fffffff, 1.0, v1 +; GFX10-NEXT: v_sub_f32_e32 v0, v0, v3 +; GFX10-NEXT: v_sub_f32_e32 v1, v1, v4 +; GFX10-NEXT: v_bfi_b32 v8, 0x7fffffff, 1.0, v2 +; GFX10-NEXT: v_sub_f32_e32 v2, v2, v6 +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v0|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v5, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v1|, 0.5 +; GFX10-NEXT: v_add_f32_e32 v0, v3, v0 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v7, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v2|, 0.5 +; GFX10-NEXT: v_add_f32_e32 v1, v4, v1 +; GFX10-NEXT: v_cndmask_b32_e32 v2, 0, v8, vcc_lo +; GFX10-NEXT: v_add_f32_e32 v2, v6, v2 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v3f32: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f32_e32 v3, v0 +; GFX11-NEXT: v_trunc_f32_e32 v4, v1 +; GFX11-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v0 +; GFX11-NEXT: v_trunc_f32_e32 v6, v2 +; GFX11-NEXT: v_bfi_b32 v7, 0x7fffffff, 1.0, v1 +; GFX11-NEXT: v_dual_sub_f32 v0, v0, v3 :: v_dual_sub_f32 v1, v1, v4 +; GFX11-NEXT: v_bfi_b32 v8, 0x7fffffff, 1.0, v2 +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v0|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v5, vcc_lo +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v1|, 0.5 +; GFX11-NEXT: v_sub_f32_e32 v2, v2, v6 +; GFX11-NEXT: v_dual_add_f32 v0, v3, v0 :: v_dual_cndmask_b32 v1, 0, v7 +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v2|, 0.5 +; GFX11-NEXT: v_dual_add_f32 v1, v4, v1 :: v_dual_cndmask_b32 v2, 0, v8 +; GFX11-NEXT: v_add_f32_e32 v2, v6, v2 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <3 x float> @llvm.roundeven.v3f32(<3 x float> %x) + ret <3 x float> %roundeven +} + +define <4 x float> @roundeven_v4f32(<4 x float> %x) { +; GFX6-LABEL: roundeven_v4f32: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_trunc_f32_e32 v5, v0 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX6-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX6-NEXT: v_trunc_f32_e32 v5, v1 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX6-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX6-NEXT: v_trunc_f32_e32 v5, v2 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX6-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX6-NEXT: v_trunc_f32_e32 v5, v3 +; GFX6-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX6-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX6-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX6-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v4f32: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_trunc_f32_e32 v5, v0 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX7-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX7-NEXT: v_trunc_f32_e32 v5, v1 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX7-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX7-NEXT: v_trunc_f32_e32 v5, v2 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX7-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX7-NEXT: v_trunc_f32_e32 v5, v3 +; GFX7-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX7-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX7-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX7-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v4f32: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_trunc_f32_e32 v5, v0 +; GFX8-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX8-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX8-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX8-NEXT: v_trunc_f32_e32 v5, v1 +; GFX8-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX8-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX8-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX8-NEXT: v_trunc_f32_e32 v5, v2 +; GFX8-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX8-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX8-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX8-NEXT: v_trunc_f32_e32 v5, v3 +; GFX8-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX8-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX8-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX8-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v4f32: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_trunc_f32_e32 v5, v0 +; GFX9-NEXT: v_bfi_b32 v4, s4, 1.0, v0 +; GFX9-NEXT: v_sub_f32_e32 v0, v0, v5 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v0|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v0, 0, v4, vcc +; GFX9-NEXT: v_add_f32_e32 v0, v5, v0 +; GFX9-NEXT: v_trunc_f32_e32 v5, v1 +; GFX9-NEXT: v_bfi_b32 v4, s4, 1.0, v1 +; GFX9-NEXT: v_sub_f32_e32 v1, v1, v5 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v1|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v4, vcc +; GFX9-NEXT: v_add_f32_e32 v1, v5, v1 +; GFX9-NEXT: v_trunc_f32_e32 v5, v2 +; GFX9-NEXT: v_bfi_b32 v4, s4, 1.0, v2 +; GFX9-NEXT: v_sub_f32_e32 v2, v2, v5 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v2|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v2, 0, v4, vcc +; GFX9-NEXT: v_add_f32_e32 v2, v5, v2 +; GFX9-NEXT: v_trunc_f32_e32 v5, v3 +; GFX9-NEXT: v_bfi_b32 v4, s4, 1.0, v3 +; GFX9-NEXT: v_sub_f32_e32 v3, v3, v5 +; GFX9-NEXT: v_cmp_ge_f32_e64 vcc, |v3|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v3, 0, v4, vcc +; GFX9-NEXT: v_add_f32_e32 v3, v5, v3 +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v4f32: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f32_e32 v4, v0 +; GFX10-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v0 +; GFX10-NEXT: v_trunc_f32_e32 v7, v1 +; GFX10-NEXT: v_bfi_b32 v6, 0x7fffffff, 1.0, v1 +; GFX10-NEXT: v_trunc_f32_e32 v8, v2 +; GFX10-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX10-NEXT: v_bfi_b32 v9, 0x7fffffff, 1.0, v2 +; GFX10-NEXT: v_sub_f32_e32 v1, v1, v7 +; GFX10-NEXT: v_trunc_f32_e32 v10, v3 +; GFX10-NEXT: v_sub_f32_e32 v2, v2, v8 +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v0|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v0, 0, v5, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v1|, 0.5 +; GFX10-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v3 +; GFX10-NEXT: v_sub_f32_e32 v3, v3, v10 +; GFX10-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v6, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v2|, 0.5 +; GFX10-NEXT: v_add_f32_e32 v1, v7, v1 +; GFX10-NEXT: v_cndmask_b32_e32 v2, 0, v9, vcc_lo +; GFX10-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v3|, 0.5 +; GFX10-NEXT: v_add_f32_e32 v2, v8, v2 +; GFX10-NEXT: v_cndmask_b32_e32 v3, 0, v5, vcc_lo +; GFX10-NEXT: v_add_f32_e32 v3, v10, v3 +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v4f32: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f32_e32 v4, v0 +; GFX11-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v0 +; GFX11-NEXT: v_trunc_f32_e32 v7, v1 +; GFX11-NEXT: v_bfi_b32 v6, 0x7fffffff, 1.0, v1 +; GFX11-NEXT: v_trunc_f32_e32 v8, v2 +; GFX11-NEXT: v_sub_f32_e32 v0, v0, v4 +; GFX11-NEXT: v_bfi_b32 v9, 0x7fffffff, 1.0, v2 +; GFX11-NEXT: v_sub_f32_e32 v1, v1, v7 +; GFX11-NEXT: v_trunc_f32_e32 v10, v3 +; GFX11-NEXT: v_sub_f32_e32 v2, v2, v8 +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v0|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v0, 0, v5, vcc_lo +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v1|, 0.5 +; GFX11-NEXT: v_bfi_b32 v5, 0x7fffffff, 1.0, v3 +; GFX11-NEXT: v_sub_f32_e32 v3, v3, v10 +; GFX11-NEXT: v_cndmask_b32_e32 v1, 0, v6, vcc_lo +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v2|, 0.5 +; GFX11-NEXT: v_add_f32_e32 v0, v4, v0 +; GFX11-NEXT: v_cndmask_b32_e32 v2, 0, v9, vcc_lo +; GFX11-NEXT: v_cmp_ge_f32_e64 vcc_lo, |v3|, 0.5 +; GFX11-NEXT: v_cndmask_b32_e32 v3, 0, v5, vcc_lo +; GFX11-NEXT: v_add_f32_e32 v3, v10, v3 +; GFX11-NEXT: v_dual_add_f32 v1, v7, v1 :: v_dual_add_f32 v2, v8, v2 +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <4 x float> @llvm.roundeven.v4f32(<4 x float> %x) + ret <4 x float> %roundeven +} + +define double @roundeven_f64(double %x) { +; GFX6-LABEL: roundeven_f64: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: v_bfe_u32 v2, v1, 20, 11 +; GFX6-NEXT: s_mov_b32 s4, -1 +; GFX6-NEXT: v_add_i32_e32 v4, vcc, 0xfffffc01, v2 +; GFX6-NEXT: s_mov_b32 s5, 0xfffff +; GFX6-NEXT: v_lshr_b64 v[2:3], s[4:5], v4 +; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v1 +; GFX6-NEXT: v_not_b32_e32 v3, v3 +; GFX6-NEXT: v_not_b32_e32 v2, v2 +; GFX6-NEXT: v_and_b32_e32 v3, v1, v3 +; GFX6-NEXT: v_and_b32_e32 v2, v0, v2 +; GFX6-NEXT: v_cmp_gt_i32_e32 vcc, 0, v4 +; GFX6-NEXT: v_cndmask_b32_e32 v3, v3, v5, vcc +; GFX6-NEXT: v_cmp_lt_i32_e64 s[4:5], 51, v4 +; GFX6-NEXT: v_cndmask_b32_e64 v2, v2, 0, vcc +; GFX6-NEXT: v_cndmask_b32_e64 v3, v3, v1, s[4:5] +; GFX6-NEXT: v_cndmask_b32_e64 v2, v2, v0, s[4:5] +; GFX6-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX6-NEXT: s_brev_b32 s4, -2 +; GFX6-NEXT: v_mov_b32_e32 v0, 0x3ff00000 +; GFX6-NEXT: v_cmp_ge_f64_e64 vcc, |v[4:5]|, 0.5 +; GFX6-NEXT: v_bfi_b32 v0, s4, v0, v1 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v0, vcc +; GFX6-NEXT: v_mov_b32_e32 v0, 0 +; GFX6-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_f64: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: v_trunc_f64_e32 v[2:3], v[0:1] +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX7-NEXT: v_mov_b32_e32 v0, 0x3ff00000 +; GFX7-NEXT: v_cmp_ge_f64_e64 vcc, |v[4:5]|, 0.5 +; GFX7-NEXT: v_bfi_b32 v0, s4, v0, v1 +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v0, vcc +; GFX7-NEXT: v_mov_b32_e32 v0, 0 +; GFX7-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_f64: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: v_trunc_f64_e32 v[2:3], v[0:1] +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_mov_b32_e32 v6, 0x3ff00000 +; GFX8-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX8-NEXT: v_bfi_b32 v1, s4, v6, v1 +; GFX8-NEXT: v_mov_b32_e32 v0, 0 +; GFX8-NEXT: v_cmp_ge_f64_e64 vcc, |v[4:5]|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX8-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_f64: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_trunc_f64_e32 v[2:3], v[0:1] +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_mov_b32_e32 v6, 0x3ff00000 +; GFX9-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX9-NEXT: v_bfi_b32 v1, s4, v6, v1 +; GFX9-NEXT: v_mov_b32_e32 v0, 0 +; GFX9-NEXT: v_cmp_ge_f64_e64 vcc, |v[4:5]|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX9-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_f64: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f64_e32 v[2:3], v[0:1] +; GFX10-NEXT: s_brev_b32 s4, -2 +; GFX10-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX10-NEXT: v_bfi_b32 v1, s4, 0x3ff00000, v1 +; GFX10-NEXT: v_mov_b32_e32 v0, 0 +; GFX10-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[4:5]|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc_lo +; GFX10-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_f64: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f64_e32 v[2:3], v[0:1] +; GFX11-NEXT: s_brev_b32 s0, -2 +; GFX11-NEXT: v_add_f64 v[4:5], v[0:1], -v[2:3] +; GFX11-NEXT: v_bfi_b32 v1, s0, 0x3ff00000, v1 +; GFX11-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[4:5]|, 0.5 +; GFX11-NEXT: v_dual_mov_b32 v0, 0 :: v_dual_cndmask_b32 v1, 0, v1 +; GFX11-NEXT: v_add_f64 v[0:1], v[2:3], v[0:1] +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call double @llvm.roundeven.f64(double %x) + ret double %roundeven +} + +define <2 x double> @roundeven_v2f64(<2 x double> %x) { +; GFX6-LABEL: roundeven_v2f64: +; GFX6: ; %bb.0: +; GFX6-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX6-NEXT: v_bfe_u32 v4, v1, 20, 11 +; GFX6-NEXT: s_mov_b32 s6, -1 +; GFX6-NEXT: v_add_i32_e32 v6, vcc, 0xfffffc01, v4 +; GFX6-NEXT: s_mov_b32 s7, 0xfffff +; GFX6-NEXT: v_lshr_b64 v[4:5], s[6:7], v6 +; GFX6-NEXT: v_and_b32_e32 v7, 0x80000000, v1 +; GFX6-NEXT: v_not_b32_e32 v5, v5 +; GFX6-NEXT: v_not_b32_e32 v4, v4 +; GFX6-NEXT: v_and_b32_e32 v5, v1, v5 +; GFX6-NEXT: v_and_b32_e32 v4, v0, v4 +; GFX6-NEXT: v_cmp_gt_i32_e32 vcc, 0, v6 +; GFX6-NEXT: v_cndmask_b32_e32 v5, v5, v7, vcc +; GFX6-NEXT: v_cmp_lt_i32_e64 s[4:5], 51, v6 +; GFX6-NEXT: v_cndmask_b32_e64 v4, v4, 0, vcc +; GFX6-NEXT: v_cndmask_b32_e64 v5, v5, v1, s[4:5] +; GFX6-NEXT: v_cndmask_b32_e64 v4, v4, v0, s[4:5] +; GFX6-NEXT: v_add_f64 v[6:7], v[0:1], -v[4:5] +; GFX6-NEXT: s_brev_b32 s9, -2 +; GFX6-NEXT: v_mov_b32_e32 v8, 0x3ff00000 +; GFX6-NEXT: v_cmp_ge_f64_e64 vcc, |v[6:7]|, 0.5 +; GFX6-NEXT: s_movk_i32 s8, 0xfc01 +; GFX6-NEXT: v_bfi_b32 v0, s9, v8, v1 +; GFX6-NEXT: v_bfe_u32 v6, v3, 20, 11 +; GFX6-NEXT: v_cndmask_b32_e32 v1, 0, v0, vcc +; GFX6-NEXT: v_add_i32_e32 v9, vcc, s8, v6 +; GFX6-NEXT: v_lshr_b64 v[6:7], s[6:7], v9 +; GFX6-NEXT: v_mov_b32_e32 v0, 0 +; GFX6-NEXT: v_add_f64 v[0:1], v[4:5], v[0:1] +; GFX6-NEXT: v_not_b32_e32 v4, v7 +; GFX6-NEXT: v_not_b32_e32 v5, v6 +; GFX6-NEXT: v_and_b32_e32 v4, v3, v4 +; GFX6-NEXT: v_and_b32_e32 v6, v2, v5 +; GFX6-NEXT: v_and_b32_e32 v5, 0x80000000, v3 +; GFX6-NEXT: v_cmp_gt_i32_e32 vcc, 0, v9 +; GFX6-NEXT: v_cndmask_b32_e32 v4, v4, v5, vcc +; GFX6-NEXT: v_cmp_lt_i32_e64 s[4:5], 51, v9 +; GFX6-NEXT: v_cndmask_b32_e64 v5, v4, v3, s[4:5] +; GFX6-NEXT: v_cndmask_b32_e64 v4, v6, 0, vcc +; GFX6-NEXT: v_cndmask_b32_e64 v4, v4, v2, s[4:5] +; GFX6-NEXT: v_add_f64 v[6:7], v[2:3], -v[4:5] +; GFX6-NEXT: v_bfi_b32 v2, s9, v8, v3 +; GFX6-NEXT: v_cmp_ge_f64_e64 vcc, |v[6:7]|, 0.5 +; GFX6-NEXT: v_cndmask_b32_e32 v3, 0, v2, vcc +; GFX6-NEXT: v_mov_b32_e32 v2, 0 +; GFX6-NEXT: v_add_f64 v[2:3], v[4:5], v[2:3] +; GFX6-NEXT: s_setpc_b64 s[30:31] +; +; GFX7-LABEL: roundeven_v2f64: +; GFX7: ; %bb.0: +; GFX7-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX7-NEXT: v_trunc_f64_e32 v[4:5], v[0:1] +; GFX7-NEXT: s_brev_b32 s4, -2 +; GFX7-NEXT: v_add_f64 v[6:7], v[0:1], -v[4:5] +; GFX7-NEXT: v_mov_b32_e32 v8, 0x3ff00000 +; GFX7-NEXT: v_cmp_ge_f64_e64 vcc, |v[6:7]|, 0.5 +; GFX7-NEXT: v_bfi_b32 v0, s4, v8, v1 +; GFX7-NEXT: v_trunc_f64_e32 v[6:7], v[2:3] +; GFX7-NEXT: v_cndmask_b32_e32 v1, 0, v0, vcc +; GFX7-NEXT: v_mov_b32_e32 v0, 0 +; GFX7-NEXT: v_add_f64 v[0:1], v[4:5], v[0:1] +; GFX7-NEXT: v_add_f64 v[4:5], v[2:3], -v[6:7] +; GFX7-NEXT: v_bfi_b32 v2, s4, v8, v3 +; GFX7-NEXT: v_cmp_ge_f64_e64 vcc, |v[4:5]|, 0.5 +; GFX7-NEXT: v_cndmask_b32_e32 v3, 0, v2, vcc +; GFX7-NEXT: v_mov_b32_e32 v2, 0 +; GFX7-NEXT: v_add_f64 v[2:3], v[6:7], v[2:3] +; GFX7-NEXT: s_setpc_b64 s[30:31] +; +; GFX8-LABEL: roundeven_v2f64: +; GFX8: ; %bb.0: +; GFX8-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX8-NEXT: v_trunc_f64_e32 v[4:5], v[0:1] +; GFX8-NEXT: v_trunc_f64_e32 v[6:7], v[2:3] +; GFX8-NEXT: s_brev_b32 s4, -2 +; GFX8-NEXT: v_mov_b32_e32 v12, 0x3ff00000 +; GFX8-NEXT: v_add_f64 v[8:9], v[0:1], -v[4:5] +; GFX8-NEXT: v_add_f64 v[10:11], v[2:3], -v[6:7] +; GFX8-NEXT: v_bfi_b32 v1, s4, v12, v1 +; GFX8-NEXT: v_bfi_b32 v3, s4, v12, v3 +; GFX8-NEXT: v_mov_b32_e32 v0, 0 +; GFX8-NEXT: v_mov_b32_e32 v2, 0 +; GFX8-NEXT: v_cmp_ge_f64_e64 vcc, |v[8:9]|, 0.5 +; GFX8-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX8-NEXT: v_cmp_ge_f64_e64 vcc, |v[10:11]|, 0.5 +; GFX8-NEXT: v_add_f64 v[0:1], v[4:5], v[0:1] +; GFX8-NEXT: v_cndmask_b32_e32 v3, 0, v3, vcc +; GFX8-NEXT: v_add_f64 v[2:3], v[6:7], v[2:3] +; GFX8-NEXT: s_setpc_b64 s[30:31] +; +; GFX9-LABEL: roundeven_v2f64: +; GFX9: ; %bb.0: +; GFX9-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX9-NEXT: v_trunc_f64_e32 v[4:5], v[0:1] +; GFX9-NEXT: v_trunc_f64_e32 v[6:7], v[2:3] +; GFX9-NEXT: s_brev_b32 s4, -2 +; GFX9-NEXT: v_mov_b32_e32 v12, 0x3ff00000 +; GFX9-NEXT: v_add_f64 v[8:9], v[0:1], -v[4:5] +; GFX9-NEXT: v_add_f64 v[10:11], v[2:3], -v[6:7] +; GFX9-NEXT: v_bfi_b32 v1, s4, v12, v1 +; GFX9-NEXT: v_bfi_b32 v3, s4, v12, v3 +; GFX9-NEXT: v_mov_b32_e32 v0, 0 +; GFX9-NEXT: v_mov_b32_e32 v2, 0 +; GFX9-NEXT: v_cmp_ge_f64_e64 vcc, |v[8:9]|, 0.5 +; GFX9-NEXT: v_cndmask_b32_e32 v1, 0, v1, vcc +; GFX9-NEXT: v_cmp_ge_f64_e64 vcc, |v[10:11]|, 0.5 +; GFX9-NEXT: v_add_f64 v[0:1], v[4:5], v[0:1] +; GFX9-NEXT: v_cndmask_b32_e32 v3, 0, v3, vcc +; GFX9-NEXT: v_add_f64 v[2:3], v[6:7], v[2:3] +; GFX9-NEXT: s_setpc_b64 s[30:31] +; +; GFX10-LABEL: roundeven_v2f64: +; GFX10: ; %bb.0: +; GFX10-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX10-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX10-NEXT: v_trunc_f64_e32 v[4:5], v[0:1] +; GFX10-NEXT: v_trunc_f64_e32 v[6:7], v[2:3] +; GFX10-NEXT: s_brev_b32 s4, -2 +; GFX10-NEXT: v_mov_b32_e32 v12, 0 +; GFX10-NEXT: v_add_f64 v[8:9], v[0:1], -v[4:5] +; GFX10-NEXT: v_add_f64 v[10:11], v[2:3], -v[6:7] +; GFX10-NEXT: v_bfi_b32 v0, s4, 0x3ff00000, v1 +; GFX10-NEXT: v_bfi_b32 v2, s4, 0x3ff00000, v3 +; GFX10-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[8:9]|, 0.5 +; GFX10-NEXT: v_cndmask_b32_e32 v13, 0, v0, vcc_lo +; GFX10-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[10:11]|, 0.5 +; GFX10-NEXT: v_add_f64 v[0:1], v[4:5], v[12:13] +; GFX10-NEXT: v_cndmask_b32_e32 v13, 0, v2, vcc_lo +; GFX10-NEXT: v_add_f64 v[2:3], v[6:7], v[12:13] +; GFX10-NEXT: s_setpc_b64 s[30:31] +; +; GFX11-LABEL: roundeven_v2f64: +; GFX11: ; %bb.0: +; GFX11-NEXT: s_waitcnt vmcnt(0) expcnt(0) lgkmcnt(0) +; GFX11-NEXT: s_waitcnt_vscnt null, 0x0 +; GFX11-NEXT: v_trunc_f64_e32 v[4:5], v[0:1] +; GFX11-NEXT: v_trunc_f64_e32 v[6:7], v[2:3] +; GFX11-NEXT: s_brev_b32 s0, -2 +; GFX11-NEXT: v_add_f64 v[8:9], v[0:1], -v[4:5] +; GFX11-NEXT: v_add_f64 v[10:11], v[2:3], -v[6:7] +; GFX11-NEXT: v_bfi_b32 v0, s0, 0x3ff00000, v1 +; GFX11-NEXT: v_bfi_b32 v2, s0, 0x3ff00000, v3 +; GFX11-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[8:9]|, 0.5 +; GFX11-NEXT: v_dual_mov_b32 v12, 0 :: v_dual_cndmask_b32 v13, 0, v0 +; GFX11-NEXT: v_cmp_ge_f64_e64 vcc_lo, |v[10:11]|, 0.5 +; GFX11-NEXT: v_add_f64 v[0:1], v[4:5], v[12:13] +; GFX11-NEXT: v_cndmask_b32_e32 v13, 0, v2, vcc_lo +; GFX11-NEXT: v_add_f64 v[2:3], v[6:7], v[12:13] +; GFX11-NEXT: s_setpc_b64 s[30:31] + %roundeven = call <2 x double> @llvm.roundeven.v2f64(<2 x double> %x) + ret <2 x double> %roundeven +} + +declare half @llvm.roundeven.f16(half) #0 +declare <2 x half> @llvm.roundeven.v2f16(<2 x half>) #0 +declare <4 x half> @llvm.roundeven.v4f16(<4 x half>) #0 + +declare float @llvm.roundeven.f32(float) #0 +declare <2 x float> @llvm.roundeven.v2f32(<2 x float>) #0 +declare <3 x float> @llvm.roundeven.v3f32(<3 x float>) #0 +declare <4 x float> @llvm.roundeven.v4f32(<4 x float>) #0 + +declare double @llvm.roundeven.f64(double) #0 +declare <2 x double> @llvm.roundeven.v2f64(<2 x double>) #0 + +attributes #0 = { nounwind readnone speculatable willreturn }