diff --git a/llvm/lib/Target/RISCV/RISCVInstrInfoVSDPatterns.td b/llvm/lib/Target/RISCV/RISCVInstrInfoVSDPatterns.td --- a/llvm/lib/Target/RISCV/RISCVInstrInfoVSDPatterns.td +++ b/llvm/lib/Target/RISCV/RISCVInstrInfoVSDPatterns.td @@ -448,6 +448,92 @@ defm : VPatWidenBinaryFPSDNode_WV_WF; } +multiclass VPatWidenFPMulAccSDNode_VV_VF { + foreach vti = AllWidenableFloatVectors in { + def : Pat<(fma (vti.Wti.Vector (fpext_oneuse (vti.Vti.Vector vti.Vti.RegClass:$rs1))), + (vti.Wti.Vector (fpext_oneuse (vti.Vti.Vector vti.Vti.RegClass:$rs2))), + (vti.Wti.Vector vti.Wti.RegClass:$rd)), + (!cast(instruction_name#"_VV_"#vti.Vti.LMul.MX) + vti.Wti.RegClass:$rd, vti.Vti.RegClass:$rs1, vti.Vti.RegClass:$rs2, + vti.Vti.AVL, vti.Vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (vti.Wti.Vector (fpext_oneuse (vti.Vti.Vector (splat_vector vti.Vti.ScalarRegClass:$rs1)))), + (vti.Wti.Vector (fpext_oneuse (vti.Vti.Vector vti.Vti.RegClass:$rs2))), + (vti.Wti.Vector vti.Wti.RegClass:$rd)), + (!cast(instruction_name#"_V"#vti.Vti.ScalarSuffix#"_"#vti.Vti.LMul.MX) + vti.Wti.RegClass:$rd, vti.Vti.ScalarRegClass:$rs1, vti.Vti.RegClass:$rs2, + vti.Vti.AVL, vti.Vti.Log2SEW, TAIL_AGNOSTIC)>; + } +} + +multiclass VPatWidenFPNegMulAccSDNode_VV_VF { + foreach vtiToWti = AllWidenableFloatVectors in { + defvar vti = vtiToWti.Vti; + defvar wti = vtiToWti.Wti; + def : Pat<(fma (fneg (wti.Vector (fpext_oneuse (vti.Vector vti.RegClass:$rs1)))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + (fneg wti.RegClass:$rd)), + (!cast(instruction_name#"_VV_"#vti.LMul.MX) + wti.RegClass:$rd, vti.RegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (fpext_oneuse (vti.Vector (splat_vector vti.ScalarRegClass:$rs1))), + (fneg (wti.Vector (fpext_oneuse (vti.Vector vti.RegClass:$rs2)))), + (fneg wti.RegClass:$rd)), + (!cast(instruction_name#"_V"#vti.ScalarSuffix#"_"#vti.LMul.MX) + wti.RegClass:$rd, vti.ScalarRegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (fneg (wti.Vector (fpext_oneuse (vti.Vector (splat_vector vti.ScalarRegClass:$rs1))))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + (fneg wti.RegClass:$rd)), + (!cast(instruction_name#"_V"#vti.ScalarSuffix#"_"#vti.LMul.MX) + wti.RegClass:$rd, vti.ScalarRegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + } +} + +multiclass VPatWidenFPMulSacSDNode_VV_VF { + foreach vtiToWti = AllWidenableFloatVectors in { + defvar vti = vtiToWti.Vti; + defvar wti = vtiToWti.Wti; + def : Pat<(fma (wti.Vector (fpext_oneuse (vti.Vector vti.RegClass:$rs1))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + (fneg wti.RegClass:$rd)), + (!cast(instruction_name#"_VV_"#vti.LMul.MX) + wti.RegClass:$rd, vti.RegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (wti.Vector (fpext_oneuse (vti.Vector (splat_vector vti.ScalarRegClass:$rs1)))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + (fneg wti.RegClass:$rd)), + (!cast(instruction_name#"_V"#vti.ScalarSuffix#"_"#vti.LMul.MX) + wti.RegClass:$rd, vti.ScalarRegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + } +} + +multiclass VPatWidenFPNegMulSacSDNode_VV_VF { + foreach vtiToWti = AllWidenableFloatVectors in { + defvar vti = vtiToWti.Vti; + defvar wti = vtiToWti.Wti; + def : Pat<(fma (fneg (wti.Vector (fpext_oneuse (vti.Vector vti.RegClass:$rs1)))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + wti.RegClass:$rd), + (!cast(instruction_name#"_VV_"#vti.LMul.MX) + wti.RegClass:$rd, vti.RegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (wti.Vector (fpext_oneuse (vti.Vector (splat_vector vti.ScalarRegClass:$rs1)))), + (fneg (wti.Vector (fpext_oneuse (vti.Vector vti.RegClass:$rs2)))), + wti.RegClass:$rd), + (!cast(instruction_name#"_V"#vti.ScalarSuffix#"_"#vti.LMul.MX) + wti.RegClass:$rd, vti.ScalarRegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + def : Pat<(fma (fneg (wti.Vector (fpext_oneuse (vti.Vector (splat_vector vti.ScalarRegClass:$rs1))))), + (fpext_oneuse (vti.Vector vti.RegClass:$rs2)), + wti.RegClass:$rd), + (!cast(instruction_name#"_V"#vti.ScalarSuffix#"_"#vti.LMul.MX) + wti.RegClass:$rd, vti.ScalarRegClass:$rs1, vti.RegClass:$rs2, + vti.AVL, vti.Log2SEW, TAIL_AGNOSTIC)>; + } +} + //===----------------------------------------------------------------------===// // Patterns. //===----------------------------------------------------------------------===// @@ -760,6 +846,12 @@ fvti.AVL, fvti.Log2SEW, TAIL_AGNOSTIC)>; } +// 14.7. Vector Widening Floating-Point Fused Multiply-Add Instructions +defm : VPatWidenFPMulAccSDNode_VV_VF<"PseudoVFWMACC">; +defm : VPatWidenFPNegMulAccSDNode_VV_VF<"PseudoVFWNMACC">; +defm : VPatWidenFPMulSacSDNode_VV_VF<"PseudoVFWMSAC">; +defm : VPatWidenFPNegMulSacSDNode_VV_VF<"PseudoVFWNMSAC">; + foreach vti = AllFloatVectors in { // 14.8. Vector Floating-Point Square-Root Instruction def : Pat<(fsqrt (vti.Vector vti.RegClass:$rs2)), diff --git a/llvm/test/CodeGen/RISCV/rvv/vfwmacc-sdnode.ll b/llvm/test/CodeGen/RISCV/rvv/vfwmacc-sdnode.ll new file mode 100644 --- /dev/null +++ b/llvm/test/CodeGen/RISCV/rvv/vfwmacc-sdnode.ll @@ -0,0 +1,1312 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=riscv32 -mattr=+d,+zfh,+v -target-abi=ilp32d \ +; RUN: -verify-machineinstrs < %s | FileCheck %s +; RUN: llc -mtriple=riscv64 -mattr=+d,+zfh,+v -target-abi=lp64d \ +; RUN: -verify-machineinstrs < %s | FileCheck %s + +declare @llvm.fma.v1f32(, , ) + +define @vfwmacc_vv_nxv1f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v1f32( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmacc_vf_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v1f32( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv1f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v1f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v1f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v1f32( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv1f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v1f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmsac_vf_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v1f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv1f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v1f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v1f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv1f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv1f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v1f32( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v2f32(, , ) + +define @vfwmacc_vv_nxv2f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v2f32( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmacc_vf_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v2f32( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv2f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v2f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v2f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v2f32( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv2f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v2f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmsac_vf_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v2f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv2f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v2f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v2f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv2f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv2f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v2f32( %vd, %vf, %va) + ret %vg +} + + +declare @llvm.fma.v4f32(, , ) + +define @vfwmacc_vv_nxv4f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v4f32( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmacc_vf_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v4f32( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv4f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v4f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v4f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v4f32( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv4f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v4f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmsac_vf_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v4f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv4f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v4f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v4f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv4f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv4f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v4f32( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v8f32(, , ) + +define @vfwmacc_vv_nxv8f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v8f32( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmacc_vf_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v8f32( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv8f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v8f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v8f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v8f32( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv8f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v8f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmsac_vf_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v8f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv8f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v8f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v8f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv8f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv8f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v8f32( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v16f32(, , ) + +define @vfwmacc_vv_nxv16f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v16f32( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmacc_vf_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v16f32( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv16f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v16f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v16f32( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v16f32( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv16f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v16f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwmsac_vf_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v16f32( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv16f32( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v16f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v16f32( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv16f32( %va, %vb, half %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv16f32: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e16, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, half %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v16f32( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v1f64(, , ) + +define @vfwmacc_vv_nxv1f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v1f64( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmacc_vf_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v1f64( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv1f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v1f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v1f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v1f64( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv1f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v1f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmsac_vf_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v1f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv1f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v9, v10 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v1f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v1f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv1f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv1f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, mf2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v9 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v1f64( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v2f64(, , ) + +define @vfwmacc_vv_nxv2f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v2f64( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmacc_vf_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v2f64( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv2f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v2f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v2f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v2f64( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv2f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v2f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmsac_vf_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v2f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv2f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v10, v11 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v2f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v2f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv2f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv2f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m1, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v10 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v2f64( %vd, %vf, %va) + ret %vg +} + + +declare @llvm.fma.v4f64(, , ) + +define @vfwmacc_vv_nxv4f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v4f64( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmacc_vf_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v4f64( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv4f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v4f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v4f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v4f64( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv4f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v4f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmsac_vf_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v4f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv4f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v12, v14 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v4f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v4f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv4f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv4f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m2, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v12 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v4f64( %vd, %vf, %va) + ret %vg +} + +declare @llvm.fma.v8f64(, , ) + +define @vfwmacc_vv_nxv8f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmacc_vv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwmacc.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = call @llvm.fma.v8f64( %vd, %ve, %va) + ret %vf +} + +define @vfwmacc_vf_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmacc_vf_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = call @llvm.fma.v8f64( %vd, %ve, %va) + ret %vf +} + +define @vfwnmacc_vv_nxv8f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmacc_vv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v8f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_vf_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_vf_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %vd + %vh = call @llvm.fma.v8f64( %vg, %ve, %vf) + ret %vh +} + +define @vfwnmacc_fv_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmacc_fv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmacc.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = fneg %ve + %vh = call @llvm.fma.v8f64( %vd, %vg, %vf) + ret %vh +} + +define @vfwmsac_vv_nxv8f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwmsac_vv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwmsac.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %va + %vg = call @llvm.fma.v8f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwmsac_vf_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwmsac_vf_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %va + %vg = call @llvm.fma.v8f64( %vd, %ve, %vf) + ret %vg +} + +define @vfwnmsac_vv_nxv8f64( %va, %vb, %vc) { +; CHECK-LABEL: vfwnmsac_vv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vv v8, v16, v20 +; CHECK-NEXT: ret + %vd = fpext %vb to + %ve = fpext %vc to + %vf = fneg %vd + %vg = call @llvm.fma.v8f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_vf_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_vf_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %vd + %vg = call @llvm.fma.v8f64( %vf, %ve, %va) + ret %vg +} + +define @vfwnmsac_fv_nxv8f64( %va, %vb, float %c) { +; CHECK-LABEL: vfwnmsac_fv_nxv8f64: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e32, m4, ta, mu +; CHECK-NEXT: vfwnmsac.vf v8, fa0, v16 +; CHECK-NEXT: ret + %head = insertelement undef, float %c, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vd = fpext %vb to + %ve = fpext %splat to + %vf = fneg %ve + %vg = call @llvm.fma.v8f64( %vd, %vf, %va) + ret %vg +}