diff --git a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.h b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.h --- a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.h +++ b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.h @@ -130,6 +130,7 @@ private: bool doPeepholeSExtW(SDNode *Node); bool doPeepholeMaskedRVV(SDNode *Node); + bool doPeepholeMergeVVMFold(); }; namespace RISCV { diff --git a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp --- a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp +++ b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp @@ -151,6 +151,8 @@ CurDAG->setRoot(Dummy.getValue()); + MadeChange |= doPeepholeMergeVVMFold(); + if (MadeChange) CurDAG->RemoveDeadNodes(); } @@ -2511,6 +2513,109 @@ return true; } +// Try to fold MERGE_VVM with unmasked intrinsic to masked intrinsic. The +// peephole only deals with MERGE_VVM which is TU and has false operand same as +// its true operand now. E.g. (VMERGE_VVM_M1_TU False, False, (VADD_M1 ...), +// ...) -> (VADD_VV_M1_MASK) +bool RISCVDAGToDAGISel::doPeepholeMergeVVMFold() { + bool MadeChange = false; + SelectionDAG::allnodes_iterator Position = CurDAG->allnodes_end(); + + while (Position != CurDAG->allnodes_begin()) { + SDNode *N = &*--Position; + if (N->use_empty() || !N->isMachineOpcode()) + continue; + + auto IsVMergeTU = [](unsigned Opcode) { + return Opcode == RISCV::PseudoVMERGE_VVM_MF8_TU || + Opcode == RISCV::PseudoVMERGE_VVM_MF4_TU || + Opcode == RISCV::PseudoVMERGE_VVM_MF2_TU || + Opcode == RISCV::PseudoVMERGE_VVM_M1_TU || + Opcode == RISCV::PseudoVMERGE_VVM_M2_TU || + Opcode == RISCV::PseudoVMERGE_VVM_M4_TU || + Opcode == RISCV::PseudoVMERGE_VVM_M8_TU; + }; + + unsigned Opc = N->getMachineOpcode(); + // TODO: Also deal with TA VMerge nodes. + if (!IsVMergeTU(Opc)) + continue; + + SDValue Merge = N->getOperand(0); + SDValue False = N->getOperand(1); + SDValue True = N->getOperand(2); + SDValue Mask = N->getOperand(3); + SDValue VL = N->getOperand(4); + + if (Merge != False) + continue; + + // Need N is the exactly one using True. + if (!True.hasOneUse()) + continue; + + if (!True.isMachineOpcode()) + continue; + + unsigned TrueOpc = True.getMachineOpcode(); + + // Skip if True has merge operand. + // TODO: Deal with True having same merge operand with N. + if (RISCVII::hasMergeOp(TII->get(TrueOpc).TSFlags)) + continue; + + // Skip if True has side effect. + // TODO: Support velff and vlsegff. + if (TII->get(TrueOpc).hasUnmodeledSideEffects()) + continue; + + // Only deal with True when True is unmasked intrinsic now. + const RISCV::RISCVMaskedPseudoInfo *Info = + RISCV::lookupMaskedIntrinsicByUnmaskedTA(TrueOpc); + + if (!Info) + continue; + + // The last operand of unmasked intrinsic should be sew or chain. + bool HasChainOp = + True.getOperand(True.getNumOperands() - 1).getValueType() == MVT::Other; + + // Need True has same VL with N. + unsigned TrueVLIndex = True.getNumOperands() - HasChainOp - 2; + SDValue TrueVL = True.getOperand(TrueVLIndex); + if (TrueVL != VL) + continue; + + SDLoc DL(N); + unsigned MaskedOpc = Info->MaskedPseudo; + SmallVector Ops; + Ops.push_back(Merge); + Ops.append(True->op_begin(), True->op_begin() + TrueVLIndex); + Ops.append({Mask, VL, /* SEW */ True.getOperand(TrueVLIndex + 1)}); + + if (RISCVII::hasVecPolicyOp(TII->get(MaskedOpc).TSFlags)) + Ops.push_back( + CurDAG->getTargetConstant(/* TUMU */ 0, DL, Subtarget->getXLenVT())); + + // Result node should have chain operand of True. + if (HasChainOp) + Ops.push_back(True.getOperand(True.getNumOperands() - 1)); + + // Result node should take over glued node of N. + if (N->getGluedNode()) + Ops.push_back(N->getOperand(N->getNumOperands() - 1)); + + SDNode *Result = + CurDAG->getMachineNode(MaskedOpc, DL, True->getVTList(), Ops); + ReplaceUses(N, Result); + + // Try to transform Result to unmasked intrinsic. + doPeepholeMaskedRVV(Result); + MadeChange = true; + } + return MadeChange; +} + // This pass converts a legalized DAG into a RISCV-specific DAG, ready // for instruction scheduling. FunctionPass *llvm::createRISCVISelDag(RISCVTargetMachine &TM, diff --git a/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td b/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td --- a/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td +++ b/llvm/lib/Target/RISCV/RISCVInstrInfoVPseudos.td @@ -472,6 +472,11 @@ Pseudo Pseudo = !cast(NAME); } +def lookupMaskedIntrinsicByUnmaskedTA : SearchIndex { + let Table = RISCVMaskedPseudosTable; + let Key = ["UnmaskedPseudo"]; +} + def RISCVVLETable : GenericTable { let FilterClass = "RISCVVLE"; let CppTypeName = "VLEPseudo"; @@ -1591,6 +1596,7 @@ VLESched; def "E" # eew # "_V_" # LInfo # "_MASK" : VPseudoUSLoadMask, + RISCVMaskedPseudo, VLESched; } } @@ -1611,6 +1617,7 @@ VLFSched; def "E" # eew # "FF_V_" # LInfo # "_MASK": VPseudoUSLoadFFMask, + RISCVMaskedPseudo, VLFSched; } } @@ -1635,8 +1642,10 @@ VLSSched; def "E" # eew # "_V_" # LInfo # "_TU": VPseudoSLoadNoMaskTU, VLSSched; - def "E" # eew # "_V_" # LInfo # "_MASK" : VPseudoSLoadMask, - VLSSched; + def "E" # eew # "_V_" # LInfo # "_MASK" : + VPseudoSLoadMask, + RISCVMaskedPseudo, + VLSSched; } } } @@ -1666,6 +1675,7 @@ VLXSched; def "EI" # eew # "_V_" # IdxLInfo # "_" # LInfo # "_MASK" : VPseudoILoadMask, + RISCVMaskedPseudo, VLXSched; } } @@ -1785,6 +1795,7 @@ def "_V_" # m.MX # "_TU": VPseudoNullaryNoMaskTU, Sched<[WriteVMIdxV, ReadVMask]>; def "_V_" # m.MX # "_MASK" : VPseudoNullaryMask, + RISCVMaskedPseudo, Sched<[WriteVMIdxV, ReadVMask]>; } } @@ -1807,6 +1818,7 @@ def "_" # m.MX # "_TU" : VPseudoUnaryNoMaskTU, Sched<[WriteVMIotV, ReadVMIotV, ReadVMask]>; def "_" # m.MX # "_MASK" : VPseudoUnaryMaskTA, + RISCVMaskedPseudo, Sched<[WriteVMIotV, ReadVMIotV, ReadVMask]>; } } @@ -2137,6 +2149,7 @@ def "_V_" # m.MX # "_TU": VPseudoUnaryNoMaskTU, Sched<[WriteVFClassV, ReadVFClassV, ReadVMask]>; def "_V_" # m.MX # "_MASK" : VPseudoUnaryMaskTA, + RISCVMaskedPseudo, Sched<[WriteVFClassV, ReadVFClassV, ReadVMask]>; } } @@ -2150,6 +2163,7 @@ def "_V_" # m.MX # "_TU": VPseudoUnaryNoMaskTU, Sched<[WriteVFSqrtV, ReadVFSqrtV, ReadVMask]>; def "_V_" # m.MX # "_MASK" : VPseudoUnaryMaskTA, + RISCVMaskedPseudo, Sched<[WriteVFSqrtV, ReadVFSqrtV, ReadVMask]>; } } @@ -2163,6 +2177,7 @@ def "_V_" # m.MX # "_TU": VPseudoUnaryNoMaskTU, Sched<[WriteVFRecpV, ReadVFRecpV, ReadVMask]>; def "_V_" # m.MX # "_MASK" : VPseudoUnaryMaskTA, + RISCVMaskedPseudo, Sched<[WriteVFRecpV, ReadVFRecpV, ReadVMask]>; } } diff --git a/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll new file mode 100644 --- /dev/null +++ b/llvm/test/CodeGen/RISCV/rvv/rvv-peephole-vmerge-vops.ll @@ -0,0 +1,320 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=riscv64 -mattr=+v | FileCheck %s + +declare @llvm.vp.merge.nxv2i16(, , , i32) +declare @llvm.vp.merge.nxv2i32(, , , i32) +declare @llvm.vp.merge.nxv2f32(, , , i32) +declare @llvm.vp.merge.nxv2f64(, , , i32) + +; Test binary operator with vp.merge and vp.smax. +declare @llvm.vp.add.nxv2i32(, , , i32) +define @vpmerge_vpadd( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vadd.vv v8, v9, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test glued node of merge should not be deleted. +declare @llvm.vp.icmp.nxv2i32(, , metadata, , i32) +define @vpmerge_vpadd2( %passthru, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpadd2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vsetvli zero, zero, e32, m1, tu, mu +; CHECK-NEXT: vadd.vv v8, v9, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %m = call @llvm.vp.icmp.nxv2i32( %x, %y, metadata !"eq", %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test vp.merge have all-ones mask. +define @vpmerge_vpadd3( %passthru, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpadd3: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vadd.vv v8, v9, v10 +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.add.nxv2i32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %mask, %a, %passthru, i32 %vl) + ret %b +} + +; Test float binary operator with vp.merge and vp.fadd. +declare @llvm.vp.fadd.nxv2f32(, , , i32) +define @vpmerge_vpfadd( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpfadd: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfadd.vv v8, v9, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fadd.nxv2f32( %x, %y, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test for binary operator with specific EEW by riscv.vrgatherei16. +declare @llvm.riscv.vrgatherei16.vv.nxv2i32.i64(, , , i64) +define @vpmerge_vrgatherei16( %passthru, %x, %y, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vrgatherei16: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, ta, mu +; CHECK-NEXT: vrgatherei16.vv v8, v9, v10 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %2 = tail call @llvm.riscv.vrgatherei16.vv.nxv2i32.i64( undef, %x, %y, i64 %1) + %3 = tail call @llvm.vp.merge.nxv2i32( %m, %2, %passthru, i32 %vl) + ret %2 +} + +; Test conversion by fptosi. +declare @llvm.vp.fptosi.nxv2i16.nxv2f32(, , i32) +define @vpmerge_vpfptosi( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpfptosi: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e16, mf2, tu, mu +; CHECK-NEXT: vfncvt.rtz.x.f.w v8, v9, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fptosi.nxv2i16.nxv2f32( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i16( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test conversion by sitofp. +declare @llvm.vp.sitofp.nxv2f32.nxv2i64(, , i32) +define @vpmerge_vpsitofp( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpsitofp: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfncvt.f.x.w v8, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.sitofp.nxv2f32.nxv2i64( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer extension by vp.zext. +declare @llvm.vp.zext.nxv2i32.nxv2i8(, , i32) +define @vpmerge_vpzext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpzext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vzext.vf4 v8, v9, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.zext.nxv2i32.nxv2i8( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer truncation by vp.trunc. +declare @llvm.vp.trunc.nxv2i32.nxv2i64(, , i32) +define @vpmerge_vptrunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vncvt.x.x.w v8, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.trunc.nxv2i32.nxv2i64( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer extension by vp.fpext. +declare @llvm.vp.fpext.nxv2f64.nxv2f32(, , i32) +define @vpmerge_vpfpext( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpfpext: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfwcvt.f.f.v v8, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fpext.nxv2f64.nxv2f32( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2f64( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test integer truncation by vp.trunc. +declare @llvm.vp.fptrunc.nxv2f32.nxv2f64(, , i32) +define @vpmerge_vpfptrunc( %passthru, %x, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpfptrunc: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfncvt.f.f.w v8, v10, v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.fptrunc.nxv2f32.nxv2f64( %x, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test load operation by vp.load. +declare @llvm.vp.load.nxv2i32.p0nxv2i32( *, , i32) +define @vpmerge_vpload( %passthru, * %p, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpload: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, tu, mu +; CHECK-NEXT: vle32.v v8, (a0), v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test result have chain and glued node. +define @vpmerge_vpload2( %passthru, * %p, %x, %y, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vpload2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vmseq.vv v0, v9, v10 +; CHECK-NEXT: vsetvli zero, zero, e32, m1, tu, mu +; CHECK-NEXT: vle32.v v8, (a0), v0.t +; CHECK-NEXT: ret + %splat = insertelement poison, i1 -1, i32 0 + %mask = shufflevector %splat, poison, zeroinitializer + %a = call @llvm.vp.load.nxv2i32.p0nxv2i32( * %p, %mask, i32 %vl) + %m = call @llvm.vp.icmp.nxv2i32( %x, %y, metadata !"eq", %mask, i32 %vl) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; FIXME: Merge vmerge.vvm and vleffN.v +declare { , i64 } @llvm.riscv.vleff.nxv2i32(, *, i64) +define @vpmerge_vleff( %passthru, * %p, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vleff: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a1, e32, m1, ta, mu +; CHECK-NEXT: vle32ff.v v9, (a0) +; CHECK-NEXT: vsetvli zero, a1, e32, m1, tu, mu +; CHECK-NEXT: vmerge.vvm v8, v8, v9, v0 +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call { , i64 } @llvm.riscv.vleff.nxv2i32( undef, * %p, i64 %1) + %b = extractvalue { , i64 } %a, 0 + %c = call @llvm.vp.merge.nxv2i32( %m, %b, %passthru, i32 %vl) + ret %c +} + +; Test strided load by riscv.vlse +declare @llvm.riscv.vlse.nxv2i32(, *, i64, i64) +define @vpmerge_vlse( %passthru, * %p, %m, i64 %s, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vlse: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a2, e32, m1, tu, mu +; CHECK-NEXT: vlse32.v v8, (a0), a1, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vlse.nxv2i32( undef, * %p, i64 %s, i64 %1) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test indexed load by riscv.vluxei +declare @llvm.riscv.vluxei.nxv2i32.nxv2i64(, *, , i64) +define @vpmerge_vluxei( %passthru, * %p, %idx, %m, i64 %s, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vluxei: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a2, e32, m1, tu, mu +; CHECK-NEXT: vluxei64.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vluxei.nxv2i32.nxv2i64( undef, * %p, %idx, i64 %1) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test vector index by riscv.vid +declare @llvm.riscv.vid.nxv2i32(, i64) +define @vpmerge_vid( %passthru, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vid: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vid.v v8, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vid.nxv2i32( undef, i64 %1) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.viota +declare @llvm.riscv.viota.nxv2i32(, , i64) +define @vpmerge_viota( %passthru, %m, %vm, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_viota: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: viota.m v8, v9, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.viota.nxv2i32( undef, %vm, i64 %1) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.vfclass +declare @llvm.riscv.vfclass.nxv2i32(, , i64) +define @vpmerge_vflcass( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vflcass: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfclass.v v8, v9, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfclass.nxv2i32( undef, %vf, i64 %1) + %b = call @llvm.vp.merge.nxv2i32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test riscv.vfsqrt +declare @llvm.riscv.vfsqrt.nxv2f32(, , i64) +define @vpmerge_vfsqrt( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vfsqrt: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfsqrt.v v8, v9, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfsqrt.nxv2f32( undef, %vf, i64 %1) + %b = call @llvm.vp.merge.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +} + +; Test reciprocal operation by riscv.vfrec7 +declare @llvm.riscv.vfrec7.nxv2f32(, , i64) +define @vpmerge_vfrec7( %passthru, %vf, %m, i32 zeroext %vl) { +; CHECK-LABEL: vpmerge_vfrec7: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli zero, a0, e32, m1, tu, mu +; CHECK-NEXT: vfrec7.v v8, v9, v0.t +; CHECK-NEXT: ret + %1 = zext i32 %vl to i64 + %a = call @llvm.riscv.vfrec7.nxv2f32( undef, %vf, i64 %1) + %b = call @llvm.vp.merge.nxv2f32( %m, %a, %passthru, i32 %vl) + ret %b +}