diff --git a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp --- a/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp +++ b/llvm/lib/Target/RISCV/RISCVISelDAGToDAG.cpp @@ -1515,6 +1515,7 @@ ReplaceNode(Node, Extract.getNode()); return; } + case ISD::SPLAT_VECTOR: case RISCVISD::VMV_V_X_VL: case RISCVISD::VFMV_V_F_VL: { // Try to match splat of a scalar load to a strided load with stride of x0. @@ -1531,7 +1532,10 @@ break; SDValue VL; - selectVLOp(Node->getOperand(1), VL); + if (Node->getOpcode() == ISD::SPLAT_VECTOR) + VL = CurDAG->getTargetConstant(RISCV::VLMaxSentinel, DL, XLenVT); + else + selectVLOp(Node->getOperand(1), VL); unsigned Log2SEW = Log2_32(VT.getScalarSizeInBits()); SDValue SEW = CurDAG->getTargetConstant(Log2SEW, DL, XLenVT); diff --git a/llvm/test/CodeGen/RISCV/rvv/vsplats-fp.ll b/llvm/test/CodeGen/RISCV/rvv/vsplats-fp.ll --- a/llvm/test/CodeGen/RISCV/rvv/vsplats-fp.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vsplats-fp.ll @@ -105,3 +105,22 @@ %splat = shufflevector %head, undef, zeroinitializer ret %splat } + +; Test that we fold this to a vlse with 0 stride. +define @vsplat_load_nxv8f32(float* %ptr) { +; RV32V-LABEL: vsplat_load_nxv8f32: +; RV32V: # %bb.0: +; RV32V-NEXT: vsetvli a1, zero, e32, m4, ta, mu +; RV32V-NEXT: vlse32.v v8, (a0), zero +; RV32V-NEXT: ret +; +; RV64V-LABEL: vsplat_load_nxv8f32: +; RV64V: # %bb.0: +; RV64V-NEXT: vsetvli a1, zero, e32, m4, ta, mu +; RV64V-NEXT: vlse32.v v8, (a0), zero +; RV64V-NEXT: ret + %f = load float, float* %ptr + %head = insertelement undef, float %f, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + ret %splat +}