diff --git a/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp b/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp --- a/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp +++ b/llvm/lib/CodeGen/SelectionDAG/SelectionDAG.cpp @@ -3647,6 +3647,12 @@ })) return true; + // Is the operand of a splat vector a constant power of two? + if (Val.getOpcode() == ISD::SPLAT_VECTOR) + if (ConstantSDNode *C = dyn_cast(Val->getOperand(0))) + if (C->getAPIntValue().zextOrTrunc(BitWidth).isPowerOf2()) + return true; + // More could be done here, though the above checks are enough // to handle some common cases. diff --git a/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv32.ll b/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv32.ll --- a/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv32.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv32.ll @@ -736,6 +736,33 @@ ret %vc } +define @vdivu_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv1i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vadd.vi v25, v9, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v25 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv2i64: ; CHECK: # %bb.0: @@ -788,6 +815,33 @@ ret %vc } +define @vdivu_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv2i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vadd.vi v26, v10, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v26 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv4i64: ; CHECK: # %bb.0: @@ -840,6 +894,33 @@ ret %vc } +define @vdivu_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv4i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vadd.vi v28, v12, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v28 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv8i64: ; CHECK: # %bb.0: @@ -892,3 +973,29 @@ ret %vc } +define @vdivu_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv8i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vadd.vi v16, v16, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v16 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vdivu-sdnode-rv64.ll @@ -699,6 +699,33 @@ ret %vc } +define @vdivu_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv1i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vadd.vi v25, v9, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v25 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv2i64: ; CHECK: # %bb.0: @@ -738,6 +765,33 @@ ret %vc } +define @vdivu_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv2i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vadd.vi v26, v10, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v26 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv4i64: ; CHECK: # %bb.0: @@ -777,6 +831,33 @@ ret %vc } +define @vdivu_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv4i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vadd.vi v28, v12, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v28 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} + define @vdivu_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vdivu_vv_nxv8i64: ; CHECK: # %bb.0: @@ -816,3 +897,29 @@ ret %vc } +define @vdivu_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vdivu_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vsrl.vi v8, v8, 1 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = udiv %va, %splat + ret %vc +} + +; fold (udiv x, (shl c, y)) -> x >>u (log2(c)+y) if c is power of 2 +define @vdivu_vi_nxv8i64_2( %va, %vb) { +; CHECK-LABEL: vdivu_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vadd.vi v16, v16, 4 +; CHECK-NEXT: vsrl.vv v8, v8, v16 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = udiv %va, %vc + ret %vd +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv32.ll b/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv32.ll --- a/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv32.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv32.ll @@ -673,6 +673,30 @@ ret %vc } +define @vmul_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv1i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv2i64: ; CHECK: # %bb.0: @@ -715,6 +739,30 @@ ret %vc } +define @vmul_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv2i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv4i64: ; CHECK: # %bb.0: @@ -757,6 +805,30 @@ ret %vc } +define @vmul_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv4i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv8i64: ; CHECK: # %bb.0: @@ -799,3 +871,26 @@ ret %vc } +define @vmul_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv8i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vmul-sdnode-rv64.ll @@ -666,6 +666,30 @@ ret %vc } +define @vmul_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv1i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv2i64: ; CHECK: # %bb.0: @@ -701,6 +725,30 @@ ret %vc } +define @vmul_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv2i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv4i64: ; CHECK: # %bb.0: @@ -736,6 +784,30 @@ ret %vc } +define @vmul_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv4i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + define @vmul_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vmul_vv_nxv8i64: ; CHECK: # %bb.0: @@ -771,3 +843,26 @@ ret %vc } +define @vmul_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vmul_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vadd.vv v8, v8, v8 +; CHECK-NEXT: ret + %head = insertelement undef, i64 2, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} + +define @vmul_vi_nxv8i64_2( %va) { +; CHECK-LABEL: vmul_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vsll.vi v8, v8, 4 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = mul %va, %splat + ret %vc +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv32.ll b/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv32.ll --- a/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv32.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv32.ll @@ -750,6 +750,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv1i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m1, ta, mu +; CHECK-NEXT: vmv.v.x v25, a0 +; CHECK-NEXT: vsll.vv v25, v25, v9 +; CHECK-NEXT: vadd.vi v25, v25, -1 +; CHECK-NEXT: vand.vv v8, v8, v25 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv2i64: ; CHECK: # %bb.0: @@ -804,6 +835,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv2i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m2, ta, mu +; CHECK-NEXT: vmv.v.x v26, a0 +; CHECK-NEXT: vsll.vv v26, v26, v10 +; CHECK-NEXT: vadd.vi v26, v26, -1 +; CHECK-NEXT: vand.vv v8, v8, v26 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv4i64: ; CHECK: # %bb.0: @@ -858,6 +920,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +;fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv4i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m4, ta, mu +; CHECK-NEXT: vmv.v.x v28, a0 +; CHECK-NEXT: vsll.vv v28, v28, v12 +; CHECK-NEXT: vadd.vi v28, v28, -1 +; CHECK-NEXT: vand.vv v8, v8, v28 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv8i64: ; CHECK: # %bb.0: @@ -911,3 +1004,34 @@ %vc = urem %va, %splat ret %vc } + +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv8i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m8, ta, mu +; CHECK-NEXT: vmv.v.x v24, a0 +; CHECK-NEXT: vsll.vv v16, v24, v16 +; CHECK-NEXT: vadd.vi v16, v16, -1 +; CHECK-NEXT: vand.vv v8, v8, v16 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vremu-sdnode-rv64.ll @@ -737,6 +737,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv1i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv1i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m1, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv1i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv1i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m1, ta, mu +; CHECK-NEXT: vmv.v.x v25, a0 +; CHECK-NEXT: vsll.vv v25, v25, v9 +; CHECK-NEXT: vadd.vi v25, v25, -1 +; CHECK-NEXT: vand.vv v8, v8, v25 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv2i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv2i64: ; CHECK: # %bb.0: @@ -778,6 +809,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv2i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv2i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m2, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv2i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv2i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m2, ta, mu +; CHECK-NEXT: vmv.v.x v26, a0 +; CHECK-NEXT: vsll.vv v26, v26, v10 +; CHECK-NEXT: vadd.vi v26, v26, -1 +; CHECK-NEXT: vand.vv v8, v8, v26 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv4i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv4i64: ; CHECK: # %bb.0: @@ -819,6 +881,37 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv4i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv4i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m4, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +;fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv4i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv4i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m4, ta, mu +; CHECK-NEXT: vmv.v.x v28, a0 +; CHECK-NEXT: vsll.vv v28, v28, v12 +; CHECK-NEXT: vadd.vi v28, v28, -1 +; CHECK-NEXT: vand.vv v8, v8, v28 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +} + define @vremu_vv_nxv8i64( %va, %vb) { ; CHECK-LABEL: vremu_vv_nxv8i64: ; CHECK: # %bb.0: @@ -860,3 +953,33 @@ ret %vc } +; fold (urem x, pow2) -> (and x, pow2-1) +define @vremu_vi_nxv8i64_1( %va) { +; CHECK-LABEL: vremu_vi_nxv8i64_1: +; CHECK: # %bb.0: +; CHECK-NEXT: vsetvli a0, zero, e64, m8, ta, mu +; CHECK-NEXT: vand.vi v8, v8, 15 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = urem %va, %splat + ret %vc +} + +; fold (urem x, (shl pow2, y)) -> (and x, (add (shl pow2, y), -1)) +define @vremu_vi_nxv8i64_2( %va, %vb) { +; CHECK-LABEL: vremu_vi_nxv8i64_2: +; CHECK: # %bb.0: +; CHECK-NEXT: addi a0, zero, 16 +; CHECK-NEXT: vsetvli a1, zero, e64, m8, ta, mu +; CHECK-NEXT: vmv.v.x v24, a0 +; CHECK-NEXT: vsll.vv v16, v24, v16 +; CHECK-NEXT: vadd.vi v16, v16, -1 +; CHECK-NEXT: vand.vv v8, v8, v16 +; CHECK-NEXT: ret + %head = insertelement undef, i64 16, i32 0 + %splat = shufflevector %head, undef, zeroinitializer + %vc = shl %splat, %vb + %vd = urem %va, %vc + ret %vd +}