diff --git a/llvm/test/CodeGen/RISCV/rvv/vloxei-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vloxei-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vloxei-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vloxei-rv64.ll @@ -1339,3 +1339,42 @@ ret %a } + +define @intrinsic_vloxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64(* %0, %1, i64 %2) nounwind { +; CHECK-LABEL: intrinsic_vloxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v9, v8 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vloxei64.v v8, (a0), v9 +; CHECK-NEXT: ret +entry: + %i = zext %1 to + %a = call @llvm.riscv.vloxei.nxv1i64.nxv1i64( + undef, + * %0, + %i, + i64 %2) + + ret %a +} + +define @intrinsic_vloxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, %3, i64 %4) nounwind { +; CHECK-LABEL: intrinsic_vloxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, mu +; CHECK-NEXT: vloxei64.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + %a = call @llvm.riscv.vloxei.mask.nxv1i64.nxv1i64( + %0, + * %1, + %i, + %3, + i64 %4, i64 1) + + ret %a +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vloxei.ll b/llvm/test/CodeGen/RISCV/rvv/vloxei.ll --- a/llvm/test/CodeGen/RISCV/rvv/vloxei.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vloxei.ll @@ -5062,3 +5062,42 @@ ret %a } + +define @intrinsic_vloxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32(* %0, %1, iXLen %2) nounwind { +; CHECK-LABEL: intrinsic_vloxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v9, v8 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vloxei32.v v8, (a0), v9 +; CHECK-NEXT: ret +entry: + %i = zext %1 to + %a = call @llvm.riscv.vloxei.nxv1i64.nxv1i32( + undef, + * %0, + %i, + iXLen %2) + + ret %a +} + +define @intrinsic_vloxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, %3, iXLen %4) nounwind { +; CHECK-LABEL: intrinsic_vloxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, mu +; CHECK-NEXT: vloxei32.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + %a = call @llvm.riscv.vloxei.mask.nxv1i64.nxv1i32( + %0, + * %1, + %i, + %3, + iXLen %4, iXLen 1) + + ret %a +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vluxei-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vluxei-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vluxei-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vluxei-rv64.ll @@ -1339,3 +1339,43 @@ ret %a } + + +define @intrinsic_vluxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64(* %0, %1, i64 %2) nounwind { +; CHECK-LABEL: intrinsic_vluxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v9, v8 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vluxei64.v v8, (a0), v9 +; CHECK-NEXT: ret +entry: + %i = zext %1 to + %a = call @llvm.riscv.vluxei.nxv1i64.nxv1i64( + undef, + * %0, + %i, + i64 %2) + + ret %a +} + +define @intrinsic_vluxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, %3, i64 %4) nounwind { +; CHECK-LABEL: intrinsic_vluxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, mu +; CHECK-NEXT: vluxei64.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + %a = call @llvm.riscv.vluxei.mask.nxv1i64.nxv1i64( + %0, + * %1, + %i, + %3, + i64 %4, i64 1) + + ret %a +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vluxei.ll b/llvm/test/CodeGen/RISCV/rvv/vluxei.ll --- a/llvm/test/CodeGen/RISCV/rvv/vluxei.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vluxei.ll @@ -5062,3 +5062,42 @@ ret %a } + +define @intrinsic_vluxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32(* %0, %1, iXLen %2) nounwind { +; CHECK-LABEL: intrinsic_vluxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v9, v8 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vluxei32.v v8, (a0), v9 +; CHECK-NEXT: ret +entry: + %i = zext %1 to + %a = call @llvm.riscv.vluxei.nxv1i64.nxv1i32( + undef, + * %0, + %i, + iXLen %2) + + ret %a +} + +define @intrinsic_vluxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, %3, iXLen %4) nounwind { +; CHECK-LABEL: intrinsic_vluxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, mu +; CHECK-NEXT: vluxei32.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + %a = call @llvm.riscv.vluxei.mask.nxv1i64.nxv1i32( + %0, + * %1, + %i, + %3, + iXLen %4, iXLen 1) + + ret %a +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vsoxei-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vsoxei-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vsoxei-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vsoxei-rv64.ll @@ -1291,3 +1291,42 @@ ret void } + +define void @intrinsic_vsoxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vsoxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsoxei64.v v8, (a0), v10 +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsoxei.nxv1i64.nxv1i64( + %0, + * %1, + %i, + i64 %3) + + ret void +} + +define void @intrinsic_vsoxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, %3, i64 %4) nounwind { +; CHECK-LABEL: intrinsic_vsoxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsoxei64.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsoxei.mask.nxv1i64.nxv1i64( + %0, + * %1, + %i, + %3, + i64 %4) + + ret void +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vsoxei.ll b/llvm/test/CodeGen/RISCV/rvv/vsoxei.ll --- a/llvm/test/CodeGen/RISCV/rvv/vsoxei.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vsoxei.ll @@ -4878,3 +4878,42 @@ ret void } + +define void @intrinsic_vsoxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, iXLen %3) nounwind { +; CHECK-LABEL: intrinsic_vsoxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsoxei32.v v8, (a0), v10 +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsoxei.nxv1i64.nxv1i32( + %0, + * %1, + %i, + iXLen %3) + + ret void +} + +define void @intrinsic_vsoxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, %3, iXLen %4) nounwind { +; CHECK-LABEL: intrinsic_vsoxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsoxei32.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsoxei.mask.nxv1i64.nxv1i32( + %0, + * %1, + %i, + %3, + iXLen %4) + + ret void +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vsuxei-rv64.ll b/llvm/test/CodeGen/RISCV/rvv/vsuxei-rv64.ll --- a/llvm/test/CodeGen/RISCV/rvv/vsuxei-rv64.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vsuxei-rv64.ll @@ -1291,3 +1291,42 @@ ret void } + +define void @intrinsic_vsuxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, i64 %3) nounwind { +; CHECK-LABEL: intrinsic_vsuxei_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsuxei64.v v8, (a0), v10 +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsuxei.nxv1i64.nxv1i64( + %0, + * %1, + %i, + i64 %3) + + ret void +} + +define void @intrinsic_vsuxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64( %0, * %1, %2, %3, i64 %4) nounwind { +; CHECK-LABEL: intrinsic_vsuxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i64: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e64, m1, ta, ma +; CHECK-NEXT: vzext.vf8 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsuxei64.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsuxei.mask.nxv1i64.nxv1i64( + %0, + * %1, + %i, + %3, + i64 %4) + + ret void +} diff --git a/llvm/test/CodeGen/RISCV/rvv/vsuxei.ll b/llvm/test/CodeGen/RISCV/rvv/vsuxei.ll --- a/llvm/test/CodeGen/RISCV/rvv/vsuxei.ll +++ b/llvm/test/CodeGen/RISCV/rvv/vsuxei.ll @@ -4878,3 +4878,42 @@ ret void } + +define void @intrinsic_vsuxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, iXLen %3) nounwind { +; CHECK-LABEL: intrinsic_vsuxei_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsuxei32.v v8, (a0), v10 +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsuxei.nxv1i64.nxv1i32( + %0, + * %1, + %i, + iXLen %3) + + ret void +} + +define void @intrinsic_vsuxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32( %0, * %1, %2, %3, iXLen %4) nounwind { +; CHECK-LABEL: intrinsic_vsuxei_mask_v_zextidx_nxv1i64_nxv1i64_nxv1i32: +; CHECK: # %bb.0: # %entry +; CHECK-NEXT: vsetvli a2, zero, e32, mf2, ta, ma +; CHECK-NEXT: vzext.vf4 v10, v9 +; CHECK-NEXT: vsetvli zero, a1, e64, m1, ta, ma +; CHECK-NEXT: vsuxei32.v v8, (a0), v10, v0.t +; CHECK-NEXT: ret +entry: + %i = zext %2 to + call void @llvm.riscv.vsuxei.mask.nxv1i64.nxv1i32( + %0, + * %1, + %i, + %3, + iXLen %4) + + ret void +}