Index: include/llvm/IR/IntrinsicsX86.td =================================================================== --- include/llvm/IR/IntrinsicsX86.td +++ include/llvm/IR/IntrinsicsX86.td @@ -2231,20 +2231,52 @@ let TargetPrefix = "x86" in { // All intrinsics start with "llvm.x86.". def int_x86_avx512_mask_vextractf32x4_512 : GCCBuiltin<"__builtin_ia32_extractf32x4_mask">, - Intrinsic<[llvm_v4f32_ty], [llvm_v16f32_ty, llvm_i8_ty, - llvm_v4f32_ty, llvm_i8_ty], [IntrNoMem]>; + Intrinsic<[llvm_v4f32_ty], [llvm_v16f32_ty, llvm_i32_ty, + llvm_v4f32_ty, llvm_i8_ty], [IntrNoMem]>; def int_x86_avx512_mask_vextracti32x4_512 : GCCBuiltin<"__builtin_ia32_extracti32x4_mask">, - Intrinsic<[llvm_v4i32_ty], [llvm_v16i32_ty, llvm_i8_ty, - llvm_v4i32_ty, llvm_i8_ty], [IntrNoMem]>; + Intrinsic<[llvm_v4i32_ty], [llvm_v16i32_ty, llvm_i32_ty, + llvm_v4i32_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextractf32x4_256 : + GCCBuiltin<"__builtin_ia32_extractf32x4_256_mask">, + Intrinsic<[llvm_v4f32_ty], [llvm_v8f32_ty, llvm_i32_ty, + llvm_v4f32_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextracti32x4_256 : + GCCBuiltin<"__builtin_ia32_extracti32x4_256_mask">, + Intrinsic<[llvm_v4i32_ty], [llvm_v8i32_ty, llvm_i32_ty, + llvm_v4i32_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextractf64x2_256 : + GCCBuiltin<"__builtin_ia32_extractf64x2_256_mask">, + Intrinsic<[llvm_v2f64_ty], [llvm_v4f64_ty, llvm_i32_ty, + llvm_v2f64_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextracti64x2_256 : + GCCBuiltin<"__builtin_ia32_extracti64x2_256_mask">, + Intrinsic<[llvm_v2i64_ty], [llvm_v4i64_ty, llvm_i32_ty, + llvm_v2i64_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextractf64x2_512 : + GCCBuiltin<"__builtin_ia32_extractf64x2_512_mask">, + Intrinsic<[llvm_v2f64_ty], [llvm_v8f64_ty, llvm_i32_ty, + llvm_v2f64_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextracti64x2_512 : + GCCBuiltin<"__builtin_ia32_extracti64x2_512_mask">, + Intrinsic<[llvm_v2i64_ty], [llvm_v8i64_ty, llvm_i32_ty, + llvm_v2i64_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextractf32x8_512 : + GCCBuiltin<"__builtin_ia32_extractf32x8_mask">, + Intrinsic<[llvm_v8f32_ty], [llvm_v16f32_ty, llvm_i32_ty, + llvm_v8f32_ty, llvm_i8_ty], [IntrNoMem]>; + def int_x86_avx512_mask_vextracti32x8_512 : + GCCBuiltin<"__builtin_ia32_extracti32x8_mask">, + Intrinsic<[llvm_v8i32_ty],[llvm_v16i32_ty, llvm_i32_ty, + llvm_v8i32_ty, llvm_i8_ty], [IntrNoMem]>; def int_x86_avx512_mask_vextractf64x4_512 : GCCBuiltin<"__builtin_ia32_extractf64x4_mask">, - Intrinsic<[llvm_v4f64_ty], [llvm_v8f64_ty, llvm_i8_ty, - llvm_v4f64_ty, llvm_i8_ty], [IntrNoMem]>; + Intrinsic<[llvm_v4f64_ty], [llvm_v8f64_ty, llvm_i32_ty, + llvm_v4f64_ty, llvm_i8_ty], [IntrNoMem]>; def int_x86_avx512_mask_vextracti64x4_512 : GCCBuiltin<"__builtin_ia32_extracti64x4_mask">, - Intrinsic<[llvm_v4i64_ty], [llvm_v8i64_ty, llvm_i8_ty, - llvm_v4i64_ty, llvm_i8_ty], [IntrNoMem]>; + Intrinsic<[llvm_v4i64_ty], [llvm_v8i64_ty, llvm_i32_ty, + llvm_v4i64_ty, llvm_i8_ty], [IntrNoMem]>; } // Conditional load ops Index: lib/Target/X86/X86InstrAVX512.td =================================================================== --- lib/Target/X86/X86InstrAVX512.td +++ lib/Target/X86/X86InstrAVX512.td @@ -566,85 +566,136 @@ // AVX-512 VECTOR EXTRACT //--- -multiclass vextract_for_size { +multiclass vextract_for_size_first_position_lowering { + // A subvector extract from the first vector position is + // a subregister copy that needs no instruction. + def NAME # To.NumElts: + Pat<(To.VT (extract_subvector (From.VT From.RC:$src),(iPTR 0))), + (To.VT (EXTRACT_SUBREG (From.VT From.RC:$src), To.SubRegIdx))>; +} + +multiclass vextract_for_size_no_alt : + vextract_for_size_first_position_lowering { + let hasSideEffects = 0, ExeDomain = To.ExeDomain in { defm rr : AVX512_maskable_in_asm, - AVX512AIi8Base, EVEX, EVEX_V512; - let mayStore = 1 in - def rm : AVX512AIi8, EVEX, EVEX_V512, EVEX_CD8; + AVX512AIi8Base, EVEX; + let mayStore = 1 in { + def rm : AVX512AIi8, EVEX; + + def rmk : AVX512AIi8, EVEX_K, EVEX; + }//mayStore = 1 } - // Codegen pattern with the alternative types, e.g. v8i64 -> v2i64 for - // vextracti32x4 - def : Pat<(vextract_extract:$ext (AltFrom.VT VR512:$src1), (iPTR imm)), - (AltTo.VT (!cast(NAME # To.EltSize # "x4rr") - VR512:$src1, - (EXTRACT_get_vextract_imm To.RC:$ext)))>; - - // A 128/256-bit subvector extract from the first 512-bit vector position is - // a subregister copy that needs no instruction. - def : Pat<(To.VT (extract_subvector (From.VT VR512:$src), (iPTR 0))), - (To.VT - (EXTRACT_SUBREG (From.VT VR512:$src), To.SubRegIdx))>; - - // And for the alternative types. - def : Pat<(AltTo.VT (extract_subvector (AltFrom.VT VR512:$src), (iPTR 0))), - (AltTo.VT - (EXTRACT_SUBREG (AltFrom.VT VR512:$src), AltTo.SubRegIdx))>; - // Intrinsic call with masking. def : Pat<(!cast("int_x86_avx512_mask_vextract" # To.EltTypeName # - "x4_512") - VR512:$src1, (iPTR imm:$idx), To.RC:$src0, GR8:$mask), - (!cast(NAME # To.EltSize # "x4rrk") To.RC:$src0, - (v4i1 (COPY_TO_REGCLASS GR8:$mask, VK4WM)), - VR512:$src1, imm:$idx)>; + "x" # To.NumElts # "_" # From.Size) + From.RC:$src1, (iPTR imm:$idx), To.RC:$src0, To.MRC:$mask), + (!cast(NAME # To.EltSize # "x" # To.NumElts # + From.ZSuffix # "rrk") + To.RC:$src0, + (COPY_TO_REGCLASS To.MRC:$mask, To.KRCWM), + From.RC:$src1, imm:$idx)>; // Intrinsic call with zero-masking. def : Pat<(!cast("int_x86_avx512_mask_vextract" # To.EltTypeName # - "x4_512") - VR512:$src1, (iPTR imm:$idx), To.ImmAllZerosV, GR8:$mask), - (!cast(NAME # To.EltSize # "x4rrkz") - (v4i1 (COPY_TO_REGCLASS GR8:$mask, VK4WM)), - VR512:$src1, imm:$idx)>; + "x" # To.NumElts # "_" # From.Size) + From.RC:$src1, (iPTR imm:$idx), To.ImmAllZerosV, To.MRC:$mask), + (!cast(NAME # To.EltSize # "x" # To.NumElts # + From.ZSuffix # "rrkz") + (COPY_TO_REGCLASS To.MRC:$mask, To.KRCWM), + From.RC:$src1, imm:$idx)>; // Intrinsic call without masking. def : Pat<(!cast("int_x86_avx512_mask_vextract" # To.EltTypeName # - "x4_512") - VR512:$src1, (iPTR imm:$idx), To.ImmAllZerosV, (i8 -1)), - (!cast(NAME # To.EltSize # "x4rr") - VR512:$src1, imm:$idx)>; + "x" # To.NumElts # "_" # From.Size) + From.RC:$src1, (iPTR imm:$idx), To.ImmAllZerosV, (i8 -1)), + (!cast(NAME # To.EltSize # "x" # To.NumElts # + From.ZSuffix # "rr") + From.RC:$src1, imm:$idx)>; } -multiclass vextract_for_type { - defm NAME # "32x4" : vextract_for_size : + vextract_for_size_no_alt, + vextract_for_size_first_position_lowering { + + // Codegen pattern with the alternative types. + // Only add this if operation not supported natively via AVX512DQ + let Predicates = [NoDQI] in + def : Pat<(vextract_extract:$ext (AltFrom.VT AltFrom.RC:$src1), (iPTR imm)), + (AltTo.VT (!cast(NAME # To.EltSize # "x" # + To.NumElts # From.ZSuffix # "rr") + AltFrom.RC:$src1, + (EXTRACT_get_vextract_imm To.RC:$ext)))>; +} + +multiclass vextract_for_type { + defm NAME # "32x4Z" : vextract_for_size, X86VectorVTInfo< 4, EltVT32, VR128X>, X86VectorVTInfo< 8, EltVT64, VR512>, X86VectorVTInfo< 2, EltVT64, VR128X>, vextract128_extract, - EXTRACT_get_vextract128_imm>; - defm NAME # "64x4" : vextract_for_size, + EVEX_V512, EVEX_CD8<32, CD8VT4>; + defm NAME # "64x4Z" : vextract_for_size, X86VectorVTInfo< 4, EltVT64, VR256X>, X86VectorVTInfo<16, EltVT32, VR512>, X86VectorVTInfo< 8, EltVT32, VR256>, vextract256_extract, - EXTRACT_get_vextract256_imm>, VEX_W; + EXTRACT_get_vextract256_imm>, + VEX_W, EVEX_V512, EVEX_CD8<64, CD8VT4>; + let Predicates = [HasVLX] in + defm NAME # "32x4Z256" : vextract_for_size, + X86VectorVTInfo< 4, EltVT32, VR128X>, + X86VectorVTInfo< 4, EltVT64, VR256X>, + X86VectorVTInfo< 2, EltVT64, VR128X>, + vextract128_extract, + EXTRACT_get_vextract128_imm>, + EVEX_V256, EVEX_CD8<32, CD8VT4>; + let Predicates = [HasVLX, HasDQI] in + defm NAME # "64x2Z256" : vextract_for_size_no_alt, + X86VectorVTInfo< 2, EltVT64, VR128X>, + vextract128_extract>, + VEX_W, EVEX_V256, EVEX_CD8<64, CD8VT2>; + let Predicates = [HasDQI] in { + defm NAME # "64x2Z" : vextract_for_size_no_alt, + X86VectorVTInfo< 2, EltVT64, VR128X>, + vextract128_extract>, + VEX_W, EVEX_V512, EVEX_CD8<64, CD8VT2>; + defm NAME # "32x8Z" : vextract_for_size_no_alt, + X86VectorVTInfo< 8, EltVT32, VR256X>, + vextract256_extract>, + EVEX_V512, EVEX_CD8<32, CD8VT8>; + } } defm VEXTRACTF : vextract_for_type; Index: test/CodeGen/X86/avx512-cvt.ll =================================================================== --- test/CodeGen/X86/avx512-cvt.ll +++ test/CodeGen/X86/avx512-cvt.ll @@ -52,14 +52,14 @@ } ; CHECK-LABEL: sltof432 -; CHECK: vcvtqq2ps +; CHECK: vcvtqq2ps define <4 x float> @sltof432(<4 x i64> %a) { %b = sitofp <4 x i64> %a to <4 x float> ret <4 x float> %b } ; CHECK-LABEL: ultof432 -; CHECK: vcvtuqq2ps +; CHECK: vcvtuqq2ps define <4 x float> @ultof432(<4 x i64> %a) { %b = uitofp <4 x i64> %a to <4 x float> ret <4 x float> %b @@ -279,12 +279,14 @@ ret i32 %res } -; CHECK-LABEL: uitof64 -; CHECK: vcvtudq2pd -; CHECK: vextracti64x4 -; CHECK: vcvtudq2pd -; CHECK: ret define <16 x double> @uitof64(<16 x i32> %a) nounwind { +; CHECK-LABEL: uitof64: +; CHECK: ## BB#0: +; CHECK-NEXT: vcvtudq2pd %ymm0, %zmm2 +; CHECK-NEXT: vextracti32x8 $1, %zmm0, %ymm0 +; CHECK-NEXT: vcvtudq2pd %ymm0, %zmm1 +; CHECK-NEXT: vmovaps %zmm2, %zmm0 +; CHECK-NEXT: retq %b = uitofp <16 x i32> %a to <16 x double> ret <16 x double> %b } @@ -407,7 +409,7 @@ } ; CHECK-LABEL: @uitofp_16i8 -; CHECK: vpmovzxbd +; CHECK: vpmovzxbd ; CHECK: vcvtudq2ps define <16 x float> @uitofp_16i8(<16 x i8>%a) { %b = uitofp <16 x i8> %a to <16 x float> Index: test/CodeGen/X86/avx512-insert-extract.ll =================================================================== --- test/CodeGen/X86/avx512-insert-extract.ll +++ test/CodeGen/X86/avx512-insert-extract.ll @@ -12,14 +12,24 @@ ret <16 x float> %rrr3 } -;CHECK-LABEL: test2: -;KNL: vinsertf32x4 $0 -;SKX: vinsertf64x2 $0 -;CHECK: vextractf32x4 $3 -;KNL: vinsertf32x4 $3 -;SKX: vinsertf64x2 $3 -;CHECK: ret define <8 x double> @test2(<8 x double> %x, double* %br, double %y) nounwind { +; KNL-LABEL: test2: +; KNL: ## BB#0: +; KNL-NEXT: vmovhpd (%rdi), %xmm0, %xmm2 +; KNL-NEXT: vinsertf32x4 $0, %xmm2, %zmm0, %zmm0 +; KNL-NEXT: vextractf32x4 $3, %zmm0, %xmm2 +; KNL-NEXT: vmovsd %xmm1, %xmm2, %xmm1 +; KNL-NEXT: vinsertf32x4 $3, %xmm1, %zmm0, %zmm0 +; KNL-NEXT: retq +; +; SKX-LABEL: test2: +; SKX: ## BB#0: +; SKX-NEXT: vmovhpd (%rdi), %xmm0, %xmm2 +; SKX-NEXT: vinsertf64x2 $0, %xmm2, %zmm0, %zmm0 +; SKX-NEXT: vextractf64x2 $3, %zmm0, %xmm2 +; SKX-NEXT: vmovsd %xmm1, %xmm2, %xmm1 +; SKX-NEXT: vinsertf64x2 $3, %xmm1, %zmm0, %zmm0 +; SKX-NEXT: retq %rrr = load double, double* %br %rrr2 = insertelement <8 x double> %x, double %rrr, i32 1 %rrr3 = insertelement <8 x double> %rrr2, double %y, i32 6 @@ -36,12 +46,22 @@ ret <16 x float> %rrr2 } -;CHECK-LABEL: test4: -;CHECK: vextracti32x4 $2 -;KNL: vinserti32x4 $0 -;SKX: vinserti64x2 $0 -;CHECK: ret define <8 x i64> @test4(<8 x i64> %x) nounwind { +; KNL-LABEL: test4: +; KNL: ## BB#0: +; KNL-NEXT: vextracti32x4 $2, %zmm0, %xmm1 +; KNL-NEXT: vmovq %xmm1, %rax +; KNL-NEXT: vpinsrq $1, %rax, %xmm0, %xmm1 +; KNL-NEXT: vinserti32x4 $0, %xmm1, %zmm0, %zmm0 +; KNL-NEXT: retq +; +; SKX-LABEL: test4: +; SKX: ## BB#0: +; SKX-NEXT: vextracti64x2 $2, %zmm0, %xmm1 +; SKX-NEXT: vmovq %xmm1, %rax +; SKX-NEXT: vpinsrq $1, %rax, %xmm0, %xmm1 +; SKX-NEXT: vinserti64x2 $0, %xmm1, %zmm0, %zmm0 +; SKX-NEXT: retq %eee = extractelement <8 x i64> %x, i32 4 %rrr2 = insertelement <8 x i64> %x, i64 %eee, i32 1 ret <8 x i64> %rrr2 @@ -142,7 +162,7 @@ ;CHECK: andl $1, %eax ;CHECK: kmovw %eax, %k0 ;CHECK: movw $-4 -;CHECK: korw +;CHECK: korw define i16 @test13(i32 %a, i32 %b) { %cmp_res = icmp ult i32 %a, %b %maskv = insertelement <16 x i1> , i1 %cmp_res, i32 0 Index: test/CodeGen/X86/avx512-intrinsics.ll =================================================================== --- test/CodeGen/X86/avx512-intrinsics.ll +++ test/CodeGen/X86/avx512-intrinsics.ll @@ -904,38 +904,38 @@ define <4 x float> @test_mask_vextractf32x4(<4 x float> %b, <16 x float> %a, i8 %mask) { ; CHECK-LABEL: test_mask_vextractf32x4: ; CHECK: vextractf32x4 $2, %zmm1, %xmm0 {%k1} - %res = call <4 x float> @llvm.x86.avx512.mask.vextractf32x4.512(<16 x float> %a, i8 2, <4 x float> %b, i8 %mask) + %res = call <4 x float> @llvm.x86.avx512.mask.vextractf32x4.512(<16 x float> %a, i32 2, <4 x float> %b, i8 %mask) ret <4 x float> %res } -declare <4 x float> @llvm.x86.avx512.mask.vextractf32x4.512(<16 x float>, i8, <4 x float>, i8) +declare <4 x float> @llvm.x86.avx512.mask.vextractf32x4.512(<16 x float>, i32, <4 x float>, i8) define <4 x i64> @test_mask_vextracti64x4(<4 x i64> %b, <8 x i64> %a, i8 %mask) { ; CHECK-LABEL: test_mask_vextracti64x4: ; CHECK: vextracti64x4 $2, %zmm1, %ymm0 {%k1} - %res = call <4 x i64> @llvm.x86.avx512.mask.vextracti64x4.512(<8 x i64> %a, i8 2, <4 x i64> %b, i8 %mask) + %res = call <4 x i64> @llvm.x86.avx512.mask.vextracti64x4.512(<8 x i64> %a, i32 2, <4 x i64> %b, i8 %mask) ret <4 x i64> %res } -declare <4 x i64> @llvm.x86.avx512.mask.vextracti64x4.512(<8 x i64>, i8, <4 x i64>, i8) +declare <4 x i64> @llvm.x86.avx512.mask.vextracti64x4.512(<8 x i64>, i32, <4 x i64>, i8) define <4 x i32> @test_maskz_vextracti32x4(<16 x i32> %a, i8 %mask) { ; CHECK-LABEL: test_maskz_vextracti32x4: ; CHECK: vextracti32x4 $2, %zmm0, %xmm0 {%k1} {z} - %res = call <4 x i32> @llvm.x86.avx512.mask.vextracti32x4.512(<16 x i32> %a, i8 2, <4 x i32> zeroinitializer, i8 %mask) + %res = call <4 x i32> @llvm.x86.avx512.mask.vextracti32x4.512(<16 x i32> %a, i32 2, <4 x i32> zeroinitializer, i8 %mask) ret <4 x i32> %res } -declare <4 x i32> @llvm.x86.avx512.mask.vextracti32x4.512(<16 x i32>, i8, <4 x i32>, i8) +declare <4 x i32> @llvm.x86.avx512.mask.vextracti32x4.512(<16 x i32>, i32, <4 x i32>, i8) define <4 x double> @test_vextractf64x4(<8 x double> %a) { ; CHECK-LABEL: test_vextractf64x4: ; CHECK: vextractf64x4 $2, %zmm0, %ymm0 ## - %res = call <4 x double> @llvm.x86.avx512.mask.vextractf64x4.512(<8 x double> %a, i8 2, <4 x double> zeroinitializer, i8 -1) + %res = call <4 x double> @llvm.x86.avx512.mask.vextractf64x4.512(<8 x double> %a, i32 2, <4 x double> zeroinitializer, i8 -1) ret <4 x double> %res } -declare <4 x double> @llvm.x86.avx512.mask.vextractf64x4.512(<8 x double>, i8, <4 x double>, i8) +declare <4 x double> @llvm.x86.avx512.mask.vextractf64x4.512(<8 x double>, i32, <4 x double>, i8) define <16 x i32> @test_x86_avx512_pslli_d(<16 x i32> %a0) { ; CHECK-LABEL: test_x86_avx512_pslli_d Index: test/CodeGen/X86/avx512dq-intrinsics.ll =================================================================== --- test/CodeGen/X86/avx512dq-intrinsics.ll +++ test/CodeGen/X86/avx512dq-intrinsics.ll @@ -315,3 +315,44 @@ %res2 = fadd <2 x double> %res, %res1 ret <2 x double> %res2 } + + +declare <2 x double> @llvm.x86.avx512.mask.vextractf64x2.512(<8 x double>, i32, <2 x double>, i8) + +define <2 x double>@test_int_x86_avx512_mask_vextractf64x2_512(<8 x double> %x0, <2 x double> %x2, i8 %x3) { +; CHECK-LABEL: test_int_x86_avx512_mask_vextractf64x2_512: +; CHECK: ## BB#0: +; CHECK-NEXT: kmovw %edi, %k1 +; CHECK-NEXT: vextractf64x2 $1, %zmm0, %xmm1 {%k1} +; CHECK-NEXT: vextractf64x2 $1, %zmm0, %xmm2 {%k1} {z} +; CHECK-NEXT: vextractf64x2 $1, %zmm0, %xmm0 +; CHECK-NEXT: vaddpd %xmm0, %xmm1, %xmm0 +; CHECK-NEXT: vaddpd %xmm0, %xmm2, %xmm0 +; CHECK-NEXT: retq + %res = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.512(<8 x double> %x0,i32 1, <2 x double> %x2, i8 %x3) + %res2 = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.512(<8 x double> %x0,i32 1, <2 x double> zeroinitializer, i8 %x3) + %res1 = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.512(<8 x double> %x0,i32 1, <2 x double> zeroinitializer, i8 -1) + %res3 = fadd <2 x double> %res, %res1 + %res4 = fadd <2 x double> %res2, %res3 + ret <2 x double> %res4 +} + +declare <8 x float> @llvm.x86.avx512.mask.vextractf32x8.512(<16 x float>, i32, <8 x float>, i8) + +define <8 x float>@test_int_x86_avx512_mask_vextractf32x8(<16 x float> %x0, <8 x float> %x2, i8 %x3) { +; CHECK-LABEL: test_int_x86_avx512_mask_vextractf32x8: +; CHECK: ## BB#0: +; CHECK-NEXT: kmovw %edi, %k1 +; CHECK-NEXT: vextractf32x8 $1, %zmm0, %ymm1 {%k1} +; CHECK-NEXT: vextractf32x8 $1, %zmm0, %ymm2 {%k1} {z} +; CHECK-NEXT: vextractf32x8 $1, %zmm0, %ymm0 +; CHECK-NEXT: vaddps %ymm0, %ymm1, %ymm0 +; CHECK-NEXT: vaddps %ymm0, %ymm2, %ymm0 +; CHECK-NEXT: retq + %res = call <8 x float> @llvm.x86.avx512.mask.vextractf32x8.512(<16 x float> %x0,i32 1, <8 x float> %x2, i8 %x3) + %res2 = call <8 x float> @llvm.x86.avx512.mask.vextractf32x8.512(<16 x float> %x0,i32 1, <8 x float> zeroinitializer, i8 %x3) + %res1 = call <8 x float> @llvm.x86.avx512.mask.vextractf32x8.512(<16 x float> %x0,i32 1, <8 x float> zeroinitializer, i8 -1) + %res3 = fadd <8 x float> %res, %res1 + %res4 = fadd <8 x float> %res2, %res3 + ret <8 x float> %res4 +} Index: test/CodeGen/X86/avx512dqvl-intrinsics.ll =================================================================== --- test/CodeGen/X86/avx512dqvl-intrinsics.ll +++ test/CodeGen/X86/avx512dqvl-intrinsics.ll @@ -1648,3 +1648,23 @@ %res2 = fadd <8 x float> %res, %res1 ret <8 x float> %res2 } + +declare <2 x double> @llvm.x86.avx512.mask.vextractf64x2.256(<4 x double>, i32, <2 x double>, i8) + +define <2 x double>@test_int_x86_avx512_mask_vextractf64x2_256(<4 x double> %x0, <2 x double> %x2, i8 %x3) { +; CHECK-LABEL: test_int_x86_avx512_mask_vextractf64x2_256: +; CHECK: ## BB#0: +; CHECK-NEXT: kmovw %edi, %k1 +; CHECK-NEXT: vextractf64x2 $1, %ymm0, %xmm1 {%k1} +; CHECK-NEXT: vextractf64x2 $1, %ymm0, %xmm2 {%k1} {z} +; CHECK-NEXT: vextractf64x2 $1, %ymm0, %xmm0 +; CHECK-NEXT: vaddpd %xmm0, %xmm1, %xmm0 +; CHECK-NEXT: vaddpd %xmm2, %xmm0, %xmm0 +; CHECK-NEXT: retq + %res = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.256(<4 x double> %x0,i32 1, <2 x double> %x2, i8 %x3) + %res2 = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.256(<4 x double> %x0,i32 1, <2 x double> zeroinitializer, i8 %x3) + %res1 = call <2 x double> @llvm.x86.avx512.mask.vextractf64x2.256(<4 x double> %x0,i32 1, <2 x double> zeroinitializer, i8 -1) + %res3 = fadd <2 x double> %res, %res1 + %res4 = fadd <2 x double> %res3, %res2 + ret <2 x double> %res4 +} Index: test/CodeGen/X86/avx512vl-intrinsics.ll =================================================================== --- test/CodeGen/X86/avx512vl-intrinsics.ll +++ test/CodeGen/X86/avx512vl-intrinsics.ll @@ -4507,3 +4507,23 @@ %res2 = fadd <8 x float> %res, %res1 ret <8 x float> %res2 } + +declare <4 x float> @llvm.x86.avx512.mask.vextractf32x4.256(<8 x float>, i32, <4 x float>, i8) + +define <4 x float>@test_int_x86_avx512_mask_vextractf32x4_256(<8 x float> %x0, <4 x float> %x2, i8 %x3) { +; CHECK-LABEL: test_int_x86_avx512_mask_vextractf32x4_256: +; CHECK: ## BB#0: +; CHECK-NEXT: kmovw %edi, %k1 +; CHECK-NEXT: vextractf32x4 $1, %ymm0, %xmm1 {%k1} +; CHECK-NEXT: vextractf32x4 $1, %ymm0, %xmm2 {%k1} {z} +; CHECK-NEXT: vextractf32x4 $1, %ymm0, %xmm0 +; CHECK-NEXT: vaddps %xmm2, %xmm1, %xmm1 +; CHECK-NEXT: vaddps %xmm1, %xmm0, %xmm0 +; CHECK-NEXT: retq + %res = call <4 x float> @llvm.x86.avx512.mask.vextractf32x4.256(<8 x float> %x0, i32 1, <4 x float> %x2, i8 %x3) + %res1 = call <4 x float> @llvm.x86.avx512.mask.vextractf32x4.256(<8 x float> %x0, i32 1, <4 x float> zeroinitializer, i8 %x3) + %res2 = call <4 x float> @llvm.x86.avx512.mask.vextractf32x4.256(<8 x float> %x0, i32 1, <4 x float> zeroinitializer, i8 -1) + %res3 = fadd <4 x float> %res, %res1 + %res4 = fadd <4 x float> %res2, %res3 + ret <4 x float> %res4 +} Index: test/MC/X86/avx512-encodings.s =================================================================== --- test/MC/X86/avx512-encodings.s +++ test/MC/X86/avx512-encodings.s @@ -14958,3 +14958,195 @@ // CHECK: encoding: [0x62,0xf2,0xc5,0x08,0x43,0x92,0xf8,0xfb,0xff,0xff] vgetexpsd -1032(%rdx), %xmm7, %xmm2 +// CHECK: vextractf32x4 $171, %zmm21, %xmm15 +// CHECK: encoding: [0x62,0xc3,0x7d,0x48,0x19,0xef,0xab] + vextractf32x4 $0xab, %zmm21, %xmm15 + +// CHECK: vextractf32x4 $171, %zmm21, %xmm15 {%k1} +// CHECK: encoding: [0x62,0xc3,0x7d,0x49,0x19,0xef,0xab] + vextractf32x4 $0xab, %zmm21, %xmm15 {%k1} + +// CHECK: vextractf32x4 $171, %zmm21, %xmm15 {%k1} {z} +// CHECK: encoding: [0x62,0xc3,0x7d,0xc9,0x19,0xef,0xab] + vextractf32x4 $0xab, %zmm21, %xmm15 {%k1} {z} + +// CHECK: vextractf32x4 $123, %zmm21, %xmm15 +// CHECK: encoding: [0x62,0xc3,0x7d,0x48,0x19,0xef,0x7b] + vextractf32x4 $0x7b, %zmm21, %xmm15 + +// CHECK: vextractf32x4 $171, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0x21,0xab] + vextractf32x4 $0xab, %zmm20, (%rcx) + +// CHECK: vextractf32x4 $171, %zmm20, (%rcx) {%k7} +// CHECK: encoding: [0x62,0xe3,0x7d,0x4f,0x19,0x21,0xab] + vextractf32x4 $0xab, %zmm20, (%rcx) {%k7} + +// CHECK: vextractf32x4 $123, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0x21,0x7b] + vextractf32x4 $0x7b, %zmm20, (%rcx) + +// CHECK: vextractf32x4 $123, %zmm20, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x19,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf32x4 $0x7b, %zmm20, 291(%rax,%r14,8) + +// CHECK: vextractf32x4 $123, %zmm20, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0x62,0x7f,0x7b] + vextractf32x4 $0x7b, %zmm20, 2032(%rdx) + +// CHECK: vextractf32x4 $123, %zmm20, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextractf32x4 $0x7b, %zmm20, 2048(%rdx) + +// CHECK: vextractf32x4 $123, %zmm20, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0x62,0x80,0x7b] + vextractf32x4 $0x7b, %zmm20, -2048(%rdx) + +// CHECK: vextractf32x4 $123, %zmm20, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x19,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextractf32x4 $0x7b, %zmm20, -2064(%rdx) + +// CHECK: vextractf64x4 $171, %zmm24, %ymm11 +// CHECK: encoding: [0x62,0x43,0xfd,0x48,0x1b,0xc3,0xab] + vextractf64x4 $0xab, %zmm24, %ymm11 + +// CHECK: vextractf64x4 $171, %zmm24, %ymm11 {%k5} +// CHECK: encoding: [0x62,0x43,0xfd,0x4d,0x1b,0xc3,0xab] + vextractf64x4 $0xab, %zmm24, %ymm11 {%k5} + +// CHECK: vextractf64x4 $171, %zmm24, %ymm11 {%k5} {z} +// CHECK: encoding: [0x62,0x43,0xfd,0xcd,0x1b,0xc3,0xab] + vextractf64x4 $0xab, %zmm24, %ymm11 {%k5} {z} + +// CHECK: vextractf64x4 $123, %zmm24, %ymm11 +// CHECK: encoding: [0x62,0x43,0xfd,0x48,0x1b,0xc3,0x7b] + vextractf64x4 $0x7b, %zmm24, %ymm11 + +// CHECK: vextractf64x4 $171, %zmm5, (%rcx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0x29,0xab] + vextractf64x4 $0xab, %zmm5, (%rcx) + +// CHECK: vextractf64x4 $171, %zmm5, (%rcx) {%k4} +// CHECK: encoding: [0x62,0xf3,0xfd,0x4c,0x1b,0x29,0xab] + vextractf64x4 $0xab, %zmm5, (%rcx) {%k4} + +// CHECK: vextractf64x4 $123, %zmm5, (%rcx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0x29,0x7b] + vextractf64x4 $0x7b, %zmm5, (%rcx) + +// CHECK: vextractf64x4 $123, %zmm5, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xb3,0xfd,0x48,0x1b,0xac,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf64x4 $0x7b, %zmm5, 291(%rax,%r14,8) + +// CHECK: vextractf64x4 $123, %zmm5, 4064(%rdx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0x6a,0x7f,0x7b] + vextractf64x4 $0x7b, %zmm5, 4064(%rdx) + +// CHECK: vextractf64x4 $123, %zmm5, 4096(%rdx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0xaa,0x00,0x10,0x00,0x00,0x7b] + vextractf64x4 $0x7b, %zmm5, 4096(%rdx) + +// CHECK: vextractf64x4 $123, %zmm5, -4096(%rdx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0x6a,0x80,0x7b] + vextractf64x4 $0x7b, %zmm5, -4096(%rdx) + +// CHECK: vextractf64x4 $123, %zmm5, -4128(%rdx) +// CHECK: encoding: [0x62,0xf3,0xfd,0x48,0x1b,0xaa,0xe0,0xef,0xff,0xff,0x7b] + vextractf64x4 $0x7b, %zmm5, -4128(%rdx) + +// CHECK: vextracti32x4 $171, %zmm16, %xmm13 +// CHECK: encoding: [0x62,0xc3,0x7d,0x48,0x39,0xc5,0xab] + vextracti32x4 $0xab, %zmm16, %xmm13 + +// CHECK: vextracti32x4 $171, %zmm16, %xmm13 {%k5} +// CHECK: encoding: [0x62,0xc3,0x7d,0x4d,0x39,0xc5,0xab] + vextracti32x4 $0xab, %zmm16, %xmm13 {%k5} + +// CHECK: vextracti32x4 $171, %zmm16, %xmm13 {%k5} {z} +// CHECK: encoding: [0x62,0xc3,0x7d,0xcd,0x39,0xc5,0xab] + vextracti32x4 $0xab, %zmm16, %xmm13 {%k5} {z} + +// CHECK: vextracti32x4 $123, %zmm16, %xmm13 +// CHECK: encoding: [0x62,0xc3,0x7d,0x48,0x39,0xc5,0x7b] + vextracti32x4 $0x7b, %zmm16, %xmm13 + +// CHECK: vextracti32x4 $171, %zmm29, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0x29,0xab] + vextracti32x4 $0xab, %zmm29, (%rcx) + +// CHECK: vextracti32x4 $171, %zmm29, (%rcx) {%k2} +// CHECK: encoding: [0x62,0x63,0x7d,0x4a,0x39,0x29,0xab] + vextracti32x4 $0xab, %zmm29, (%rcx) {%k2} + +// CHECK: vextracti32x4 $123, %zmm29, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0x29,0x7b] + vextracti32x4 $0x7b, %zmm29, (%rcx) + +// CHECK: vextracti32x4 $123, %zmm29, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0x7d,0x48,0x39,0xac,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti32x4 $0x7b, %zmm29, 291(%rax,%r14,8) + +// CHECK: vextracti32x4 $123, %zmm29, 2032(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0x6a,0x7f,0x7b] + vextracti32x4 $0x7b, %zmm29, 2032(%rdx) + +// CHECK: vextracti32x4 $123, %zmm29, 2048(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0xaa,0x00,0x08,0x00,0x00,0x7b] + vextracti32x4 $0x7b, %zmm29, 2048(%rdx) + +// CHECK: vextracti32x4 $123, %zmm29, -2048(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0x6a,0x80,0x7b] + vextracti32x4 $0x7b, %zmm29, -2048(%rdx) + +// CHECK: vextracti32x4 $123, %zmm29, -2064(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x39,0xaa,0xf0,0xf7,0xff,0xff,0x7b] + vextracti32x4 $0x7b, %zmm29, -2064(%rdx) + +// CHECK: vextracti64x4 $171, %zmm16, %ymm13 +// CHECK: encoding: [0x62,0xc3,0xfd,0x48,0x3b,0xc5,0xab] + vextracti64x4 $0xab, %zmm16, %ymm13 + +// CHECK: vextracti64x4 $171, %zmm16, %ymm13 {%k3} +// CHECK: encoding: [0x62,0xc3,0xfd,0x4b,0x3b,0xc5,0xab] + vextracti64x4 $0xab, %zmm16, %ymm13 {%k3} + +// CHECK: vextracti64x4 $171, %zmm16, %ymm13 {%k3} {z} +// CHECK: encoding: [0x62,0xc3,0xfd,0xcb,0x3b,0xc5,0xab] + vextracti64x4 $0xab, %zmm16, %ymm13 {%k3} {z} + +// CHECK: vextracti64x4 $123, %zmm16, %ymm13 +// CHECK: encoding: [0x62,0xc3,0xfd,0x48,0x3b,0xc5,0x7b] + vextracti64x4 $0x7b, %zmm16, %ymm13 + +// CHECK: vextracti64x4 $171, %zmm30, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0x31,0xab] + vextracti64x4 $0xab, %zmm30, (%rcx) + +// CHECK: vextracti64x4 $171, %zmm30, (%rcx) {%k4} +// CHECK: encoding: [0x62,0x63,0xfd,0x4c,0x3b,0x31,0xab] + vextracti64x4 $0xab, %zmm30, (%rcx) {%k4} + +// CHECK: vextracti64x4 $123, %zmm30, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0x31,0x7b] + vextracti64x4 $0x7b, %zmm30, (%rcx) + +// CHECK: vextracti64x4 $123, %zmm30, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0xfd,0x48,0x3b,0xb4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti64x4 $0x7b, %zmm30, 291(%rax,%r14,8) + +// CHECK: vextracti64x4 $123, %zmm30, 4064(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0x72,0x7f,0x7b] + vextracti64x4 $0x7b, %zmm30, 4064(%rdx) + +// CHECK: vextracti64x4 $123, %zmm30, 4096(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0xb2,0x00,0x10,0x00,0x00,0x7b] + vextracti64x4 $0x7b, %zmm30, 4096(%rdx) + +// CHECK: vextracti64x4 $123, %zmm30, -4096(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0x72,0x80,0x7b] + vextracti64x4 $0x7b, %zmm30, -4096(%rdx) + +// CHECK: vextracti64x4 $123, %zmm30, -4128(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x3b,0xb2,0xe0,0xef,0xff,0xff,0x7b] + vextracti64x4 $0x7b, %zmm30, -4128(%rdx) + Index: test/MC/X86/x86-64-avx512dq.s =================================================================== --- test/MC/X86/x86-64-avx512dq.s +++ test/MC/X86/x86-64-avx512dq.s @@ -2371,3 +2371,387 @@ // CHECK: encoding: [0x62,0xa1,0xff,0xca,0x7a,0xd5] vcvtuqq2ps %zmm21, %ymm18 {%k2} {z} +// CHECK: vextractf32x8 $171, %zmm18, %ymm21 +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x1b,0xd5,0xab] + vextractf32x8 $0xab, %zmm18, %ymm21 + +// CHECK: vextractf32x8 $171, %zmm18, %ymm21 {%k1} +// CHECK: encoding: [0x62,0xa3,0x7d,0x49,0x1b,0xd5,0xab] + vextractf32x8 $0xab, %zmm18, %ymm21 {%k1} + +// CHECK: vextractf32x8 $171, %zmm18, %ymm21 {%k1} {z} +// CHECK: encoding: [0x62,0xa3,0x7d,0xc9,0x1b,0xd5,0xab] + vextractf32x8 $0xab, %zmm18, %ymm21 {%k1} {z} + +// CHECK: vextractf32x8 $123, %zmm18, %ymm21 +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x1b,0xd5,0x7b] + vextractf32x8 $0x7b, %zmm18, %ymm21 + +// CHECK: vextractf32x8 $171, %zmm21, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x29,0xab] + vextractf32x8 $0xab, %zmm21,(%rcx) + +// CHECK: vextractf32x8 $171, %zmm21, (%rcx) {%k3} +// CHECK: encoding: [0x62,0xe3,0x7d,0x4b,0x1b,0x29,0xab] + vextractf32x8 $0xab, %zmm21,(%rcx) {%k3} + +// CHECK: vextractf32x8 $123, %zmm21, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x29,0x7b] + vextractf32x8 $0x7b, %zmm21,(%rcx) + +// CHECK: vextractf32x8 $123, %zmm21, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x1b,0xac,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf32x8 $0x7b, %zmm21,291(%rax,%r14,8) + +// CHECK: vextractf32x8 $123, %zmm21, 4064(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x6a,0x7f,0x7b] + vextractf32x8 $0x7b, %zmm21,4064(%rdx) + +// CHECK: vextractf32x8 $123, %zmm21, 4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0xaa,0x00,0x10,0x00,0x00,0x7b] + vextractf32x8 $0x7b, %zmm21,4096(%rdx) + +// CHECK: vextractf32x8 $123, %zmm21, -4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x6a,0x80,0x7b] + vextractf32x8 $0x7b, %zmm21,-4096(%rdx) + +// CHECK: vextractf32x8 $123, %zmm21, -4128(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0xaa,0xe0,0xef,0xff,0xff,0x7b] + vextractf32x8 $0x7b, %zmm21,-4128(%rdx) + +// CHECK: vextractf32x8 $171, %zmm26, %ymm30 +// CHECK: encoding: [0x62,0x03,0x7d,0x48,0x1b,0xd6,0xab] + vextractf32x8 $0xab, %zmm26, %ymm30 + +// CHECK: vextractf32x8 $171, %zmm26, %ymm30 {%k3} +// CHECK: encoding: [0x62,0x03,0x7d,0x4b,0x1b,0xd6,0xab] + vextractf32x8 $0xab, %zmm26, %ymm30 {%k3} + +// CHECK: vextractf32x8 $171, %zmm26, %ymm30 {%k3} {z} +// CHECK: encoding: [0x62,0x03,0x7d,0xcb,0x1b,0xd6,0xab] + vextractf32x8 $0xab, %zmm26, %ymm30 {%k3} {z} + +// CHECK: vextractf32x8 $123, %zmm26, %ymm30 +// CHECK: encoding: [0x62,0x03,0x7d,0x48,0x1b,0xd6,0x7b] + vextractf32x8 $0x7b, %zmm26, %ymm30 + +// CHECK: vextractf32x8 $171, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x21,0xab] + vextractf32x8 $0xab, %zmm20,(%rcx) + +// CHECK: vextractf32x8 $171, %zmm20, (%rcx) {%k3} +// CHECK: encoding: [0x62,0xe3,0x7d,0x4b,0x1b,0x21,0xab] + vextractf32x8 $0xab, %zmm20,(%rcx) {%k3} + +// CHECK: vextractf32x8 $123, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x21,0x7b] + vextractf32x8 $0x7b, %zmm20,(%rcx) + +// CHECK: vextractf32x8 $123, %zmm20, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x1b,0xa4,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextractf32x8 $0x7b, %zmm20,4660(%rax,%r14,8) + +// CHECK: vextractf32x8 $123, %zmm20, 4064(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x62,0x7f,0x7b] + vextractf32x8 $0x7b, %zmm20,4064(%rdx) + +// CHECK: vextractf32x8 $123, %zmm20, 4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0xa2,0x00,0x10,0x00,0x00,0x7b] + vextractf32x8 $0x7b, %zmm20,4096(%rdx) + +// CHECK: vextractf32x8 $123, %zmm20, -4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0x62,0x80,0x7b] + vextractf32x8 $0x7b, %zmm20,-4096(%rdx) + +// CHECK: vextractf32x8 $123, %zmm20, -4128(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x1b,0xa2,0xe0,0xef,0xff,0xff,0x7b] + vextractf32x8 $0x7b, %zmm20,-4128(%rdx) + +// CHECK: vextractf64x2 $171, %zmm26, %xmm28 +// CHECK: encoding: [0x62,0x03,0xfd,0x48,0x19,0xd4,0xab] + vextractf64x2 $0xab, %zmm26, %xmm28 + +// CHECK: vextractf64x2 $171, %zmm26, %xmm28 {%k5} +// CHECK: encoding: [0x62,0x03,0xfd,0x4d,0x19,0xd4,0xab] + vextractf64x2 $0xab, %zmm26, %xmm28 {%k5} + +// CHECK: vextractf64x2 $171, %zmm26, %xmm28 {%k5} {z} +// CHECK: encoding: [0x62,0x03,0xfd,0xcd,0x19,0xd4,0xab] + vextractf64x2 $0xab, %zmm26, %xmm28 {%k5} {z} + +// CHECK: vextractf64x2 $123, %zmm26, %xmm28 +// CHECK: encoding: [0x62,0x03,0xfd,0x48,0x19,0xd4,0x7b] + vextractf64x2 $0x7b, %zmm26, %xmm28 + +// CHECK: vextractf64x2 $171, %zmm28, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0x21,0xab] + vextractf64x2 $0xab, %zmm28,(%rcx) + +// CHECK: vextractf64x2 $171, %zmm28, (%rcx) {%k3} +// CHECK: encoding: [0x62,0x63,0xfd,0x4b,0x19,0x21,0xab] + vextractf64x2 $0xab, %zmm28,(%rcx) {%k3} + +// CHECK: vextractf64x2 $123, %zmm28, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0x21,0x7b] + vextractf64x2 $0x7b, %zmm28,(%rcx) + +// CHECK: vextractf64x2 $123, %zmm28, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0xfd,0x48,0x19,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %zmm28,291(%rax,%r14,8) + +// CHECK: vextractf64x2 $123, %zmm28, 2032(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0x62,0x7f,0x7b] + vextractf64x2 $0x7b, %zmm28,2032(%rdx) + +// CHECK: vextractf64x2 $123, %zmm28, 2048(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %zmm28,2048(%rdx) + +// CHECK: vextractf64x2 $123, %zmm28, -2048(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0x62,0x80,0x7b] + vextractf64x2 $0x7b, %zmm28,-2048(%rdx) + +// CHECK: vextractf64x2 $123, %zmm28, -2064(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x19,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextractf64x2 $0x7b, %zmm28,-2064(%rdx) + +// CHECK: vextractf64x2 $171, %zmm26, %xmm19 +// CHECK: encoding: [0x62,0x23,0xfd,0x48,0x19,0xd3,0xab] + vextractf64x2 $0xab, %zmm26, %xmm19 + +// CHECK: vextractf64x2 $171, %zmm26, %xmm19 {%k3} +// CHECK: encoding: [0x62,0x23,0xfd,0x4b,0x19,0xd3,0xab] + vextractf64x2 $0xab, %zmm26, %xmm19 {%k3} + +// CHECK: vextractf64x2 $171, %zmm26, %xmm19 {%k3} {z} +// CHECK: encoding: [0x62,0x23,0xfd,0xcb,0x19,0xd3,0xab] + vextractf64x2 $0xab, %zmm26, %xmm19 {%k3} {z} + +// CHECK: vextractf64x2 $123, %zmm26, %xmm19 +// CHECK: encoding: [0x62,0x23,0xfd,0x48,0x19,0xd3,0x7b] + vextractf64x2 $0x7b, %zmm26, %xmm19 + +// CHECK: vextractf64x2 $171, %zmm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x09,0xab] + vextractf64x2 $0xab, %zmm17,(%rcx) + +// CHECK: vextractf64x2 $171, %zmm17, (%rcx) {%k1} +// CHECK: encoding: [0x62,0xe3,0xfd,0x49,0x19,0x09,0xab] + vextractf64x2 $0xab, %zmm17,(%rcx) {%k1} + +// CHECK: vextractf64x2 $123, %zmm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x09,0x7b] + vextractf64x2 $0x7b, %zmm17,(%rcx) + +// CHECK: vextractf64x2 $123, %zmm17, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x48,0x19,0x8c,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %zmm17,4660(%rax,%r14,8) + +// CHECK: vextractf64x2 $123, %zmm17, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x4a,0x7f,0x7b] + vextractf64x2 $0x7b, %zmm17,2032(%rdx) + +// CHECK: vextractf64x2 $123, %zmm17, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x8a,0x00,0x08,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %zmm17,2048(%rdx) + +// CHECK: vextractf64x2 $123, %zmm17, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x4a,0x80,0x7b] + vextractf64x2 $0x7b, %zmm17,-2048(%rdx) + +// CHECK: vextractf64x2 $123, %zmm17, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x19,0x8a,0xf0,0xf7,0xff,0xff,0x7b] + vextractf64x2 $0x7b, %zmm17,-2064(%rdx) + +// CHECK: vextracti32x8 $171, %zmm24, %ymm20 +// CHECK: encoding: [0x62,0x23,0x7d,0x48,0x3b,0xc4,0xab] + vextracti32x8 $0xab, %zmm24, %ymm20 + +// CHECK: vextracti32x8 $171, %zmm24, %ymm20 {%k1} +// CHECK: encoding: [0x62,0x23,0x7d,0x49,0x3b,0xc4,0xab] + vextracti32x8 $0xab, %zmm24, %ymm20 {%k1} + +// CHECK: vextracti32x8 $171, %zmm24, %ymm20 {%k1} {z} +// CHECK: encoding: [0x62,0x23,0x7d,0xc9,0x3b,0xc4,0xab] + vextracti32x8 $0xab, %zmm24, %ymm20 {%k1} {z} + +// CHECK: vextracti32x8 $123, %zmm24, %ymm20 +// CHECK: encoding: [0x62,0x23,0x7d,0x48,0x3b,0xc4,0x7b] + vextracti32x8 $0x7b, %zmm24, %ymm20 + +// CHECK: vextracti32x8 $171, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0x21,0xab] + vextracti32x8 $0xab, %zmm20,(%rcx) + +// CHECK: vextracti32x8 $171, %zmm20, (%rcx) {%k3} +// CHECK: encoding: [0x62,0xe3,0x7d,0x4b,0x3b,0x21,0xab] + vextracti32x8 $0xab, %zmm20,(%rcx) {%k3} + +// CHECK: vextracti32x8 $123, %zmm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0x21,0x7b] + vextracti32x8 $0x7b, %zmm20,(%rcx) + +// CHECK: vextracti32x8 $123, %zmm20, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0x7d,0x48,0x3b,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti32x8 $0x7b, %zmm20,291(%rax,%r14,8) + +// CHECK: vextracti32x8 $123, %zmm20, 4064(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0x62,0x7f,0x7b] + vextracti32x8 $0x7b, %zmm20,4064(%rdx) + +// CHECK: vextracti32x8 $123, %zmm20, 4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0xa2,0x00,0x10,0x00,0x00,0x7b] + vextracti32x8 $0x7b, %zmm20,4096(%rdx) + +// CHECK: vextracti32x8 $123, %zmm20, -4096(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0x62,0x80,0x7b] + vextracti32x8 $0x7b, %zmm20,-4096(%rdx) + +// CHECK: vextracti32x8 $123, %zmm20, -4128(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x48,0x3b,0xa2,0xe0,0xef,0xff,0xff,0x7b] + vextracti32x8 $0x7b, %zmm20,-4128(%rdx) + +// CHECK: vextracti32x8 $171, %zmm29, %ymm27 +// CHECK: encoding: [0x62,0x03,0x7d,0x48,0x3b,0xeb,0xab] + vextracti32x8 $0xab, %zmm29, %ymm27 + +// CHECK: vextracti32x8 $171, %zmm29, %ymm27 {%k2} +// CHECK: encoding: [0x62,0x03,0x7d,0x4a,0x3b,0xeb,0xab] + vextracti32x8 $0xab, %zmm29, %ymm27 {%k2} + +// CHECK: vextracti32x8 $171, %zmm29, %ymm27 {%k2} {z} +// CHECK: encoding: [0x62,0x03,0x7d,0xca,0x3b,0xeb,0xab] + vextracti32x8 $0xab, %zmm29, %ymm27 {%k2} {z} + +// CHECK: vextracti32x8 $123, %zmm29, %ymm27 +// CHECK: encoding: [0x62,0x03,0x7d,0x48,0x3b,0xeb,0x7b] + vextracti32x8 $0x7b, %zmm29, %ymm27 + +// CHECK: vextracti32x8 $171, %zmm26, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x11,0xab] + vextracti32x8 $0xab, %zmm26,(%rcx) + +// CHECK: vextracti32x8 $171, %zmm26, (%rcx) {%k2} +// CHECK: encoding: [0x62,0x63,0x7d,0x4a,0x3b,0x11,0xab] + vextracti32x8 $0xab, %zmm26,(%rcx) {%k2} + +// CHECK: vextracti32x8 $123, %zmm26, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x11,0x7b] + vextracti32x8 $0x7b, %zmm26,(%rcx) + +// CHECK: vextracti32x8 $123, %zmm26, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0x7d,0x48,0x3b,0x94,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextracti32x8 $0x7b, %zmm26,4660(%rax,%r14,8) + +// CHECK: vextracti32x8 $123, %zmm26, 4064(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x52,0x7f,0x7b] + vextracti32x8 $0x7b, %zmm26,4064(%rdx) + +// CHECK: vextracti32x8 $123, %zmm26, 4096(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x92,0x00,0x10,0x00,0x00,0x7b] + vextracti32x8 $0x7b, %zmm26,4096(%rdx) + +// CHECK: vextracti32x8 $123, %zmm26, -4096(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x52,0x80,0x7b] + vextracti32x8 $0x7b, %zmm26,-4096(%rdx) + +// CHECK: vextracti32x8 $123, %zmm26, -4128(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x48,0x3b,0x92,0xe0,0xef,0xff,0xff,0x7b] + vextracti32x8 $0x7b, %zmm26,-4128(%rdx) + +// CHECK: vextracti64x2 $171, %zmm20, %xmm17 +// CHECK: encoding: [0x62,0xa3,0xfd,0x48,0x39,0xe1,0xab] + vextracti64x2 $0xab, %zmm20, %xmm17 + +// CHECK: vextracti64x2 $171, %zmm20, %xmm17 {%k2} +// CHECK: encoding: [0x62,0xa3,0xfd,0x4a,0x39,0xe1,0xab] + vextracti64x2 $0xab, %zmm20, %xmm17 {%k2} + +// CHECK: vextracti64x2 $171, %zmm20, %xmm17 {%k2} {z} +// CHECK: encoding: [0x62,0xa3,0xfd,0xca,0x39,0xe1,0xab] + vextracti64x2 $0xab, %zmm20, %xmm17 {%k2} {z} + +// CHECK: vextracti64x2 $123, %zmm20, %xmm17 +// CHECK: encoding: [0x62,0xa3,0xfd,0x48,0x39,0xe1,0x7b] + vextracti64x2 $0x7b, %zmm20, %xmm17 + +// CHECK: vextracti64x2 $171, %zmm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x09,0xab] + vextracti64x2 $0xab, %zmm17,(%rcx) + +// CHECK: vextracti64x2 $171, %zmm17, (%rcx) {%k5} +// CHECK: encoding: [0x62,0xe3,0xfd,0x4d,0x39,0x09,0xab] + vextracti64x2 $0xab, %zmm17,(%rcx) {%k5} + +// CHECK: vextracti64x2 $123, %zmm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x09,0x7b] + vextracti64x2 $0x7b, %zmm17,(%rcx) + +// CHECK: vextracti64x2 $123, %zmm17, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x48,0x39,0x8c,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %zmm17,291(%rax,%r14,8) + +// CHECK: vextracti64x2 $123, %zmm17, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x4a,0x7f,0x7b] + vextracti64x2 $0x7b, %zmm17,2032(%rdx) + +// CHECK: vextracti64x2 $123, %zmm17, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x8a,0x00,0x08,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %zmm17,2048(%rdx) + +// CHECK: vextracti64x2 $123, %zmm17, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x4a,0x80,0x7b] + vextracti64x2 $0x7b, %zmm17,-2048(%rdx) + +// CHECK: vextracti64x2 $123, %zmm17, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x48,0x39,0x8a,0xf0,0xf7,0xff,0xff,0x7b] + vextracti64x2 $0x7b, %zmm17,-2064(%rdx) + +// CHECK: vextracti64x2 $171, %zmm23, %xmm27 +// CHECK: encoding: [0x62,0x83,0xfd,0x48,0x39,0xfb,0xab] + vextracti64x2 $0xab, %zmm23, %xmm27 + +// CHECK: vextracti64x2 $171, %zmm23, %xmm27 {%k5} +// CHECK: encoding: [0x62,0x83,0xfd,0x4d,0x39,0xfb,0xab] + vextracti64x2 $0xab, %zmm23, %xmm27 {%k5} + +// CHECK: vextracti64x2 $171, %zmm23, %xmm27 {%k5} {z} +// CHECK: encoding: [0x62,0x83,0xfd,0xcd,0x39,0xfb,0xab] + vextracti64x2 $0xab, %zmm23, %xmm27 {%k5} {z} + +// CHECK: vextracti64x2 $123, %zmm23, %xmm27 +// CHECK: encoding: [0x62,0x83,0xfd,0x48,0x39,0xfb,0x7b] + vextracti64x2 $0x7b, %zmm23, %xmm27 + +// CHECK: vextracti64x2 $171, %zmm24, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x01,0xab] + vextracti64x2 $0xab, %zmm24,(%rcx) + +// CHECK: vextracti64x2 $171, %zmm24, (%rcx) {%k3} +// CHECK: encoding: [0x62,0x63,0xfd,0x4b,0x39,0x01,0xab] + vextracti64x2 $0xab, %zmm24,(%rcx) {%k3} + +// CHECK: vextracti64x2 $123, %zmm24, (%rcx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x01,0x7b] + vextracti64x2 $0x7b, %zmm24,(%rcx) + +// CHECK: vextracti64x2 $123, %zmm24, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0xfd,0x48,0x39,0x84,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %zmm24,4660(%rax,%r14,8) + +// CHECK: vextracti64x2 $123, %zmm24, 2032(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x42,0x7f,0x7b] + vextracti64x2 $0x7b, %zmm24,2032(%rdx) + +// CHECK: vextracti64x2 $123, %zmm24, 2048(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x82,0x00,0x08,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %zmm24,2048(%rdx) + +// CHECK: vextracti64x2 $123, %zmm24, -2048(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x42,0x80,0x7b] + vextracti64x2 $0x7b, %zmm24,-2048(%rdx) + +// CHECK: vextracti64x2 $123, %zmm24, -2064(%rdx) +// CHECK: encoding: [0x62,0x63,0xfd,0x48,0x39,0x82,0xf0,0xf7,0xff,0xff,0x7b] + vextracti64x2 $0x7b, %zmm24,-2064(%rdx) + Index: test/MC/X86/x86-64-avx512dq_vl.s =================================================================== --- test/MC/X86/x86-64-avx512dq_vl.s +++ test/MC/X86/x86-64-avx512dq_vl.s @@ -3584,3 +3584,195 @@ // CHECK: encoding: [0x62,0x61,0xff,0x38,0x7a,0xa2,0xf8,0xfb,0xff,0xff] vcvtuqq2ps -1032(%rdx){1to4}, %xmm28 +// CHECK: vextractf64x2 $171, %ymm21, %xmm27 +// CHECK: encoding: [0x62,0x83,0xfd,0x28,0x19,0xeb,0xab] + vextractf64x2 $0xab, %ymm21, %xmm27 + +// CHECK: vextractf64x2 $171, %ymm21, %xmm27 {%k7} +// CHECK: encoding: [0x62,0x83,0xfd,0x2f,0x19,0xeb,0xab] + vextractf64x2 $0xab, %ymm21, %xmm27 {%k7} + +// CHECK: vextractf64x2 $171, %ymm21, %xmm27 {%k7} {z} +// CHECK: encoding: [0x62,0x83,0xfd,0xaf,0x19,0xeb,0xab] + vextractf64x2 $0xab, %ymm21, %xmm27 {%k7} {z} + +// CHECK: vextractf64x2 $123, %ymm21, %xmm27 +// CHECK: encoding: [0x62,0x83,0xfd,0x28,0x19,0xeb,0x7b] + vextractf64x2 $0x7b, %ymm21, %xmm27 + +// CHECK: vextractf64x2 $171, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x21,0xab] + vextractf64x2 $0xab, %ymm20,(%rcx) + +// CHECK: vextractf64x2 $171, %ymm20, (%rcx) {%k1} +// CHECK: encoding: [0x62,0xe3,0xfd,0x29,0x19,0x21,0xab] + vextractf64x2 $0xab, %ymm20,(%rcx) {%k1} + +// CHECK: vextractf64x2 $123, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x21,0x7b] + vextractf64x2 $0x7b, %ymm20,(%rcx) + +// CHECK: vextractf64x2 $123, %ymm20, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x28,0x19,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %ymm20,291(%rax,%r14,8) + +// CHECK: vextractf64x2 $123, %ymm20, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x62,0x7f,0x7b] + vextractf64x2 $0x7b, %ymm20,2032(%rdx) + +// CHECK: vextractf64x2 $123, %ymm20, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %ymm20,2048(%rdx) + +// CHECK: vextractf64x2 $123, %ymm20, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x62,0x80,0x7b] + vextractf64x2 $0x7b, %ymm20,-2048(%rdx) + +// CHECK: vextractf64x2 $123, %ymm20, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextractf64x2 $0x7b, %ymm20,-2064(%rdx) + +// CHECK: vextractf64x2 $171, %ymm26, %xmm28 +// CHECK: encoding: [0x62,0x03,0xfd,0x28,0x19,0xd4,0xab] + vextractf64x2 $0xab, %ymm26, %xmm28 + +// CHECK: vextractf64x2 $171, %ymm26, %xmm28 {%k4} +// CHECK: encoding: [0x62,0x03,0xfd,0x2c,0x19,0xd4,0xab] + vextractf64x2 $0xab, %ymm26, %xmm28 {%k4} + +// CHECK: vextractf64x2 $171, %ymm26, %xmm28 {%k4} {z} +// CHECK: encoding: [0x62,0x03,0xfd,0xac,0x19,0xd4,0xab] + vextractf64x2 $0xab, %ymm26, %xmm28 {%k4} {z} + +// CHECK: vextractf64x2 $123, %ymm26, %xmm28 +// CHECK: encoding: [0x62,0x03,0xfd,0x28,0x19,0xd4,0x7b] + vextractf64x2 $0x7b, %ymm26, %xmm28 + +// CHECK: vextractf64x2 $171, %ymm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x09,0xab] + vextractf64x2 $0xab, %ymm17,(%rcx) + +// CHECK: vextractf64x2 $171, %ymm17, (%rcx) {%k2} +// CHECK: encoding: [0x62,0xe3,0xfd,0x2a,0x19,0x09,0xab] + vextractf64x2 $0xab, %ymm17,(%rcx) {%k2} + +// CHECK: vextractf64x2 $123, %ymm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x09,0x7b] + vextractf64x2 $0x7b, %ymm17,(%rcx) + +// CHECK: vextractf64x2 $123, %ymm17, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x28,0x19,0x8c,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %ymm17,4660(%rax,%r14,8) + +// CHECK: vextractf64x2 $123, %ymm17, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x4a,0x7f,0x7b] + vextractf64x2 $0x7b, %ymm17,2032(%rdx) + +// CHECK: vextractf64x2 $123, %ymm17, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x8a,0x00,0x08,0x00,0x00,0x7b] + vextractf64x2 $0x7b, %ymm17,2048(%rdx) + +// CHECK: vextractf64x2 $123, %ymm17, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x4a,0x80,0x7b] + vextractf64x2 $0x7b, %ymm17,-2048(%rdx) + +// CHECK: vextractf64x2 $123, %ymm17, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x19,0x8a,0xf0,0xf7,0xff,0xff,0x7b] + vextractf64x2 $0x7b, %ymm17,-2064(%rdx) + +// CHECK: vextracti64x2 $171, %ymm24, %xmm29 +// CHECK: encoding: [0x62,0x03,0xfd,0x28,0x39,0xc5,0xab] + vextracti64x2 $0xab, %ymm24, %xmm29 + +// CHECK: vextracti64x2 $171, %ymm24, %xmm29 {%k7} +// CHECK: encoding: [0x62,0x03,0xfd,0x2f,0x39,0xc5,0xab] + vextracti64x2 $0xab, %ymm24, %xmm29 {%k7} + +// CHECK: vextracti64x2 $171, %ymm24, %xmm29 {%k7} {z} +// CHECK: encoding: [0x62,0x03,0xfd,0xaf,0x39,0xc5,0xab] + vextracti64x2 $0xab, %ymm24, %xmm29 {%k7} {z} + +// CHECK: vextracti64x2 $123, %ymm24, %xmm29 +// CHECK: encoding: [0x62,0x03,0xfd,0x28,0x39,0xc5,0x7b] + vextracti64x2 $0x7b, %ymm24, %xmm29 + +// CHECK: vextracti64x2 $171, %ymm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x09,0xab] + vextracti64x2 $0xab, %ymm17,(%rcx) + +// CHECK: vextracti64x2 $171, %ymm17, (%rcx) {%k1} +// CHECK: encoding: [0x62,0xe3,0xfd,0x29,0x39,0x09,0xab] + vextracti64x2 $0xab, %ymm17,(%rcx) {%k1} + +// CHECK: vextracti64x2 $123, %ymm17, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x09,0x7b] + vextracti64x2 $0x7b, %ymm17,(%rcx) + +// CHECK: vextracti64x2 $123, %ymm17, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x28,0x39,0x8c,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %ymm17,291(%rax,%r14,8) + +// CHECK: vextracti64x2 $123, %ymm17, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x4a,0x7f,0x7b] + vextracti64x2 $0x7b, %ymm17,2032(%rdx) + +// CHECK: vextracti64x2 $123, %ymm17, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x8a,0x00,0x08,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %ymm17,2048(%rdx) + +// CHECK: vextracti64x2 $123, %ymm17, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x4a,0x80,0x7b] + vextracti64x2 $0x7b, %ymm17,-2048(%rdx) + +// CHECK: vextracti64x2 $123, %ymm17, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x8a,0xf0,0xf7,0xff,0xff,0x7b] + vextracti64x2 $0x7b, %ymm17,-2064(%rdx) + +// CHECK: vextracti64x2 $171, %ymm17, %xmm29 +// CHECK: encoding: [0x62,0x83,0xfd,0x28,0x39,0xcd,0xab] + vextracti64x2 $0xab, %ymm17, %xmm29 + +// CHECK: vextracti64x2 $171, %ymm17, %xmm29 {%k5} +// CHECK: encoding: [0x62,0x83,0xfd,0x2d,0x39,0xcd,0xab] + vextracti64x2 $0xab, %ymm17, %xmm29 {%k5} + +// CHECK: vextracti64x2 $171, %ymm17, %xmm29 {%k5} {z} +// CHECK: encoding: [0x62,0x83,0xfd,0xad,0x39,0xcd,0xab] + vextracti64x2 $0xab, %ymm17, %xmm29 {%k5} {z} + +// CHECK: vextracti64x2 $123, %ymm17, %xmm29 +// CHECK: encoding: [0x62,0x83,0xfd,0x28,0x39,0xcd,0x7b] + vextracti64x2 $0x7b, %ymm17, %xmm29 + +// CHECK: vextracti64x2 $171, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x21,0xab] + vextracti64x2 $0xab, %ymm20,(%rcx) + +// CHECK: vextracti64x2 $171, %ymm20, (%rcx) {%k2} +// CHECK: encoding: [0x62,0xe3,0xfd,0x2a,0x39,0x21,0xab] + vextracti64x2 $0xab, %ymm20,(%rcx) {%k2} + +// CHECK: vextracti64x2 $123, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x21,0x7b] + vextracti64x2 $0x7b, %ymm20,(%rcx) + +// CHECK: vextracti64x2 $123, %ymm20, 4660(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0xfd,0x28,0x39,0xa4,0xf0,0x34,0x12,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %ymm20,4660(%rax,%r14,8) + +// CHECK: vextracti64x2 $123, %ymm20, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x62,0x7f,0x7b] + vextracti64x2 $0x7b, %ymm20,2032(%rdx) + +// CHECK: vextracti64x2 $123, %ymm20, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextracti64x2 $0x7b, %ymm20,2048(%rdx) + +// CHECK: vextracti64x2 $123, %ymm20, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0x62,0x80,0x7b] + vextracti64x2 $0x7b, %ymm20,-2048(%rdx) + +// CHECK: vextracti64x2 $123, %ymm20, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0xfd,0x28,0x39,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextracti64x2 $0x7b, %ymm20,-2064(%rdx) + Index: test/MC/X86/x86-64-avx512f_vl.s =================================================================== --- test/MC/X86/x86-64-avx512f_vl.s +++ test/MC/X86/x86-64-avx512f_vl.s @@ -19739,3 +19739,99 @@ // CHECK: encoding: [0x62,0xe1,0xe5,0x30,0x6d,0xa2,0xf8,0xfb,0xff,0xff] vpunpckhqdq -1032(%rdx){1to4}, %ymm19, %ymm20 +// CHECK: vextractf32x4 $171, %ymm17, %xmm28 +// CHECK: encoding: [0x62,0x83,0x7d,0x28,0x19,0xcc,0xab] + vextractf32x4 $0xab, %ymm17, %xmm28 + +// CHECK: vextractf32x4 $171, %ymm17, %xmm28 {%k6} +// CHECK: encoding: [0x62,0x83,0x7d,0x2e,0x19,0xcc,0xab] + vextractf32x4 $0xab, %ymm17, %xmm28 {%k6} + +// CHECK: vextractf32x4 $171, %ymm17, %xmm28 {%k6} {z} +// CHECK: encoding: [0x62,0x83,0x7d,0xae,0x19,0xcc,0xab] + vextractf32x4 $0xab, %ymm17, %xmm28 {%k6} {z} + +// CHECK: vextractf32x4 $123, %ymm17, %xmm28 +// CHECK: encoding: [0x62,0x83,0x7d,0x28,0x19,0xcc,0x7b] + vextractf32x4 $0x7b, %ymm17, %xmm28 + +// CHECK: vextractf32x4 $171, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0x21,0xab] + vextractf32x4 $0xab, %ymm20, (%rcx) + +// CHECK: vextractf32x4 $171, %ymm20, (%rcx) {%k2} +// CHECK: encoding: [0x62,0xe3,0x7d,0x2a,0x19,0x21,0xab] + vextractf32x4 $0xab, %ymm20, (%rcx) {%k2} + +// CHECK: vextractf32x4 $123, %ymm20, (%rcx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0x21,0x7b] + vextractf32x4 $0x7b, %ymm20, (%rcx) + +// CHECK: vextractf32x4 $123, %ymm20, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0xa3,0x7d,0x28,0x19,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextractf32x4 $0x7b, %ymm20, 291(%rax,%r14,8) + +// CHECK: vextractf32x4 $123, %ymm20, 2032(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0x62,0x7f,0x7b] + vextractf32x4 $0x7b, %ymm20, 2032(%rdx) + +// CHECK: vextractf32x4 $123, %ymm20, 2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextractf32x4 $0x7b, %ymm20, 2048(%rdx) + +// CHECK: vextractf32x4 $123, %ymm20, -2048(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0x62,0x80,0x7b] + vextractf32x4 $0x7b, %ymm20, -2048(%rdx) + +// CHECK: vextractf32x4 $123, %ymm20, -2064(%rdx) +// CHECK: encoding: [0x62,0xe3,0x7d,0x28,0x19,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextractf32x4 $0x7b, %ymm20, -2064(%rdx) + +// CHECK: vextracti32x4 $171, %ymm21, %xmm20 +// CHECK: encoding: [0x62,0xa3,0x7d,0x28,0x39,0xec,0xab] + vextracti32x4 $0xab, %ymm21, %xmm20 + +// CHECK: vextracti32x4 $171, %ymm21, %xmm20 {%k4} +// CHECK: encoding: [0x62,0xa3,0x7d,0x2c,0x39,0xec,0xab] + vextracti32x4 $0xab, %ymm21, %xmm20 {%k4} + +// CHECK: vextracti32x4 $171, %ymm21, %xmm20 {%k4} {z} +// CHECK: encoding: [0x62,0xa3,0x7d,0xac,0x39,0xec,0xab] + vextracti32x4 $0xab, %ymm21, %xmm20 {%k4} {z} + +// CHECK: vextracti32x4 $123, %ymm21, %xmm20 +// CHECK: encoding: [0x62,0xa3,0x7d,0x28,0x39,0xec,0x7b] + vextracti32x4 $0x7b, %ymm21, %xmm20 + +// CHECK: vextracti32x4 $171, %ymm28, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0x21,0xab] + vextracti32x4 $0xab, %ymm28, (%rcx) + +// CHECK: vextracti32x4 $171, %ymm28, (%rcx) {%k6} +// CHECK: encoding: [0x62,0x63,0x7d,0x2e,0x39,0x21,0xab] + vextracti32x4 $0xab, %ymm28, (%rcx) {%k6} + +// CHECK: vextracti32x4 $123, %ymm28, (%rcx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0x21,0x7b] + vextracti32x4 $0x7b, %ymm28, (%rcx) + +// CHECK: vextracti32x4 $123, %ymm28, 291(%rax,%r14,8) +// CHECK: encoding: [0x62,0x23,0x7d,0x28,0x39,0xa4,0xf0,0x23,0x01,0x00,0x00,0x7b] + vextracti32x4 $0x7b, %ymm28, 291(%rax,%r14,8) + +// CHECK: vextracti32x4 $123, %ymm28, 2032(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0x62,0x7f,0x7b] + vextracti32x4 $0x7b, %ymm28, 2032(%rdx) + +// CHECK: vextracti32x4 $123, %ymm28, 2048(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0xa2,0x00,0x08,0x00,0x00,0x7b] + vextracti32x4 $0x7b, %ymm28, 2048(%rdx) + +// CHECK: vextracti32x4 $123, %ymm28, -2048(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0x62,0x80,0x7b] + vextracti32x4 $0x7b, %ymm28, -2048(%rdx) + +// CHECK: vextracti32x4 $123, %ymm28, -2064(%rdx) +// CHECK: encoding: [0x62,0x63,0x7d,0x28,0x39,0xa2,0xf0,0xf7,0xff,0xff,0x7b] + vextracti32x4 $0x7b, %ymm28, -2064(%rdx) +