Index: lib/Target/X86/X86ISelLowering.cpp =================================================================== --- lib/Target/X86/X86ISelLowering.cpp +++ lib/Target/X86/X86ISelLowering.cpp @@ -1225,6 +1225,9 @@ setOperationAction(ISD::MSTORE, VT, Custom); } } + if (Subtarget.hasDQI() && Subtarget.hasVLX() && Subtarget.hasAVX512() && + Subtarget.hasBWI()) + setOperationAction(ISD::SELECT, MVT::v1i1, Custom); if (Subtarget.hasDQI()) { for (auto VT : { MVT::v2i64, MVT::v4i64, MVT::v8i64 }) { Index: test/CodeGen/X86/avx512-load-trunc-store-i1.ll =================================================================== --- /dev/null +++ test/CodeGen/X86/avx512-load-trunc-store-i1.ll @@ -0,0 +1,99 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=+avx512f,+avx512bw,+avx512vl,+avx512dq -O2 | FileCheck %s --check-prefix=AVX512 + + +define void @load_v1i2_trunc_v1i1_store(<1 x i2>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i2_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: movb (%rdi), %al +; AVX512-NEXT: testb %al, %al +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i2>, <1 x i2>* %a0 + %d1 = trunc <1 x i2> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i3_trunc_v1i1_store(<1 x i3>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i3_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: movb (%rdi), %al +; AVX512-NEXT: testb %al, %al +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i3>, <1 x i3>* %a0 + %d1 = trunc <1 x i3> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i4_trunc_v1i1_store(<1 x i4>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i4_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: movb (%rdi), %al +; AVX512-NEXT: testb %al, %al +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i4>, <1 x i4>* %a0 + %d1 = trunc <1 x i4> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i8_trunc_v1i1_store(<1 x i8>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i8_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: cmpb $0, (%rdi) +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i8>, <1 x i8>* %a0 + %d1 = trunc <1 x i8> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i16_trunc_v1i1_store(<1 x i16>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i16_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: cmpb $0, (%rdi) +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i16>, <1 x i16>* %a0 + %d1 = trunc <1 x i16> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i32_trunc_v1i1_store(<1 x i32>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i32_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: cmpb $0, (%rdi) +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i32>, <1 x i32>* %a0 + %d1 = trunc <1 x i32> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +define void @load_v1i64_trunc_v1i1_store(<1 x i64>* %a0,<1 x i1>* %a1) { +; AVX512-LABEL: load_v1i64_trunc_v1i1_store: +; AVX512: # BB#0: +; AVX512-NEXT: cmpb $0, (%rdi) +; AVX512-NEXT: setne %al +; AVX512-NEXT: kmovd %eax, %k0 +; AVX512-NEXT: kmovb %k0, (%rsi) +; AVX512-NEXT: retq + %d0 = load <1 x i64>, <1 x i64>* %a0 + %d1 = trunc <1 x i64> %d0 to <1 x i1> + store <1 x i1> %d1, <1 x i1>* %a1 + ret void +} +