diff --git a/llvm/lib/Target/ARM/ARMISelDAGToDAG.cpp b/llvm/lib/Target/ARM/ARMISelDAGToDAG.cpp --- a/llvm/lib/Target/ARM/ARMISelDAGToDAG.cpp +++ b/llvm/lib/Target/ARM/ARMISelDAGToDAG.cpp @@ -2078,6 +2078,7 @@ // Double-register operations: case MVT::v8i8: OpcodeIndex = 0; break; case MVT::v4f16: + case MVT::v4bf16: case MVT::v4i16: OpcodeIndex = 1; break; case MVT::v2f32: case MVT::v2i32: OpcodeIndex = 2; break; @@ -2085,6 +2086,7 @@ // Quad-register operations: case MVT::v16i8: OpcodeIndex = 0; break; case MVT::v8f16: + case MVT::v8bf16: case MVT::v8i16: OpcodeIndex = 1; break; case MVT::v4f32: case MVT::v4i32: OpcodeIndex = 2; break; @@ -2221,6 +2223,7 @@ // Double-register operations: case MVT::v8i8: OpcodeIndex = 0; break; case MVT::v4f16: + case MVT::v4bf16: case MVT::v4i16: OpcodeIndex = 1; break; case MVT::v2f32: case MVT::v2i32: OpcodeIndex = 2; break; @@ -2228,6 +2231,7 @@ // Quad-register operations: case MVT::v16i8: OpcodeIndex = 0; break; case MVT::v8f16: + case MVT::v8bf16: case MVT::v8i16: OpcodeIndex = 1; break; case MVT::v4f32: case MVT::v4i32: OpcodeIndex = 2; break; @@ -2389,11 +2393,13 @@ // Double-register operations: case MVT::v8i8: OpcodeIndex = 0; break; case MVT::v4f16: + case MVT::v4bf16: case MVT::v4i16: OpcodeIndex = 1; break; case MVT::v2f32: case MVT::v2i32: OpcodeIndex = 2; break; // Quad-register operations: case MVT::v8f16: + case MVT::v8bf16: case MVT::v8i16: OpcodeIndex = 0; break; case MVT::v4f32: case MVT::v4i32: OpcodeIndex = 1; break; @@ -2923,6 +2929,8 @@ case MVT::v8i16: case MVT::v4f16: case MVT::v8f16: + case MVT::v4bf16: + case MVT::v8bf16: OpcodeIndex = 1; break; case MVT::v2f32: case MVT::v2i32: diff --git a/llvm/lib/Target/ARM/ARMInstrNEON.td b/llvm/lib/Target/ARM/ARMInstrNEON.td --- a/llvm/lib/Target/ARM/ARMInstrNEON.td +++ b/llvm/lib/Target/ARM/ARMInstrNEON.td @@ -1074,6 +1074,12 @@ def : Pat<(vector_insert (v8f16 QPR:$src), (f16 (load addrmode6:$addr)), imm:$lane), (VLD1LNq16Pseudo addrmode6:$addr, QPR:$src, imm:$lane)>; +def : Pat<(vector_insert (v4bf16 DPR:$src), + (bf16 (load addrmode6:$addr)), imm:$lane), + (VLD1LNd16 addrmode6:$addr, DPR:$src, imm:$lane)>; +def : Pat<(vector_insert (v8bf16 QPR:$src), + (bf16 (load addrmode6:$addr)), imm:$lane), + (VLD1LNq16Pseudo addrmode6:$addr, QPR:$src, imm:$lane)>; def : Pat<(vector_insert (v2f32 DPR:$src), (f32 (load addrmode6:$addr)), imm:$lane), (VLD1LNd32 addrmode6:$addr, DPR:$src, imm:$lane)>; diff --git a/llvm/test/CodeGen/ARM/bf16-intrinsics-ld-st.ll b/llvm/test/CodeGen/ARM/bf16-intrinsics-ld-st.ll new file mode 100644 --- /dev/null +++ b/llvm/test/CodeGen/ARM/bf16-intrinsics-ld-st.ll @@ -0,0 +1,1062 @@ +; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py +; RUN: llc -mtriple=armv8.6a-arm-none-eabi -mattr=+bf16,+neon,+fullfp16 < %s | FileCheck %s +; FIXME: Remove fullfp16 once bfloat arguments and returns lowering stops +; depending on it. + +define arm_aapcs_vfpcc <4 x bfloat> @test_vld1_bf16(bfloat* nocapture readonly %ptr) { +; CHECK-LABEL: test_vld1_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to <4 x bfloat>* + %1 = load <4 x bfloat>, <4 x bfloat>* %0, align 2 + ret <4 x bfloat> %1 +} + +define arm_aapcs_vfpcc <8 x bfloat> @test_vld1q_bf16(bfloat* nocapture readonly %ptr) { +; CHECK-LABEL: test_vld1q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to <8 x bfloat>* + %1 = load <8 x bfloat>, <8 x bfloat>* %0, align 2 + ret <8 x bfloat> %1 +} + +define arm_aapcs_vfpcc <4 x bfloat> @test_vld1_lane_bf16(bfloat* nocapture readonly %ptr, <4 x bfloat> %src) { +; CHECK-LABEL: test_vld1_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0[0]}, [r0:16] +; CHECK-NEXT: bx lr +entry: + %0 = load bfloat, bfloat* %ptr, align 2 + %vld1_lane = insertelement <4 x bfloat> %src, bfloat %0, i32 0 + ret <4 x bfloat> %vld1_lane +} + +define arm_aapcs_vfpcc <8 x bfloat> @test_vld1q_lane_bf16(bfloat* nocapture readonly %ptr, <8 x bfloat> %src) { +; CHECK-LABEL: test_vld1q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d1[3]}, [r0:16] +; CHECK-NEXT: bx lr +entry: + %0 = load bfloat, bfloat* %ptr, align 2 + %vld1_lane = insertelement <8 x bfloat> %src, bfloat %0, i32 7 + ret <8 x bfloat> %vld1_lane +} + +define arm_aapcs_vfpcc <4 x bfloat> @test_vld1_dup_bf16(bfloat* nocapture readonly %ptr) { +; CHECK-LABEL: test_vld1_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0[]}, [r0:16] +; CHECK-NEXT: bx lr +entry: + %0 = load bfloat, bfloat* %ptr, align 2 + %1 = insertelement <4 x bfloat> undef, bfloat %0, i32 0 + %lane = shufflevector <4 x bfloat> %1, <4 x bfloat> undef, <4 x i32> zeroinitializer + ret <4 x bfloat> %lane +} + +define arm_aapcs_vfpcc [2 x <2 x i32>] @test_vld1_bf16_x2(bfloat* %ptr) { +; CHECK-LABEL: test_vld1_bf16_x2: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x2.v4bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld1xN, 1 + %0 = bitcast <4 x bfloat> %vld1xN.fca.0.extract to <2 x i32> + %1 = bitcast <4 x bfloat> %vld1xN.fca.1.extract to <2 x i32> + %.fca.0.insert = insertvalue [2 x <2 x i32>] undef, <2 x i32> %0, 0 + %.fca.1.insert = insertvalue [2 x <2 x i32>] %.fca.0.insert, <2 x i32> %1, 1 + ret [2 x <2 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [2 x <4 x i32>] @test_vld1q_bf16_x2(bfloat* %ptr) { +; CHECK-LABEL: test_vld1q_bf16_x2: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1, d2, d3}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x2.v8bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld1xN, 1 + %0 = bitcast <8 x bfloat> %vld1xN.fca.0.extract to <4 x i32> + %1 = bitcast <8 x bfloat> %vld1xN.fca.1.extract to <4 x i32> + %.fca.0.insert = insertvalue [2 x <4 x i32>] undef, <4 x i32> %0, 0 + %.fca.1.insert = insertvalue [2 x <4 x i32>] %.fca.0.insert, <4 x i32> %1, 1 + ret [2 x <4 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [3 x <2 x i32>] @test_vld1_bf16_x3(bfloat* %ptr) { +; CHECK-LABEL: test_vld1_bf16_x3: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1, d2}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x3.v4bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 1 + %vld1xN.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 2 + %0 = bitcast <4 x bfloat> %vld1xN.fca.0.extract to <2 x i32> + %1 = bitcast <4 x bfloat> %vld1xN.fca.1.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld1xN.fca.2.extract to <2 x i32> + %.fca.0.insert = insertvalue [3 x <2 x i32>] undef, <2 x i32> %0, 0 + %.fca.1.insert = insertvalue [3 x <2 x i32>] %.fca.0.insert, <2 x i32> %1, 1 + %.fca.2.insert = insertvalue [3 x <2 x i32>] %.fca.1.insert, <2 x i32> %2, 2 + ret [3 x <2 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [3 x <4 x i32>] @test_vld1q_bf16_x3(bfloat* %ptr) { +; CHECK-LABEL: test_vld1q_bf16_x3: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1, d2}, [r0:64]! +; CHECK-NEXT: vld1.16 {d3, d4, d5}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x3.v8bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 1 + %vld1xN.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 2 + %0 = bitcast <8 x bfloat> %vld1xN.fca.0.extract to <4 x i32> + %1 = bitcast <8 x bfloat> %vld1xN.fca.1.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld1xN.fca.2.extract to <4 x i32> + %.fca.0.insert = insertvalue [3 x <4 x i32>] undef, <4 x i32> %0, 0 + %.fca.1.insert = insertvalue [3 x <4 x i32>] %.fca.0.insert, <4 x i32> %1, 1 + %.fca.2.insert = insertvalue [3 x <4 x i32>] %.fca.1.insert, <4 x i32> %2, 2 + ret [3 x <4 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [4 x <2 x i32>] @test_vld1_bf16_x4(bfloat* %ptr) { +; CHECK-LABEL: test_vld1_bf16_x4: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1, d2, d3}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x4.v4bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 1 + %vld1xN.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 2 + %vld1xN.fca.3.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld1xN, 3 + %0 = bitcast <4 x bfloat> %vld1xN.fca.0.extract to <2 x i32> + %1 = bitcast <4 x bfloat> %vld1xN.fca.1.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld1xN.fca.2.extract to <2 x i32> + %3 = bitcast <4 x bfloat> %vld1xN.fca.3.extract to <2 x i32> + %.fca.0.insert = insertvalue [4 x <2 x i32>] undef, <2 x i32> %0, 0 + %.fca.1.insert = insertvalue [4 x <2 x i32>] %.fca.0.insert, <2 x i32> %1, 1 + %.fca.2.insert = insertvalue [4 x <2 x i32>] %.fca.1.insert, <2 x i32> %2, 2 + %.fca.3.insert = insertvalue [4 x <2 x i32>] %.fca.2.insert, <2 x i32> %3, 3 + ret [4 x <2 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [4 x <4 x i32>] @test_vld1q_bf16_x4(bfloat* %ptr) { +; CHECK-LABEL: test_vld1q_bf16_x4: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0, d1, d2, d3}, [r0:256]! +; CHECK-NEXT: vld1.16 {d4, d5, d6, d7}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %vld1xN = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x4.v8bf16.p0bf16(bfloat* %ptr) + %vld1xN.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 0 + %vld1xN.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 1 + %vld1xN.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 2 + %vld1xN.fca.3.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld1xN, 3 + %0 = bitcast <8 x bfloat> %vld1xN.fca.0.extract to <4 x i32> + %1 = bitcast <8 x bfloat> %vld1xN.fca.1.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld1xN.fca.2.extract to <4 x i32> + %3 = bitcast <8 x bfloat> %vld1xN.fca.3.extract to <4 x i32> + %.fca.0.insert = insertvalue [4 x <4 x i32>] undef, <4 x i32> %0, 0 + %.fca.1.insert = insertvalue [4 x <4 x i32>] %.fca.0.insert, <4 x i32> %1, 1 + %.fca.2.insert = insertvalue [4 x <4 x i32>] %.fca.1.insert, <4 x i32> %2, 2 + %.fca.3.insert = insertvalue [4 x <4 x i32>] %.fca.2.insert, <4 x i32> %3, 3 + ret [4 x <4 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc <8 x bfloat> @test_vld1q_dup_bf16(bfloat* nocapture readonly %ptr) { +; CHECK-LABEL: test_vld1q_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld1.16 {d0[], d1[]}, [r0:16] +; CHECK-NEXT: bx lr +entry: + %0 = load bfloat, bfloat* %ptr, align 2 + %1 = insertelement <8 x bfloat> undef, bfloat %0, i32 0 + %lane = shufflevector <8 x bfloat> %1, <8 x bfloat> undef, <8 x i32> zeroinitializer + ret <8 x bfloat> %lane +} + +define arm_aapcs_vfpcc [2 x <2 x i32>] @test_vld2_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld2_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld2.16 {d0, d1}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld2_v = tail call { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2.v4bf16.p0i8(i8* %0, i32 2) + %vld2_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_v, 0 + %vld2_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_v, 1 + %1 = bitcast <4 x bfloat> %vld2_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld2_v.fca.1.extract to <2 x i32> + %.fca.0.insert = insertvalue [2 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [2 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + ret [2 x <2 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [2 x <4 x i32>] @test_vld2q_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld2q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld2.16 {d0, d1, d2, d3}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld2q_v = tail call { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2.v8bf16.p0i8(i8* %0, i32 2) + %vld2q_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_v, 0 + %vld2q_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_v, 1 + %1 = bitcast <8 x bfloat> %vld2q_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld2q_v.fca.1.extract to <4 x i32> + %.fca.0.insert = insertvalue [2 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [2 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + ret [2 x <4 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [2 x <2 x i32>] @test_vld2_lane_bf16(bfloat* %ptr, [2 x <2 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld2_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0 def $q0 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0 def $q0 +; CHECK-NEXT: vld2.16 {d0[1], d1[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [2 x <2 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [2 x <2 x i32>] %src.coerce, 1 + %0 = bitcast <2 x i32> %src.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %src.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + %vld2_lane_v = tail call { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2lane.v4bf16.p0i8(i8* %2, <4 x bfloat> %0, <4 x bfloat> %1, i32 1, i32 2) + %vld2_lane_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_lane_v, 0 + %vld2_lane_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_lane_v, 1 + %3 = bitcast <4 x bfloat> %vld2_lane_v.fca.0.extract to <2 x i32> + %4 = bitcast <4 x bfloat> %vld2_lane_v.fca.1.extract to <2 x i32> + %.fca.0.insert = insertvalue [2 x <2 x i32>] undef, <2 x i32> %3, 0 + %.fca.1.insert = insertvalue [2 x <2 x i32>] %.fca.0.insert, <2 x i32> %4, 1 + ret [2 x <2 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [2 x <4 x i32>] @test_vld2q_lane_bf16(bfloat* %ptr, [2 x <4 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld2q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vld2.16 {d1[3], d3[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [2 x <4 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [2 x <4 x i32>] %src.coerce, 1 + %0 = bitcast <4 x i32> %src.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %src.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + %vld2q_lane_v = tail call { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2lane.v8bf16.p0i8(i8* %2, <8 x bfloat> %0, <8 x bfloat> %1, i32 7, i32 2) + %vld2q_lane_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_lane_v, 0 + %vld2q_lane_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_lane_v, 1 + %3 = bitcast <8 x bfloat> %vld2q_lane_v.fca.0.extract to <4 x i32> + %4 = bitcast <8 x bfloat> %vld2q_lane_v.fca.1.extract to <4 x i32> + %.fca.0.insert = insertvalue [2 x <4 x i32>] undef, <4 x i32> %3, 0 + %.fca.1.insert = insertvalue [2 x <4 x i32>] %.fca.0.insert, <4 x i32> %4, 1 + ret [2 x <4 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [3 x <2 x i32>] @test_vld3_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld3_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld3.16 {d0, d1, d2}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld3_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3.v4bf16.p0i8(i8* %0, i32 2) + %vld3_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_v, 0 + %vld3_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_v, 1 + %vld3_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_v, 2 + %1 = bitcast <4 x bfloat> %vld3_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld3_v.fca.1.extract to <2 x i32> + %3 = bitcast <4 x bfloat> %vld3_v.fca.2.extract to <2 x i32> + %.fca.0.insert = insertvalue [3 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [3 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + %.fca.2.insert = insertvalue [3 x <2 x i32>] %.fca.1.insert, <2 x i32> %3, 2 + ret [3 x <2 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [3 x <4 x i32>] @test_vld3q_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld3q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld3.16 {d0, d2, d4}, [r0]! +; CHECK-NEXT: vld3.16 {d1, d3, d5}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld3q_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3.v8bf16.p0i8(i8* %0, i32 2) + %vld3q_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_v, 0 + %vld3q_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_v, 1 + %vld3q_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_v, 2 + %1 = bitcast <8 x bfloat> %vld3q_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld3q_v.fca.1.extract to <4 x i32> + %3 = bitcast <8 x bfloat> %vld3q_v.fca.2.extract to <4 x i32> + %.fca.0.insert = insertvalue [3 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [3 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + %.fca.2.insert = insertvalue [3 x <4 x i32>] %.fca.1.insert, <4 x i32> %3, 2 + ret [3 x <4 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [3 x <2 x i32>] @test_vld3_lane_bf16(bfloat* %ptr, [3 x <2 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld3_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vld3.16 {d0[1], d1[1], d2[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [3 x <2 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [3 x <2 x i32>] %src.coerce, 1 + %src.coerce.fca.2.extract = extractvalue [3 x <2 x i32>] %src.coerce, 2 + %0 = bitcast <2 x i32> %src.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %src.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %src.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + %vld3_lane_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3lane.v4bf16.p0i8(i8* %3, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, i32 1, i32 2) + %vld3_lane_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_lane_v, 0 + %vld3_lane_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_lane_v, 1 + %vld3_lane_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_lane_v, 2 + %4 = bitcast <4 x bfloat> %vld3_lane_v.fca.0.extract to <2 x i32> + %5 = bitcast <4 x bfloat> %vld3_lane_v.fca.1.extract to <2 x i32> + %6 = bitcast <4 x bfloat> %vld3_lane_v.fca.2.extract to <2 x i32> + %.fca.0.insert = insertvalue [3 x <2 x i32>] undef, <2 x i32> %4, 0 + %.fca.1.insert = insertvalue [3 x <2 x i32>] %.fca.0.insert, <2 x i32> %5, 1 + %.fca.2.insert = insertvalue [3 x <2 x i32>] %.fca.1.insert, <2 x i32> %6, 2 + ret [3 x <2 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [3 x <4 x i32>] @test_vld3q_lane_bf16(bfloat* %ptr, [3 x <4 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld3q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vld3.16 {d1[3], d3[3], d5[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [3 x <4 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [3 x <4 x i32>] %src.coerce, 1 + %src.coerce.fca.2.extract = extractvalue [3 x <4 x i32>] %src.coerce, 2 + %0 = bitcast <4 x i32> %src.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %src.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %src.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + %vld3q_lane_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3lane.v8bf16.p0i8(i8* %3, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, i32 7, i32 2) + %vld3q_lane_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_lane_v, 0 + %vld3q_lane_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_lane_v, 1 + %vld3q_lane_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_lane_v, 2 + %4 = bitcast <8 x bfloat> %vld3q_lane_v.fca.0.extract to <4 x i32> + %5 = bitcast <8 x bfloat> %vld3q_lane_v.fca.1.extract to <4 x i32> + %6 = bitcast <8 x bfloat> %vld3q_lane_v.fca.2.extract to <4 x i32> + %.fca.0.insert = insertvalue [3 x <4 x i32>] undef, <4 x i32> %4, 0 + %.fca.1.insert = insertvalue [3 x <4 x i32>] %.fca.0.insert, <4 x i32> %5, 1 + %.fca.2.insert = insertvalue [3 x <4 x i32>] %.fca.1.insert, <4 x i32> %6, 2 + ret [3 x <4 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [4 x <2 x i32>] @test_vld4_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld4_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld4.16 {d0, d1, d2, d3}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld4_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4.v4bf16.p0i8(i8* %0, i32 2) + %vld4_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_v, 0 + %vld4_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_v, 1 + %vld4_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_v, 2 + %vld4_v.fca.3.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_v, 3 + %1 = bitcast <4 x bfloat> %vld4_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld4_v.fca.1.extract to <2 x i32> + %3 = bitcast <4 x bfloat> %vld4_v.fca.2.extract to <2 x i32> + %4 = bitcast <4 x bfloat> %vld4_v.fca.3.extract to <2 x i32> + %.fca.0.insert = insertvalue [4 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [4 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + %.fca.2.insert = insertvalue [4 x <2 x i32>] %.fca.1.insert, <2 x i32> %3, 2 + %.fca.3.insert = insertvalue [4 x <2 x i32>] %.fca.2.insert, <2 x i32> %4, 3 + ret [4 x <2 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [4 x <4 x i32>] @test_vld4q_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld4q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld4.16 {d0, d2, d4, d6}, [r0]! +; CHECK-NEXT: vld4.16 {d1, d3, d5, d7}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld4q_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4.v8bf16.p0i8(i8* %0, i32 2) + %vld4q_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_v, 0 + %vld4q_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_v, 1 + %vld4q_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_v, 2 + %vld4q_v.fca.3.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_v, 3 + %1 = bitcast <8 x bfloat> %vld4q_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld4q_v.fca.1.extract to <4 x i32> + %3 = bitcast <8 x bfloat> %vld4q_v.fca.2.extract to <4 x i32> + %4 = bitcast <8 x bfloat> %vld4q_v.fca.3.extract to <4 x i32> + %.fca.0.insert = insertvalue [4 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [4 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + %.fca.2.insert = insertvalue [4 x <4 x i32>] %.fca.1.insert, <4 x i32> %3, 2 + %.fca.3.insert = insertvalue [4 x <4 x i32>] %.fca.2.insert, <4 x i32> %4, 3 + ret [4 x <4 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [4 x <2 x i32>] @test_vld4_lane_bf16(bfloat* %ptr, [4 x <2 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld4_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d3 killed $d3 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vld4.16 {d0[1], d1[1], d2[1], d3[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [4 x <2 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [4 x <2 x i32>] %src.coerce, 1 + %src.coerce.fca.2.extract = extractvalue [4 x <2 x i32>] %src.coerce, 2 + %src.coerce.fca.3.extract = extractvalue [4 x <2 x i32>] %src.coerce, 3 + %0 = bitcast <2 x i32> %src.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %src.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %src.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast <2 x i32> %src.coerce.fca.3.extract to <4 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + %vld4_lane_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4lane.v4bf16.p0i8(i8* %4, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, <4 x bfloat> %3, i32 1, i32 2) + %vld4_lane_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_lane_v, 0 + %vld4_lane_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_lane_v, 1 + %vld4_lane_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_lane_v, 2 + %vld4_lane_v.fca.3.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_lane_v, 3 + %5 = bitcast <4 x bfloat> %vld4_lane_v.fca.0.extract to <2 x i32> + %6 = bitcast <4 x bfloat> %vld4_lane_v.fca.1.extract to <2 x i32> + %7 = bitcast <4 x bfloat> %vld4_lane_v.fca.2.extract to <2 x i32> + %8 = bitcast <4 x bfloat> %vld4_lane_v.fca.3.extract to <2 x i32> + %.fca.0.insert = insertvalue [4 x <2 x i32>] undef, <2 x i32> %5, 0 + %.fca.1.insert = insertvalue [4 x <2 x i32>] %.fca.0.insert, <2 x i32> %6, 1 + %.fca.2.insert = insertvalue [4 x <2 x i32>] %.fca.1.insert, <2 x i32> %7, 2 + %.fca.3.insert = insertvalue [4 x <2 x i32>] %.fca.2.insert, <2 x i32> %8, 3 + ret [4 x <2 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [4 x <4 x i32>] @test_vld4q_lane_bf16(bfloat* %ptr, [4 x <4 x i32>] %src.coerce) { +; CHECK-LABEL: test_vld4q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q3 killed $q3 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vld4.16 {d1[3], d3[3], d5[3], d7[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %src.coerce.fca.0.extract = extractvalue [4 x <4 x i32>] %src.coerce, 0 + %src.coerce.fca.1.extract = extractvalue [4 x <4 x i32>] %src.coerce, 1 + %src.coerce.fca.2.extract = extractvalue [4 x <4 x i32>] %src.coerce, 2 + %src.coerce.fca.3.extract = extractvalue [4 x <4 x i32>] %src.coerce, 3 + %0 = bitcast <4 x i32> %src.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %src.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %src.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast <4 x i32> %src.coerce.fca.3.extract to <8 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + %vld4q_lane_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4lane.v8bf16.p0i8(i8* %4, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, <8 x bfloat> %3, i32 7, i32 2) + %vld4q_lane_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_lane_v, 0 + %vld4q_lane_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_lane_v, 1 + %vld4q_lane_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_lane_v, 2 + %vld4q_lane_v.fca.3.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_lane_v, 3 + %5 = bitcast <8 x bfloat> %vld4q_lane_v.fca.0.extract to <4 x i32> + %6 = bitcast <8 x bfloat> %vld4q_lane_v.fca.1.extract to <4 x i32> + %7 = bitcast <8 x bfloat> %vld4q_lane_v.fca.2.extract to <4 x i32> + %8 = bitcast <8 x bfloat> %vld4q_lane_v.fca.3.extract to <4 x i32> + %.fca.0.insert = insertvalue [4 x <4 x i32>] undef, <4 x i32> %5, 0 + %.fca.1.insert = insertvalue [4 x <4 x i32>] %.fca.0.insert, <4 x i32> %6, 1 + %.fca.2.insert = insertvalue [4 x <4 x i32>] %.fca.1.insert, <4 x i32> %7, 2 + %.fca.3.insert = insertvalue [4 x <4 x i32>] %.fca.2.insert, <4 x i32> %8, 3 + ret [4 x <4 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [2 x <2 x i32>] @test_vld2_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld2_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld2.16 {d0[], d1[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld2_dup_v = tail call { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2dup.v4bf16.p0i8(i8* %0, i32 2) + %vld2_dup_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_dup_v, 0 + %vld2_dup_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat> } %vld2_dup_v, 1 + %1 = bitcast <4 x bfloat> %vld2_dup_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld2_dup_v.fca.1.extract to <2 x i32> + %.fca.0.insert = insertvalue [2 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [2 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + ret [2 x <2 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [2 x <4 x i32>] @test_vld2q_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld2q_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld2.16 {d16[], d18[]}, [r0] +; CHECK-NEXT: vld2.16 {d1[], d3[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld2q_dup_v = tail call { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2dup.v8bf16.p0i8(i8* %0, i32 2) + %vld2q_dup_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_dup_v, 0 + %vld2q_dup_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat> } %vld2q_dup_v, 1 + %1 = bitcast <8 x bfloat> %vld2q_dup_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld2q_dup_v.fca.1.extract to <4 x i32> + %.fca.0.insert = insertvalue [2 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [2 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + ret [2 x <4 x i32>] %.fca.1.insert +} + +define arm_aapcs_vfpcc [3 x <2 x i32>] @test_vld3_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld3_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld3.16 {d0[], d1[], d2[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld3_dup_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3dup.v4bf16.p0i8(i8* %0, i32 2) + %vld3_dup_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_dup_v, 0 + %vld3_dup_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_dup_v, 1 + %vld3_dup_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld3_dup_v, 2 + %1 = bitcast <4 x bfloat> %vld3_dup_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld3_dup_v.fca.1.extract to <2 x i32> + %3 = bitcast <4 x bfloat> %vld3_dup_v.fca.2.extract to <2 x i32> + %.fca.0.insert = insertvalue [3 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [3 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + %.fca.2.insert = insertvalue [3 x <2 x i32>] %.fca.1.insert, <2 x i32> %3, 2 + ret [3 x <2 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [3 x <4 x i32>] @test_vld3q_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld3q_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld3.16 {d0[], d2[], d4[]}, [r0] +; CHECK-NEXT: vld3.16 {d1[], d3[], d5[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld3q_dup_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3dup.v8bf16.p0i8(i8* %0, i32 2) + %vld3q_dup_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_dup_v, 0 + %vld3q_dup_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_dup_v, 1 + %vld3q_dup_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld3q_dup_v, 2 + %1 = bitcast <8 x bfloat> %vld3q_dup_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld3q_dup_v.fca.1.extract to <4 x i32> + %3 = bitcast <8 x bfloat> %vld3q_dup_v.fca.2.extract to <4 x i32> + %.fca.0.insert = insertvalue [3 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [3 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + %.fca.2.insert = insertvalue [3 x <4 x i32>] %.fca.1.insert, <4 x i32> %3, 2 + ret [3 x <4 x i32>] %.fca.2.insert +} + +define arm_aapcs_vfpcc [4 x <2 x i32>] @test_vld4_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld4_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld4.16 {d0[], d1[], d2[], d3[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld4_dup_v = tail call { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4dup.v4bf16.p0i8(i8* %0, i32 2) + %vld4_dup_v.fca.0.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_dup_v, 0 + %vld4_dup_v.fca.1.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_dup_v, 1 + %vld4_dup_v.fca.2.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_dup_v, 2 + %vld4_dup_v.fca.3.extract = extractvalue { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } %vld4_dup_v, 3 + %1 = bitcast <4 x bfloat> %vld4_dup_v.fca.0.extract to <2 x i32> + %2 = bitcast <4 x bfloat> %vld4_dup_v.fca.1.extract to <2 x i32> + %3 = bitcast <4 x bfloat> %vld4_dup_v.fca.2.extract to <2 x i32> + %4 = bitcast <4 x bfloat> %vld4_dup_v.fca.3.extract to <2 x i32> + %.fca.0.insert = insertvalue [4 x <2 x i32>] undef, <2 x i32> %1, 0 + %.fca.1.insert = insertvalue [4 x <2 x i32>] %.fca.0.insert, <2 x i32> %2, 1 + %.fca.2.insert = insertvalue [4 x <2 x i32>] %.fca.1.insert, <2 x i32> %3, 2 + %.fca.3.insert = insertvalue [4 x <2 x i32>] %.fca.2.insert, <2 x i32> %4, 3 + ret [4 x <2 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc [4 x <4 x i32>] @test_vld4q_dup_bf16(bfloat* %ptr) { +; CHECK-LABEL: test_vld4q_dup_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vld4.16 {d0[], d2[], d4[], d6[]}, [r0] +; CHECK-NEXT: vld4.16 {d1[], d3[], d5[], d7[]}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + %vld4q_dup_v = tail call { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4dup.v8bf16.p0i8(i8* %0, i32 2) + %vld4q_dup_v.fca.0.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_dup_v, 0 + %vld4q_dup_v.fca.1.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_dup_v, 1 + %vld4q_dup_v.fca.2.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_dup_v, 2 + %vld4q_dup_v.fca.3.extract = extractvalue { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } %vld4q_dup_v, 3 + %1 = bitcast <8 x bfloat> %vld4q_dup_v.fca.0.extract to <4 x i32> + %2 = bitcast <8 x bfloat> %vld4q_dup_v.fca.1.extract to <4 x i32> + %3 = bitcast <8 x bfloat> %vld4q_dup_v.fca.2.extract to <4 x i32> + %4 = bitcast <8 x bfloat> %vld4q_dup_v.fca.3.extract to <4 x i32> + %.fca.0.insert = insertvalue [4 x <4 x i32>] undef, <4 x i32> %1, 0 + %.fca.1.insert = insertvalue [4 x <4 x i32>] %.fca.0.insert, <4 x i32> %2, 1 + %.fca.2.insert = insertvalue [4 x <4 x i32>] %.fca.1.insert, <4 x i32> %3, 2 + %.fca.3.insert = insertvalue [4 x <4 x i32>] %.fca.2.insert, <4 x i32> %4, 3 + ret [4 x <4 x i32>] %.fca.3.insert +} + +define arm_aapcs_vfpcc void @test_vst1_bf16(bfloat* %ptr, <4 x bfloat> %val) { +; CHECK-LABEL: test_vst1_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vst1.16 {d0}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst1.p0i8.v4bf16(i8* %0, <4 x bfloat> %val, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1q_bf16(bfloat* %ptr, <8 x bfloat> %val) { +; CHECK-LABEL: test_vst1q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vst1.16 {d0, d1}, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst1.p0i8.v8bf16(i8* %0, <8 x bfloat> %val, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1_lane_bf16(bfloat* nocapture %ptr, <4 x bfloat> %val) { +; CHECK-LABEL: test_vst1_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vmovx.f16 s0, s0 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = extractelement <4 x bfloat> %val, i32 1 + store bfloat %0, bfloat* %ptr, align 2 + ret void +} + +define arm_aapcs_vfpcc void @test_vst1q_lane_bf16(bfloat* nocapture %ptr, <8 x bfloat> %val) { +; CHECK-LABEL: test_vst1q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: vmovx.f16 s0, s3 +; CHECK-NEXT: vstr.16 s0, [r0] +; CHECK-NEXT: bx lr +entry: + %0 = extractelement <8 x bfloat> %val, i32 7 + store bfloat %0, bfloat* %ptr, align 2 + ret void +} + +define arm_aapcs_vfpcc void @test_vst1_bf16_x2(bfloat* nocapture %ptr, [2 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1_bf16_x2: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0 def $q0 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0 def $q0 +; CHECK-NEXT: vst1.16 {d0, d1}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <2 x i32>] %val.coerce, 1 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + tail call void @llvm.arm.neon.vst1x2.p0bf16.v4bf16(bfloat* %ptr, <4 x bfloat> %0, <4 x bfloat> %1) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1q_bf16_x2(bfloat* nocapture %ptr, [2 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1q_bf16_x2: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst1.16 {d0, d1, d2, d3}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <4 x i32>] %val.coerce, 1 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + tail call void @llvm.arm.neon.vst1x2.p0bf16.v8bf16(bfloat* %ptr, <8 x bfloat> %0, <8 x bfloat> %1) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1_bf16_x3(bfloat* nocapture %ptr, [3 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1_bf16_x3: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst1.16 {d0, d1, d2}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <2 x i32>] %val.coerce, 2 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + tail call void @llvm.arm.neon.vst1x3.p0bf16.v4bf16(bfloat* %ptr, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1q_bf16_x3(bfloat* nocapture %ptr, [3 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1q_bf16_x3: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst1.16 {d0, d1, d2}, [r0:64]! +; CHECK-NEXT: vst1.16 {d3, d4, d5}, [r0:64] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <4 x i32>] %val.coerce, 2 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + tail call void @llvm.arm.neon.vst1x3.p0bf16.v8bf16(bfloat* %ptr, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1_bf16_x4(bfloat* nocapture %ptr, [4 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1_bf16_x4: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d3 killed $d3 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst1.16 {d0, d1, d2, d3}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <2 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <2 x i32>] %val.coerce, 3 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast <2 x i32> %val.coerce.fca.3.extract to <4 x bfloat> + tail call void @llvm.arm.neon.vst1x4.p0bf16.v4bf16(bfloat* %ptr, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, <4 x bfloat> %3) + ret void +} + +define arm_aapcs_vfpcc void @test_vst1q_bf16_x4(bfloat* nocapture %ptr, [4 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst1q_bf16_x4: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q3 killed $q3 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst1.16 {d0, d1, d2, d3}, [r0:256]! +; CHECK-NEXT: vst1.16 {d4, d5, d6, d7}, [r0:256] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <4 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <4 x i32>] %val.coerce, 3 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast <4 x i32> %val.coerce.fca.3.extract to <8 x bfloat> + tail call void @llvm.arm.neon.vst1x4.p0bf16.v8bf16(bfloat* %ptr, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, <8 x bfloat> %3) + ret void +} + +define arm_aapcs_vfpcc void @test_vst2_bf16(bfloat* %ptr, [2 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst2_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0 def $q0 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0 def $q0 +; CHECK-NEXT: vst2.16 {d0, d1}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <2 x i32>] %val.coerce, 1 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst2.p0i8.v4bf16(i8* %2, <4 x bfloat> %0, <4 x bfloat> %1, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst2q_bf16(bfloat* %ptr, [2 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst2q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst2.16 {d0, d1, d2, d3}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <4 x i32>] %val.coerce, 1 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst2.p0i8.v8bf16(i8* %2, <8 x bfloat> %0, <8 x bfloat> %1, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst2_lane_bf16(bfloat* %ptr, [2 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst2_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0 def $q0 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0 def $q0 +; CHECK-NEXT: vst2.16 {d0[1], d1[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <2 x i32>] %val.coerce, 1 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst2lane.p0i8.v4bf16(i8* %2, <4 x bfloat> %0, <4 x bfloat> %1, i32 1, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst2q_lane_bf16(bfloat* %ptr, [2 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst2q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst2.16 {d1[3], d3[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [2 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [2 x <4 x i32>] %val.coerce, 1 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst2lane.p0i8.v8bf16(i8* %2, <8 x bfloat> %0, <8 x bfloat> %1, i32 7, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst3_bf16(bfloat* %ptr, [3 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst3_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst3.16 {d0, d1, d2}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <2 x i32>] %val.coerce, 2 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst3.p0i8.v4bf16(i8* %3, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst3q_bf16(bfloat* %ptr, [3 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst3q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst3.16 {d0, d2, d4}, [r0]! +; CHECK-NEXT: vst3.16 {d1, d3, d5}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <4 x i32>] %val.coerce, 2 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst3.p0i8.v8bf16(i8* %3, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst3_lane_bf16(bfloat* %ptr, [3 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst3_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst3.16 {d0[1], d1[1], d2[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <2 x i32>] %val.coerce, 2 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst3lane.p0i8.v4bf16(i8* %3, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, i32 1, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst3q_lane_bf16(bfloat* %ptr, [3 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst3q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst3.16 {d1[3], d3[3], d5[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [3 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [3 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [3 x <4 x i32>] %val.coerce, 2 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst3lane.p0i8.v8bf16(i8* %3, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, i32 7, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst4_bf16(bfloat* %ptr, [4 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst4_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d3 killed $d3 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst4.16 {d0, d1, d2, d3}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <2 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <2 x i32>] %val.coerce, 3 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast <2 x i32> %val.coerce.fca.3.extract to <4 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst4.p0i8.v4bf16(i8* %4, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, <4 x bfloat> %3, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst4q_bf16(bfloat* %ptr, [4 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst4q_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q3 killed $q3 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst4.16 {d0, d2, d4, d6}, [r0]! +; CHECK-NEXT: vst4.16 {d1, d3, d5, d7}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <4 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <4 x i32>] %val.coerce, 3 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast <4 x i32> %val.coerce.fca.3.extract to <8 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst4.p0i8.v8bf16(i8* %4, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, <8 x bfloat> %3, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst4_lane_bf16(bfloat* %ptr, [4 x <2 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst4_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $d3 killed $d3 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d2 killed $d2 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d1 killed $d1 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: @ kill: def $d0 killed $d0 killed $q0_q1 def $q0_q1 +; CHECK-NEXT: vst4.16 {d0[1], d1[1], d2[1], d3[1]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <2 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <2 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <2 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <2 x i32>] %val.coerce, 3 + %0 = bitcast <2 x i32> %val.coerce.fca.0.extract to <4 x bfloat> + %1 = bitcast <2 x i32> %val.coerce.fca.1.extract to <4 x bfloat> + %2 = bitcast <2 x i32> %val.coerce.fca.2.extract to <4 x bfloat> + %3 = bitcast <2 x i32> %val.coerce.fca.3.extract to <4 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst4lane.p0i8.v4bf16(i8* %4, <4 x bfloat> %0, <4 x bfloat> %1, <4 x bfloat> %2, <4 x bfloat> %3, i32 1, i32 2) + ret void +} + +define arm_aapcs_vfpcc void @test_vst4q_lane_bf16(bfloat* %ptr, [4 x <4 x i32>] %val.coerce) { +; CHECK-LABEL: test_vst4q_lane_bf16: +; CHECK: @ %bb.0: @ %entry +; CHECK-NEXT: @ kill: def $q3 killed $q3 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q2 killed $q2 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q1 killed $q1 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: @ kill: def $q0 killed $q0 killed $q0_q1_q2_q3 def $q0_q1_q2_q3 +; CHECK-NEXT: vst4.16 {d1[3], d3[3], d5[3], d7[3]}, [r0] +; CHECK-NEXT: bx lr +entry: + %val.coerce.fca.0.extract = extractvalue [4 x <4 x i32>] %val.coerce, 0 + %val.coerce.fca.1.extract = extractvalue [4 x <4 x i32>] %val.coerce, 1 + %val.coerce.fca.2.extract = extractvalue [4 x <4 x i32>] %val.coerce, 2 + %val.coerce.fca.3.extract = extractvalue [4 x <4 x i32>] %val.coerce, 3 + %0 = bitcast <4 x i32> %val.coerce.fca.0.extract to <8 x bfloat> + %1 = bitcast <4 x i32> %val.coerce.fca.1.extract to <8 x bfloat> + %2 = bitcast <4 x i32> %val.coerce.fca.2.extract to <8 x bfloat> + %3 = bitcast <4 x i32> %val.coerce.fca.3.extract to <8 x bfloat> + %4 = bitcast bfloat* %ptr to i8* + tail call void @llvm.arm.neon.vst4lane.p0i8.v8bf16(i8* %4, <8 x bfloat> %0, <8 x bfloat> %1, <8 x bfloat> %2, <8 x bfloat> %3, i32 7, i32 2) + ret void +} + +declare { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2.v8bf16.p0i8(i8*, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3.v8bf16.p0i8(i8*, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4.v8bf16.p0i8(i8*, i32) + +declare { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2dup.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2dup.v8bf16.p0i8(i8*, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3dup.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3dup.v8bf16.p0i8(i8*, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4dup.v4bf16.p0i8(i8*, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4dup.v8bf16.p0i8(i8*, i32) + +declare { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x2.v4bf16.p0bf16(bfloat*) +declare { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x2.v8bf16.p0bf16(bfloat*) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x3.v4bf16.p0bf16(bfloat*) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x3.v8bf16.p0bf16(bfloat*) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld1x4.v4bf16.p0bf16(bfloat*) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld1x4.v8bf16.p0bf16(bfloat*) + +declare { <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld2lane.v4bf16.p0i8(i8*, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare { <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld2lane.v8bf16.p0i8(i8*, <8 x bfloat>, <8 x bfloat>, i32, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld3lane.v4bf16.p0i8(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld3lane.v8bf16.p0i8(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32, i32) +declare { <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat> } @llvm.arm.neon.vld4lane.v4bf16.p0i8(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare { <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat> } @llvm.arm.neon.vld4lane.v8bf16.p0i8(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32, i32) + +declare void @llvm.arm.neon.vst1.p0i8.v4bf16(i8*, <4 x bfloat>, i32) +declare void @llvm.arm.neon.vst1.p0i8.v8bf16(i8*, <8 x bfloat>, i32) +declare void @llvm.arm.neon.vst2.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, i32) +declare void @llvm.arm.neon.vst2.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, i32) +declare void @llvm.arm.neon.vst3.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32) +declare void @llvm.arm.neon.vst3.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32) +declare void @llvm.arm.neon.vst4.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32) +declare void @llvm.arm.neon.vst4.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32) + +declare void @llvm.arm.neon.vst1x2.p0bf16.v4bf16(bfloat* nocapture, <4 x bfloat>, <4 x bfloat>) +declare void @llvm.arm.neon.vst1x2.p0bf16.v8bf16(bfloat* nocapture, <8 x bfloat>, <8 x bfloat>) +declare void @llvm.arm.neon.vst1x3.p0bf16.v4bf16(bfloat* nocapture, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>) +declare void @llvm.arm.neon.vst1x3.p0bf16.v8bf16(bfloat* nocapture, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>) +declare void @llvm.arm.neon.vst1x4.p0bf16.v4bf16(bfloat* nocapture, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>) +declare void @llvm.arm.neon.vst1x4.p0bf16.v8bf16(bfloat* nocapture, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>) + +declare void @llvm.arm.neon.vst2lane.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare void @llvm.arm.neon.vst2lane.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, i32, i32) +declare void @llvm.arm.neon.vst3lane.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare void @llvm.arm.neon.vst3lane.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32, i32) +declare void @llvm.arm.neon.vst4lane.p0i8.v4bf16(i8*, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, <4 x bfloat>, i32, i32) +declare void @llvm.arm.neon.vst4lane.p0i8.v8bf16(i8*, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, <8 x bfloat>, i32, i32)