Index: llvm/test/Transforms/LoopVectorize/AArch64/sve-inv-loads.ll =================================================================== --- /dev/null +++ llvm/test/Transforms/LoopVectorize/AArch64/sve-inv-loads.ll @@ -0,0 +1,36 @@ +; RUN: opt -S -loop-vectorize -mattr=+sve -mtriple aarch64 < %s | FileCheck %s + +define void @invariant_load(i64 %n, i32* noalias nocapture %a, i32* nocapture readonly %b) { +; CHECK-LABEL: @invariant_load +; CHECK: vector.body: +; CHECK: %[[GEP:.*]] = getelementptr inbounds i32, i32* %b, i64 42 +; CHECK: %[[INVLOAD:.*]] = load i32, i32* %[[GEP]] +; CHECK: %[[SPLATINS:.*]] = insertelement poison, i32 %[[INVLOAD]], i32 0 +; CHECK: %[[SPLAT:.*]] = shufflevector %[[SPLATINS]], poison, zeroinitializer +; CHECK: %[[LOAD:.*]] = load , * +; CHECK: %[[ADD:.*]] = add nsw %[[SPLAT]], %[[LOAD]] +; CHECK: store %[[ADD]], * +entry: + br label %for.body + +for.body: ; preds = %for.body.lr.ph, %for.body + %iv = phi i64 [ 0, %entry ], [ %iv.next, %for.body ] + %arrayidx = getelementptr inbounds i32, i32* %b, i64 42 + %0 = load i32, i32* %arrayidx, align 4 + %arrayidx1 = getelementptr inbounds i32, i32* %b, i64 %iv + %1 = load i32, i32* %arrayidx1, align 4 + %add = add nsw i32 %0, %1 + %arrayidx2 = getelementptr inbounds i32, i32* %a, i64 %iv + store i32 %add, i32* %arrayidx2, align 4 + %iv.next = add nuw nsw i64 %iv, 1 + %exitcond.not = icmp eq i64 %iv.next, %n + br i1 %exitcond.not, label %for.end, label %for.body, !llvm.loop !1 + +for.end: ; preds = %for.body + ret void +} + +!1 = distinct !{!1, !2, !3, !4} +!2 = !{!"llvm.loop.vectorize.width", i32 4} +!3 = !{!"llvm.loop.interleave.count", i32 1} +!4 = !{!"llvm.loop.vectorize.scalable.enable", i1 true}