diff --git a/llvm/lib/Transforms/Utils/ScalarEvolutionExpander.cpp b/llvm/lib/Transforms/Utils/ScalarEvolutionExpander.cpp --- a/llvm/lib/Transforms/Utils/ScalarEvolutionExpander.cpp +++ b/llvm/lib/Transforms/Utils/ScalarEvolutionExpander.cpp @@ -287,14 +287,14 @@ if (const SCEVMulExpr *M = dyn_cast(S)) { // Size is known, check if there is a constant operand which is a multiple // of the given factor. If so, we can factor it. - const SCEVConstant *FC = cast(Factor); - if (const SCEVConstant *C = dyn_cast(M->getOperand(0))) - if (!C->getAPInt().srem(FC->getAPInt())) { - SmallVector NewMulOps(M->op_begin(), M->op_end()); - NewMulOps[0] = SE.getConstant(C->getAPInt().sdiv(FC->getAPInt())); - S = SE.getMulExpr(NewMulOps); - return true; - } + if (const SCEVConstant *FC = dyn_cast(Factor)) + if (const SCEVConstant *C = dyn_cast(M->getOperand(0))) + if (!C->getAPInt().srem(FC->getAPInt())) { + SmallVector NewMulOps(M->op_begin(), M->op_end()); + NewMulOps[0] = SE.getConstant(C->getAPInt().sdiv(FC->getAPInt())); + S = SE.getMulExpr(NewMulOps); + return true; + } } // In an AddRec, check if both start and step are divisible. diff --git a/llvm/test/Transforms/LoopStrengthReduce/vscale-factor-out-constant.ll b/llvm/test/Transforms/LoopStrengthReduce/vscale-factor-out-constant.ll new file mode 100644 --- /dev/null +++ b/llvm/test/Transforms/LoopStrengthReduce/vscale-factor-out-constant.ll @@ -0,0 +1,46 @@ +; RUN: opt -S -loop-reduce < %s | FileCheck %s + +target datalayout = "e-m:e-i8:8:32-i16:16:32-i64:64-i128:128-n32:64-S128" +target triple = "aarch64-unknown-linux-gnu" + +; This test check SCEVExpander FactorOutConstant() is not crashing with blind cast Factor to SCEVConstant. + +; CHECK-LABEL: test +define void @test(i32* %a, i32 %v, i64 %n) { +entry: + %scalar_vector = alloca , align 16 + %num_elm = call i64 @llvm.aarch64.sve.cntw(i32 31) + %scalar_count = and i64 %num_elm, -4 + br label %loop_header + +exit: + ret void + +loop_header: + %indvar = phi i64 [ 0, %entry ], [ %indvar_next, %for_loop ] + %gep_a_0 = getelementptr i32, i32* %a, i64 0 + %gep_vec_0 = getelementptr inbounds , * %scalar_vector, i64 0, i64 0 + br label %scalar_loop + +scalar_loop: + %gep_vec = phi i32* [ %gep_vec_0, %loop_header ], [ %gep_vec_inc, %scalar_loop ] + %scalar_iv = phi i64 [ 0, %loop_header ], [ %scalar_iv_next, %scalar_loop ] + store i32 %v, i32* %gep_vec, align 4 + %scalar_iv_next = add i64 %scalar_iv, 1 + %gep_vec_inc = getelementptr i32, i32* %gep_vec, i64 1 + %scalar_exit = icmp eq i64 %scalar_iv_next, %scalar_count + br i1 %scalar_exit, label %for_loop, label %scalar_loop + +for_loop: + %vector = load , * %scalar_vector, align 16 + %gep_a = getelementptr i32, i32* %gep_a_0, i64 %indvar + %vector_ptr = bitcast i32* %gep_a to * + call void @llvm.masked.store.nxv4i32.p0nxv4i32( %vector, * %vector_ptr, i32 4, undef) + %indvar_next = add nsw i64 %indvar, %scalar_count + %exit_cond = icmp eq i64 %indvar_next, %n + br i1 %exit_cond, label %exit, label %loop_header +} + +declare i64 @llvm.aarch64.sve.cntw(i32 immarg) + +declare void @llvm.masked.store.nxv4i32.p0nxv4i32(, *, i32 immarg, )