Index: llvm/lib/Analysis/Loads.cpp =================================================================== --- llvm/lib/Analysis/Loads.cpp +++ llvm/lib/Analysis/Loads.cpp @@ -383,10 +383,6 @@ MaxInstsToScan = ~0U; const DataLayout &DL = ScanBB->getModule()->getDataLayout(); - - // Try to get the store size for the type. - auto AccessSize = LocationSize::precise(DL.getTypeStoreSize(AccessTy)); - Value *StrippedPtr = Ptr->stripPointerCasts(); while (ScanFrom != ScanBB->begin()) { @@ -425,6 +421,9 @@ return LI; } + // Try to get the store size for the type. + auto AccessSize = LocationSize::precise(DL.getTypeStoreSize(AccessTy)); + if (StoreInst *SI = dyn_cast(Inst)) { Value *StorePtr = SI->getPointerOperand()->stripPointerCasts(); // If this is a store through Ptr, the value is available! Index: llvm/lib/Transforms/InstCombine/InstCombineLoadStoreAlloca.cpp =================================================================== --- llvm/lib/Transforms/InstCombine/InstCombineLoadStoreAlloca.cpp +++ llvm/lib/Transforms/InstCombine/InstCombineLoadStoreAlloca.cpp @@ -586,6 +586,7 @@ // Do not perform canonicalization if minmax pattern is found (to avoid // infinite loop). if (!Ty->isIntegerTy() && Ty->isSized() && + !(Ty->isVectorTy() && Ty->getVectorIsScalable()) && DL.isLegalInteger(DL.getTypeStoreSizeInBits(Ty)) && DL.typeSizeEqualsStoreSize(Ty) && !DL.isNonIntegralPointerType(Ty) && Index: llvm/test/Transforms/InstCombine/load.ll =================================================================== --- llvm/test/Transforms/InstCombine/load.ll +++ llvm/test/Transforms/InstCombine/load.ll @@ -237,6 +237,42 @@ ret void } +define void @test16-vect(i8* %x, i8* %a, i8* %b, i8* %c) { +; CHECK-LABEL: @test16-vect( +; CHECK-NEXT: entry: +; CHECK-NEXT: [[C_CAST:%.*]] = bitcast i8* [[C:%.*]] to i32* +; CHECK-NEXT: [[TMP0:%.*]] = bitcast i8* [[X:%.*]] to i32* +; CHECK-NEXT: [[X11:%.*]] = load i32, i32* [[TMP0]], align 4 +; CHECK-NEXT: [[TMP1:%.*]] = bitcast i8* [[A:%.*]] to i32* +; CHECK-NEXT: store i32 [[X11]], i32* [[TMP1]], align 4 +; CHECK-NEXT: [[TMP2:%.*]] = bitcast i8* [[B:%.*]] to i32* +; CHECK-NEXT: store i32 [[X11]], i32* [[TMP2]], align 4 +; CHECK-NEXT: [[TMP3:%.*]] = bitcast i8* [[X]] to i32* +; CHECK-NEXT: [[X22:%.*]] = load i32, i32* [[TMP3]], align 4 +; CHECK-NEXT: [[TMP4:%.*]] = bitcast i8* [[B]] to i32* +; CHECK-NEXT: store i32 [[X22]], i32* [[TMP4]], align 4 +; CHECK-NEXT: store i32 [[X22]], i32* [[C_CAST]], align 4 +; CHECK-NEXT: ret void +; +entry: + %x.cast = bitcast i8* %x to <4 x i8>* + %a.cast = bitcast i8* %a to <4 x i8>* + %b.cast = bitcast i8* %b to <4 x i8>* + %c.cast = bitcast i8* %c to i32* + + %x1 = load <4 x i8>, <4 x i8>* %x.cast + store <4 x i8> %x1, <4 x i8>* %a.cast + store <4 x i8> %x1, <4 x i8>* %b.cast + + %x2 = load <4 x i8>, <4 x i8>* %x.cast + store <4 x i8> %x2, <4 x i8>* %b.cast + %x2.cast = bitcast <4 x i8> %x2 to i32 + store i32 %x2.cast, i32* %c.cast + + ret void +} + + ; Check that in cases similar to @test16 we don't try to rewrite a load when ; its only use is a store but it is used as the pointer to that store rather ; than the value. @@ -300,3 +336,10 @@ store %swift.error* %err.res, %swift.error** %err, align 8 ret void } + +; Make sure we don't canonicalize accesses to scalable vectors. +define void @test20(* %addr) { + %1 = load , * %addr, align 1 + store %1, * %addr, align 1 + ret void +}