Index: llvm/trunk/lib/IR/ConstantFold.cpp =================================================================== --- llvm/trunk/lib/IR/ConstantFold.cpp +++ llvm/trunk/lib/IR/ConstantFold.cpp @@ -2193,7 +2193,7 @@ for (unsigned i = 1, e = Idxs.size(); i != e; Prev = Ty, Ty = cast(Ty)->getTypeAtIndex(Idxs[i]), ++i) { if (ConstantInt *CI = dyn_cast(Idxs[i])) { - if (isa(Ty) || isa(Ty)) + if (isa(Ty)) if (CI->getSExtValue() > 0 && !isIndexInRangeOfSequentialType(cast(Ty), CI)) { if (isa(Prev)) { Index: llvm/trunk/test/Transforms/InstCombine/getelementptr-folding.ll =================================================================== --- llvm/trunk/test/Transforms/InstCombine/getelementptr-folding.ll +++ llvm/trunk/test/Transforms/InstCombine/getelementptr-folding.ll @@ -0,0 +1,13 @@ +; RUN: opt -instcombine -S < %s | FileCheck %s + +%struct.matrix_float3x3 = type { [3 x <3 x float>] } + +; We used to fold this by rewriting the indices to 0, 0, 2, 0. This is +; invalid because there is a 4-byte padding after each <3 x float> field. + +target datalayout = "e-m:o-i64:64-f80:128-n8:16:32:64-S128" +target triple = "x86_64-apple-macosx10.11.0" + +@matrix_identity_float3x3 = external global %struct.matrix_float3x3, align 16 +@bbb = global float* getelementptr inbounds (%struct.matrix_float3x3, %struct.matrix_float3x3* @matrix_identity_float3x3, i64 0, i32 0, i64 1, i64 3) +; CHECK: @bbb = global float* getelementptr inbounds (%struct.matrix_float3x3, %struct.matrix_float3x3* @matrix_identity_float3x3, i64 0, i32 0, i64 1, i64 3)