Changeset View
Changeset View
Standalone View
Standalone View
clang/test/CodeGen/alloc-align-attr.c
Context not available. | |||||
__INT32_TYPE__*m1(__INT32_TYPE__ i) __attribute__((alloc_align(1))); | __INT32_TYPE__*m1(__INT32_TYPE__ i) __attribute__((alloc_align(1))); | ||||
// Condition where parameter to m1 is not size_t. | // Condition where parameter to m1 is not size_t. | ||||
// CHECK-LABEL: define {{[^@]+}}@test1 | // CHECK-LABEL: @test1( | ||||
// CHECK-SAME: (i32 [[A:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4 | ||||
// CHECK-NEXT: store i32 [[A]], i32* [[A_ADDR]], align 4 | // CHECK-NEXT: store i32 [[A:%.*]], i32* [[A_ADDR]], align 4 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = load i32, i32* [[A_ADDR]], align 4 | // CHECK-NEXT: [[TMP0:%.*]] = load i32, i32* [[A_ADDR]], align 4 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m1(i32 [[TMP0]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m1(i32 noundef [[TMP0]]) | ||||
// CHECK-NEXT: [[CASTED_ALIGN:%.*]] = zext i32 [[TMP0]] to i64 | // CHECK-NEXT: [[CASTED_ALIGN:%.*]] = zext i32 [[TMP0]] to i64 | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | ||||
// CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
Context not available. | |||||
return *m1(a); | return *m1(a); | ||||
} | } | ||||
// Condition where test2 param needs casting. | // Condition where test2 param needs casting. | ||||
// CHECK-LABEL: define {{[^@]+}}@test2 | // CHECK-LABEL: @test2( | ||||
// CHECK-SAME: (i64 [[A:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i64, align 8 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i64, align 8 | ||||
// CHECK-NEXT: store i64 [[A]], i64* [[A_ADDR]], align 8 | // CHECK-NEXT: store i64 [[A:%.*]], i64* [[A_ADDR]], align 8 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = load i64, i64* [[A_ADDR]], align 8 | // CHECK-NEXT: [[TMP0:%.*]] = load i64, i64* [[A_ADDR]], align 8 | ||||
// CHECK-NEXT: [[CONV:%.*]] = trunc i64 [[TMP0]] to i32 | // CHECK-NEXT: [[CONV:%.*]] = trunc i64 [[TMP0]] to i32 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m1(i32 [[CONV]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m1(i32 noundef [[CONV]]) | ||||
// CHECK-NEXT: [[CASTED_ALIGN:%.*]] = zext i32 [[CONV]] to i64 | // CHECK-NEXT: [[CASTED_ALIGN:%.*]] = zext i32 [[CONV]] to i64 | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | ||||
// CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
Context not available. | |||||
__INT32_TYPE__ *m2(__SIZE_TYPE__ i) __attribute__((alloc_align(1))); | __INT32_TYPE__ *m2(__SIZE_TYPE__ i) __attribute__((alloc_align(1))); | ||||
// test3 param needs casting, but 'm2' is correct. | // test3 param needs casting, but 'm2' is correct. | ||||
// CHECK-LABEL: define {{[^@]+}}@test3 | // CHECK-LABEL: @test3( | ||||
// CHECK-SAME: (i32 [[A:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i32, align 4 | ||||
// CHECK-NEXT: store i32 [[A]], i32* [[A_ADDR]], align 4 | // CHECK-NEXT: store i32 [[A:%.*]], i32* [[A_ADDR]], align 4 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = load i32, i32* [[A_ADDR]], align 4 | // CHECK-NEXT: [[TMP0:%.*]] = load i32, i32* [[A_ADDR]], align 4 | ||||
// CHECK-NEXT: [[CONV:%.*]] = sext i32 [[TMP0]] to i64 | // CHECK-NEXT: [[CONV:%.*]] = sext i32 [[TMP0]] to i64 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m2(i64 [[CONV]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m2(i64 noundef [[CONV]]) | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CONV]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CONV]]) ] | ||||
// CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
// CHECK-NEXT: ret i32 [[TMP1]] | // CHECK-NEXT: ret i32 [[TMP1]] | ||||
Context not available. | |||||
} | } | ||||
// Every type matches, canonical example. | // Every type matches, canonical example. | ||||
// CHECK-LABEL: define {{[^@]+}}@test4 | // CHECK-LABEL: @test4( | ||||
// CHECK-SAME: (i64 [[A:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i64, align 8 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i64, align 8 | ||||
// CHECK-NEXT: store i64 [[A]], i64* [[A_ADDR]], align 8 | // CHECK-NEXT: store i64 [[A:%.*]], i64* [[A_ADDR]], align 8 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = load i64, i64* [[A_ADDR]], align 8 | // CHECK-NEXT: [[TMP0:%.*]] = load i64, i64* [[A_ADDR]], align 8 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m2(i64 [[TMP0]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m2(i64 noundef [[TMP0]]) | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[TMP0]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[TMP0]]) ] | ||||
// CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP1:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
// CHECK-NEXT: ret i32 [[TMP1]] | // CHECK-NEXT: ret i32 [[TMP1]] | ||||
Context not available. | |||||
// Struct parameter doesn't take up an IR parameter, 'i' takes up 2. | // Struct parameter doesn't take up an IR parameter, 'i' takes up 2. | ||||
// Truncation to i64 is permissible, since alignments of greater than 2^64 are insane. | // Truncation to i64 is permissible, since alignments of greater than 2^64 are insane. | ||||
__INT32_TYPE__ *m3(struct Empty s, __int128_t i) __attribute__((alloc_align(2))); | __INT32_TYPE__ *m3(struct Empty s, __int128_t i) __attribute__((alloc_align(2))); | ||||
// CHECK-LABEL: define {{[^@]+}}@test5 | // CHECK-LABEL: @test5( | ||||
// CHECK-SAME: (i64 [[A_COERCE0:%.*]], i64 [[A_COERCE1:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[A:%.*]] = alloca i128, align 16 | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i128, align 16 | ||||
Context not available. | |||||
// CHECK-NEXT: [[COERCE:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[COERCE:%.*]] = alloca i128, align 16 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = bitcast i128* [[A]] to { i64, i64 }* | // CHECK-NEXT: [[TMP0:%.*]] = bitcast i128* [[A]] to { i64, i64 }* | ||||
// CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 0 | // CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 0 | ||||
// CHECK-NEXT: store i64 [[A_COERCE0]], i64* [[TMP1]], align 16 | // CHECK-NEXT: store i64 [[A_COERCE0:%.*]], i64* [[TMP1]], align 16 | ||||
// CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 1 | // CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 1 | ||||
// CHECK-NEXT: store i64 [[A_COERCE1]], i64* [[TMP2]], align 8 | // CHECK-NEXT: store i64 [[A_COERCE1:%.*]], i64* [[TMP2]], align 8 | ||||
// CHECK-NEXT: [[A1:%.*]] = load i128, i128* [[A]], align 16 | // CHECK-NEXT: [[A1:%.*]] = load i128, i128* [[A]], align 16 | ||||
// CHECK-NEXT: store i128 [[A1]], i128* [[A_ADDR]], align 16 | // CHECK-NEXT: store i128 [[A1]], i128* [[A_ADDR]], align 16 | ||||
// CHECK-NEXT: [[TMP3:%.*]] = load i128, i128* [[A_ADDR]], align 16 | // CHECK-NEXT: [[TMP3:%.*]] = load i128, i128* [[A_ADDR]], align 16 | ||||
Context not available. | |||||
// CHECK-NEXT: [[TMP6:%.*]] = load i64, i64* [[TMP5]], align 16 | // CHECK-NEXT: [[TMP6:%.*]] = load i64, i64* [[TMP5]], align 16 | ||||
// CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP4]], i32 0, i32 1 | // CHECK-NEXT: [[TMP7:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP4]], i32 0, i32 1 | ||||
// CHECK-NEXT: [[TMP8:%.*]] = load i64, i64* [[TMP7]], align 8 | // CHECK-NEXT: [[TMP8:%.*]] = load i64, i64* [[TMP7]], align 8 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m3(i64 [[TMP6]], i64 [[TMP8]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m3(i64 noundef [[TMP6]], i64 noundef [[TMP8]]) | ||||
// CHECK-NEXT: [[CASTED_ALIGN:%.*]] = trunc i128 [[TMP3]] to i64 | // CHECK-NEXT: [[CASTED_ALIGN:%.*]] = trunc i128 [[TMP3]] to i64 | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | ||||
// CHECK-NEXT: [[TMP9:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP9:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
Context not available. | |||||
} | } | ||||
// Struct parameter takes up 2 parameters, 'i' takes up 2. | // Struct parameter takes up 2 parameters, 'i' takes up 2. | ||||
__INT32_TYPE__ *m4(struct MultiArgs s, __int128_t i) __attribute__((alloc_align(2))); | __INT32_TYPE__ *m4(struct MultiArgs s, __int128_t i) __attribute__((alloc_align(2))); | ||||
// CHECK-LABEL: define {{[^@]+}}@test6 | // CHECK-LABEL: @test6( | ||||
// CHECK-SAME: (i64 [[A_COERCE0:%.*]], i64 [[A_COERCE1:%.*]]) #0 | |||||
// CHECK-NEXT: entry: | // CHECK-NEXT: entry: | ||||
// CHECK-NEXT: [[A:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[A:%.*]] = alloca i128, align 16 | ||||
// CHECK-NEXT: [[A_ADDR:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[A_ADDR:%.*]] = alloca i128, align 16 | ||||
Context not available. | |||||
// CHECK-NEXT: [[COERCE:%.*]] = alloca i128, align 16 | // CHECK-NEXT: [[COERCE:%.*]] = alloca i128, align 16 | ||||
// CHECK-NEXT: [[TMP0:%.*]] = bitcast i128* [[A]] to { i64, i64 }* | // CHECK-NEXT: [[TMP0:%.*]] = bitcast i128* [[A]] to { i64, i64 }* | ||||
// CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 0 | // CHECK-NEXT: [[TMP1:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 0 | ||||
// CHECK-NEXT: store i64 [[A_COERCE0]], i64* [[TMP1]], align 16 | // CHECK-NEXT: store i64 [[A_COERCE0:%.*]], i64* [[TMP1]], align 16 | ||||
// CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 1 | // CHECK-NEXT: [[TMP2:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP0]], i32 0, i32 1 | ||||
// CHECK-NEXT: store i64 [[A_COERCE1]], i64* [[TMP2]], align 8 | // CHECK-NEXT: store i64 [[A_COERCE1:%.*]], i64* [[TMP2]], align 8 | ||||
// CHECK-NEXT: [[A1:%.*]] = load i128, i128* [[A]], align 16 | // CHECK-NEXT: [[A1:%.*]] = load i128, i128* [[A]], align 16 | ||||
// CHECK-NEXT: store i128 [[A1]], i128* [[A_ADDR]], align 16 | // CHECK-NEXT: store i128 [[A1]], i128* [[A_ADDR]], align 16 | ||||
// CHECK-NEXT: [[TMP3:%.*]] = load i128, i128* [[A_ADDR]], align 16 | // CHECK-NEXT: [[TMP3:%.*]] = load i128, i128* [[A_ADDR]], align 16 | ||||
Context not available. | |||||
// CHECK-NEXT: [[TMP11:%.*]] = load i64, i64* [[TMP10]], align 16 | // CHECK-NEXT: [[TMP11:%.*]] = load i64, i64* [[TMP10]], align 16 | ||||
// CHECK-NEXT: [[TMP12:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP9]], i32 0, i32 1 | // CHECK-NEXT: [[TMP12:%.*]] = getelementptr inbounds { i64, i64 }, { i64, i64 }* [[TMP9]], i32 0, i32 1 | ||||
// CHECK-NEXT: [[TMP13:%.*]] = load i64, i64* [[TMP12]], align 8 | // CHECK-NEXT: [[TMP13:%.*]] = load i64, i64* [[TMP12]], align 8 | ||||
// CHECK-NEXT: [[CALL:%.*]] = call i32* @m4(i64 [[TMP6]], i64 [[TMP8]], i64 [[TMP11]], i64 [[TMP13]]) | // CHECK-NEXT: [[CALL:%.*]] = call noundef i32* @m4(i64 [[TMP6]], i64 [[TMP8]], i64 noundef [[TMP11]], i64 noundef [[TMP13]]) | ||||
// CHECK-NEXT: [[CASTED_ALIGN:%.*]] = trunc i128 [[TMP3]] to i64 | // CHECK-NEXT: [[CASTED_ALIGN:%.*]] = trunc i128 [[TMP3]] to i64 | ||||
// CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | // CHECK-NEXT: call void @llvm.assume(i1 true) [ "align"(i32* [[CALL]], i64 [[CASTED_ALIGN]]) ] | ||||
// CHECK-NEXT: [[TMP14:%.*]] = load i32, i32* [[CALL]], align 4 | // CHECK-NEXT: [[TMP14:%.*]] = load i32, i32* [[CALL]], align 4 | ||||
Context not available. |