218 lines
11 KiB
C
218 lines
11 KiB
C
// NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py
|
|
/// Check the code generation for the alignment builtins
|
|
/// To make the test case easier to read, run SROA after generating IR to remove the alloca instructions.
|
|
// RUN: %clang_cc1 -triple=x86_64-unknown-unknown -DTEST_VOID_PTR \
|
|
// RUN: -o - -emit-llvm %s -disable-O0-optnone | opt -S -sroa | \
|
|
// RUN: FileCheck %s -check-prefixes CHECK,CHECK-VOID_PTR \
|
|
// RUN: -enable-var-scope '-D$PTRTYPE=i8'
|
|
// RUN: %clang_cc1 -triple=x86_64-unknown-unknown -DTEST_FLOAT_PTR \
|
|
// RUN: -o - -emit-llvm %s -disable-O0-optnone | opt -S -sroa | \
|
|
// RUN: FileCheck %s -check-prefixes CHECK,CHECK-FLOAT_PTR \
|
|
// RUN: -enable-var-scope '-D$PTRTYPE=f32'
|
|
// RUN: %clang_cc1 -triple=x86_64-unknown-unknown -DTEST_LONG \
|
|
// RUN: -o - -emit-llvm %s -disable-O0-optnone | opt -S -sroa | \
|
|
// RUN: FileCheck %s -check-prefixes CHECK,CHECK-LONG -enable-var-scope
|
|
/// Check that we can handle the case where the alignment parameter is wider
|
|
/// than the source type (generate a trunc on alignment instead of zext)
|
|
// RUN: %clang_cc1 -triple=x86_64-unknown-unknown -DTEST_USHORT \
|
|
// RUN: -o - -emit-llvm %s -disable-O0-optnone | opt -S -sroa | \
|
|
// RUN: FileCheck %s -check-prefixes CHECK,CHECK-USHORT -enable-var-scope
|
|
|
|
|
|
#ifdef TEST_VOID_PTR
|
|
#define TYPE void *
|
|
#elif defined(TEST_FLOAT_PTR)
|
|
#define TYPE float *
|
|
#elif defined(TEST_LONG)
|
|
#define TYPE long
|
|
#elif defined(TEST_USHORT)
|
|
#define TYPE unsigned short
|
|
#else
|
|
#error MISSING TYPE
|
|
#endif
|
|
|
|
/// Check that constant initializers work and are correct
|
|
_Bool aligned_true = __builtin_is_aligned(1024, 512);
|
|
// CHECK: @aligned_true ={{.*}} global i8 1, align 1
|
|
_Bool aligned_false = __builtin_is_aligned(123, 512);
|
|
// CHECK: @aligned_false ={{.*}} global i8 0, align 1
|
|
|
|
int down_1 = __builtin_align_down(1023, 32);
|
|
// CHECK: @down_1 ={{.*}} global i32 992, align 4
|
|
int down_2 = __builtin_align_down(256, 32);
|
|
// CHECK: @down_2 ={{.*}} global i32 256, align 4
|
|
|
|
int up_1 = __builtin_align_up(1023, 32);
|
|
// CHECK: @up_1 ={{.*}} global i32 1024, align 4
|
|
int up_2 = __builtin_align_up(256, 32);
|
|
// CHECK: @up_2 ={{.*}} global i32 256, align 4
|
|
|
|
/// Capture the IR type here to use in the remaining FileCheck captures:
|
|
// CHECK-VOID_PTR-LABEL: define {{[^@]+}}@get_type() #0
|
|
// CHECK-VOID_PTR-NEXT: entry:
|
|
// CHECK-VOID_PTR-NEXT: ret i8* null
|
|
//
|
|
// CHECK-FLOAT_PTR-LABEL: define {{[^@]+}}@get_type() #0
|
|
// CHECK-FLOAT_PTR-NEXT: entry:
|
|
// CHECK-FLOAT_PTR-NEXT: ret float* null
|
|
//
|
|
// CHECK-LONG-LABEL: define {{[^@]+}}@get_type() #0
|
|
// CHECK-LONG-NEXT: entry:
|
|
// CHECK-LONG-NEXT: ret i64 0
|
|
//
|
|
// CHECK-USHORT-LABEL: define {{[^@]+}}@get_type() #0
|
|
// CHECK-USHORT-NEXT: entry:
|
|
// CHECK-USHORT-NEXT: ret i16 0
|
|
//
|
|
TYPE get_type(void) {
|
|
return (TYPE)0;
|
|
}
|
|
|
|
// CHECK-VOID_PTR-LABEL: define {{[^@]+}}@is_aligned
|
|
// CHECK-VOID_PTR-SAME: (i8* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-VOID_PTR-NEXT: entry:
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-VOID_PTR-NEXT: [[SRC_ADDR:%.*]] = ptrtoint i8* [[PTR]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[SET_BITS:%.*]] = and i64 [[SRC_ADDR]], [[MASK]]
|
|
// CHECK-VOID_PTR-NEXT: [[IS_ALIGNED:%.*]] = icmp eq i64 [[SET_BITS]], 0
|
|
// CHECK-VOID_PTR-NEXT: ret i1 [[IS_ALIGNED]]
|
|
//
|
|
// CHECK-FLOAT_PTR-LABEL: define {{[^@]+}}@is_aligned
|
|
// CHECK-FLOAT_PTR-SAME: (float* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-FLOAT_PTR-NEXT: entry:
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-FLOAT_PTR-NEXT: [[SRC_ADDR:%.*]] = ptrtoint float* [[PTR]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[SET_BITS:%.*]] = and i64 [[SRC_ADDR]], [[MASK]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[IS_ALIGNED:%.*]] = icmp eq i64 [[SET_BITS]], 0
|
|
// CHECK-FLOAT_PTR-NEXT: ret i1 [[IS_ALIGNED]]
|
|
//
|
|
// CHECK-LONG-LABEL: define {{[^@]+}}@is_aligned
|
|
// CHECK-LONG-SAME: (i64 [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-LONG-NEXT: entry:
|
|
// CHECK-LONG-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-LONG-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-LONG-NEXT: [[SET_BITS:%.*]] = and i64 [[PTR]], [[MASK]]
|
|
// CHECK-LONG-NEXT: [[IS_ALIGNED:%.*]] = icmp eq i64 [[SET_BITS]], 0
|
|
// CHECK-LONG-NEXT: ret i1 [[IS_ALIGNED]]
|
|
//
|
|
// CHECK-USHORT-LABEL: define {{[^@]+}}@is_aligned
|
|
// CHECK-USHORT-SAME: (i16 zeroext [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-USHORT-NEXT: entry:
|
|
// CHECK-USHORT-NEXT: [[ALIGNMENT:%.*]] = trunc i32 [[ALIGN]] to i16
|
|
// CHECK-USHORT-NEXT: [[MASK:%.*]] = sub i16 [[ALIGNMENT]], 1
|
|
// CHECK-USHORT-NEXT: [[SET_BITS:%.*]] = and i16 [[PTR]], [[MASK]]
|
|
// CHECK-USHORT-NEXT: [[IS_ALIGNED:%.*]] = icmp eq i16 [[SET_BITS]], 0
|
|
// CHECK-USHORT-NEXT: ret i1 [[IS_ALIGNED]]
|
|
//
|
|
_Bool is_aligned(TYPE ptr, unsigned align) {
|
|
return __builtin_is_aligned(ptr, align);
|
|
}
|
|
|
|
// NOTYET-POINTER-NEXT: [[ALIGNED_RESULT:%.*]] = call [[$TYPE]] @llvm.ptrmask.p0[[$PTRTYPE]].p0i8.i64(i8* [[OVER_BOUNDARY]], [[ALIGN_TYPE]] [[INVERTED_MASK]])
|
|
// CHECK-VOID_PTR-LABEL: define {{[^@]+}}@align_up
|
|
// CHECK-VOID_PTR-SAME: (i8* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-VOID_PTR-NEXT: entry:
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-VOID_PTR-NEXT: [[INTPTR:%.*]] = ptrtoint i8* [[PTR]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[OVER_BOUNDARY:%.*]] = add i64 [[INTPTR]], [[MASK]]
|
|
// CHECK-VOID_PTR-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNED_INTPTR:%.*]] = and i64 [[OVER_BOUNDARY]], [[INVERTED_MASK]]
|
|
// CHECK-VOID_PTR-NEXT: [[DIFF:%.*]] = sub i64 [[ALIGNED_INTPTR]], [[INTPTR]]
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNED_RESULT:%.*]] = getelementptr inbounds i8, i8* [[PTR]], i64 [[DIFF]]
|
|
// CHECK-VOID_PTR-NEXT: call void @llvm.assume(i1 true) [ "align"(i8* [[ALIGNED_RESULT]], i64 [[ALIGNMENT]]) ]
|
|
// CHECK-VOID_PTR-NEXT: ret i8* [[ALIGNED_RESULT]]
|
|
//
|
|
// CHECK-FLOAT_PTR-LABEL: define {{[^@]+}}@align_up
|
|
// CHECK-FLOAT_PTR-SAME: (float* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-FLOAT_PTR-NEXT: entry:
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-FLOAT_PTR-NEXT: [[INTPTR:%.*]] = ptrtoint float* [[PTR]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[OVER_BOUNDARY:%.*]] = add i64 [[INTPTR]], [[MASK]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNED_INTPTR:%.*]] = and i64 [[OVER_BOUNDARY]], [[INVERTED_MASK]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[DIFF:%.*]] = sub i64 [[ALIGNED_INTPTR]], [[INTPTR]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[TMP0:%.*]] = bitcast float* [[PTR]] to i8*
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNED_RESULT:%.*]] = getelementptr inbounds i8, i8* [[TMP0]], i64 [[DIFF]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[TMP1:%.*]] = bitcast i8* [[ALIGNED_RESULT]] to float*
|
|
// CHECK-FLOAT_PTR-NEXT: call void @llvm.assume(i1 true) [ "align"(float* [[TMP1]], i64 [[ALIGNMENT]]) ]
|
|
// CHECK-FLOAT_PTR-NEXT: ret float* [[TMP1]]
|
|
//
|
|
// CHECK-LONG-LABEL: define {{[^@]+}}@align_up
|
|
// CHECK-LONG-SAME: (i64 [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-LONG-NEXT: entry:
|
|
// CHECK-LONG-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-LONG-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-LONG-NEXT: [[OVER_BOUNDARY:%.*]] = add i64 [[PTR]], [[MASK]]
|
|
// CHECK-LONG-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-LONG-NEXT: [[ALIGNED_RESULT:%.*]] = and i64 [[OVER_BOUNDARY]], [[INVERTED_MASK]]
|
|
// CHECK-LONG-NEXT: ret i64 [[ALIGNED_RESULT]]
|
|
//
|
|
// CHECK-USHORT-LABEL: define {{[^@]+}}@align_up
|
|
// CHECK-USHORT-SAME: (i16 zeroext [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-USHORT-NEXT: entry:
|
|
// CHECK-USHORT-NEXT: [[ALIGNMENT:%.*]] = trunc i32 [[ALIGN]] to i16
|
|
// CHECK-USHORT-NEXT: [[MASK:%.*]] = sub i16 [[ALIGNMENT]], 1
|
|
// CHECK-USHORT-NEXT: [[OVER_BOUNDARY:%.*]] = add i16 [[PTR]], [[MASK]]
|
|
// CHECK-USHORT-NEXT: [[INVERTED_MASK:%.*]] = xor i16 [[MASK]], -1
|
|
// CHECK-USHORT-NEXT: [[ALIGNED_RESULT:%.*]] = and i16 [[OVER_BOUNDARY]], [[INVERTED_MASK]]
|
|
// CHECK-USHORT-NEXT: ret i16 [[ALIGNED_RESULT]]
|
|
//
|
|
TYPE align_up(TYPE ptr, unsigned align) {
|
|
return __builtin_align_up(ptr, align);
|
|
}
|
|
|
|
// NOTYET-POINTER-NEXT: [[ALIGNED_RESULT:%.*]] = call [[$TYPE]] @llvm.ptrmask.p0[[$PTRTYPE]].p0[[$PTRTYPE]].i64([[$TYPE]] [[PTR]], [[ALIGN_TYPE]] [[INVERTED_MASK]])
|
|
// CHECK-VOID_PTR-LABEL: define {{[^@]+}}@align_down
|
|
// CHECK-VOID_PTR-SAME: (i8* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-VOID_PTR-NEXT: entry:
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-VOID_PTR-NEXT: [[INTPTR:%.*]] = ptrtoint i8* [[PTR]] to i64
|
|
// CHECK-VOID_PTR-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNED_INTPTR:%.*]] = and i64 [[INTPTR]], [[INVERTED_MASK]]
|
|
// CHECK-VOID_PTR-NEXT: [[DIFF:%.*]] = sub i64 [[ALIGNED_INTPTR]], [[INTPTR]]
|
|
// CHECK-VOID_PTR-NEXT: [[ALIGNED_RESULT:%.*]] = getelementptr inbounds i8, i8* [[PTR]], i64 [[DIFF]]
|
|
// CHECK-VOID_PTR-NEXT: call void @llvm.assume(i1 true) [ "align"(i8* [[ALIGNED_RESULT]], i64 [[ALIGNMENT]]) ]
|
|
// CHECK-VOID_PTR-NEXT: ret i8* [[ALIGNED_RESULT]]
|
|
//
|
|
// CHECK-FLOAT_PTR-LABEL: define {{[^@]+}}@align_down
|
|
// CHECK-FLOAT_PTR-SAME: (float* [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-FLOAT_PTR-NEXT: entry:
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-FLOAT_PTR-NEXT: [[INTPTR:%.*]] = ptrtoint float* [[PTR]] to i64
|
|
// CHECK-FLOAT_PTR-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNED_INTPTR:%.*]] = and i64 [[INTPTR]], [[INVERTED_MASK]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[DIFF:%.*]] = sub i64 [[ALIGNED_INTPTR]], [[INTPTR]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[TMP0:%.*]] = bitcast float* [[PTR]] to i8*
|
|
// CHECK-FLOAT_PTR-NEXT: [[ALIGNED_RESULT:%.*]] = getelementptr inbounds i8, i8* [[TMP0]], i64 [[DIFF]]
|
|
// CHECK-FLOAT_PTR-NEXT: [[TMP1:%.*]] = bitcast i8* [[ALIGNED_RESULT]] to float*
|
|
// CHECK-FLOAT_PTR-NEXT: call void @llvm.assume(i1 true) [ "align"(float* [[TMP1]], i64 [[ALIGNMENT]]) ]
|
|
// CHECK-FLOAT_PTR-NEXT: ret float* [[TMP1]]
|
|
//
|
|
// CHECK-LONG-LABEL: define {{[^@]+}}@align_down
|
|
// CHECK-LONG-SAME: (i64 [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-LONG-NEXT: entry:
|
|
// CHECK-LONG-NEXT: [[ALIGNMENT:%.*]] = zext i32 [[ALIGN]] to i64
|
|
// CHECK-LONG-NEXT: [[MASK:%.*]] = sub i64 [[ALIGNMENT]], 1
|
|
// CHECK-LONG-NEXT: [[INVERTED_MASK:%.*]] = xor i64 [[MASK]], -1
|
|
// CHECK-LONG-NEXT: [[ALIGNED_RESULT:%.*]] = and i64 [[PTR]], [[INVERTED_MASK]]
|
|
// CHECK-LONG-NEXT: ret i64 [[ALIGNED_RESULT]]
|
|
//
|
|
// CHECK-USHORT-LABEL: define {{[^@]+}}@align_down
|
|
// CHECK-USHORT-SAME: (i16 zeroext [[PTR:%.*]], i32 [[ALIGN:%.*]]) #0
|
|
// CHECK-USHORT-NEXT: entry:
|
|
// CHECK-USHORT-NEXT: [[ALIGNMENT:%.*]] = trunc i32 [[ALIGN]] to i16
|
|
// CHECK-USHORT-NEXT: [[MASK:%.*]] = sub i16 [[ALIGNMENT]], 1
|
|
// CHECK-USHORT-NEXT: [[INVERTED_MASK:%.*]] = xor i16 [[MASK]], -1
|
|
// CHECK-USHORT-NEXT: [[ALIGNED_RESULT:%.*]] = and i16 [[PTR]], [[INVERTED_MASK]]
|
|
// CHECK-USHORT-NEXT: ret i16 [[ALIGNED_RESULT]]
|
|
//
|
|
TYPE align_down(TYPE ptr, unsigned align) {
|
|
return __builtin_align_down(ptr, align);
|
|
}
|