// RUN: %clang_cc1 %s -emit-llvm -o - -ffreestanding -ffake-address-space-map -triple=i686-apple-darwin9 | FileCheck %s // REQUIRES: x86-registered-target // Also test serialization of atomic operations here, to avoid duplicating the // test. // RUN: %clang_cc1 %s -emit-pch -o %t -ffreestanding -ffake-address-space-map -triple=i686-apple-darwin9 // RUN: %clang_cc1 %s -include-pch %t -ffreestanding -ffake-address-space-map -triple=i686-apple-darwin9 -emit-llvm -o - | FileCheck %s #ifndef ALREADY_INCLUDED #define ALREADY_INCLUDED #include // Basic IRGen tests for __c11_atomic_* and GNU __atomic_* int fi1(_Atomic(int) *i) { // CHECK-LABEL: @fi1 // CHECK: load atomic i32, i32* {{.*}} seq_cst return __c11_atomic_load(i, memory_order_seq_cst); } int fi1a(int *i) { // CHECK-LABEL: @fi1a // CHECK: load atomic i32, i32* {{.*}} seq_cst int v; __atomic_load(i, &v, memory_order_seq_cst); return v; } int fi1b(int *i) { // CHECK-LABEL: @fi1b // CHECK: load atomic i32, i32* {{.*}} seq_cst return __atomic_load_n(i, memory_order_seq_cst); } int fi1c(atomic_int *i) { // CHECK-LABEL: @fi1c // CHECK: load atomic i32, i32* {{.*}} seq_cst return atomic_load(i); } void fi2(_Atomic(int) *i) { // CHECK-LABEL: @fi2 // CHECK: store atomic i32 {{.*}} seq_cst __c11_atomic_store(i, 1, memory_order_seq_cst); } void fi2a(int *i) { // CHECK-LABEL: @fi2a // CHECK: store atomic i32 {{.*}} seq_cst int v = 1; __atomic_store(i, &v, memory_order_seq_cst); } void fi2b(int *i) { // CHECK-LABEL: @fi2b // CHECK: store atomic i32 {{.*}} seq_cst __atomic_store_n(i, 1, memory_order_seq_cst); } void fi2c(atomic_int *i) { // CHECK-LABEL: @fi2c // CHECK: store atomic i32 {{.*}} seq_cst atomic_store(i, 1); } int fi3(_Atomic(int) *i) { // CHECK-LABEL: @fi3 // CHECK: atomicrmw and // CHECK-NOT: and return __c11_atomic_fetch_and(i, 1, memory_order_seq_cst); } int fi3a(int *i) { // CHECK-LABEL: @fi3a // CHECK: atomicrmw xor // CHECK-NOT: xor return __atomic_fetch_xor(i, 1, memory_order_seq_cst); } int fi3b(int *i) { // CHECK-LABEL: @fi3b // CHECK: atomicrmw add // CHECK: add return __atomic_add_fetch(i, 1, memory_order_seq_cst); } int fi3c(int *i) { // CHECK-LABEL: @fi3c // CHECK: atomicrmw nand // CHECK-NOT: and return __atomic_fetch_nand(i, 1, memory_order_seq_cst); } int fi3d(int *i) { // CHECK-LABEL: @fi3d // CHECK: atomicrmw nand // CHECK: and // CHECK: xor return __atomic_nand_fetch(i, 1, memory_order_seq_cst); } int fi3e(atomic_int *i) { // CHECK-LABEL: @fi3e // CHECK: atomicrmw or // CHECK-NOT: {{ or }} return atomic_fetch_or(i, 1); } int fi3f(int *i) { // CHECK-LABEL: @fi3f // CHECK-NOT: store volatile // CHECK: atomicrmw or // CHECK-NOT: {{ or }} return __atomic_fetch_or(i, (short)1, memory_order_seq_cst); } _Bool fi4(_Atomic(int) *i) { // CHECK-LABEL: @fi4( // CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0 // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1 // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] // CHECK: store i32 [[OLD]] int cmp = 0; return __c11_atomic_compare_exchange_strong(i, &cmp, 1, memory_order_acquire, memory_order_acquire); } _Bool fi4a(int *i) { // CHECK-LABEL: @fi4a // CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0 // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1 // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] // CHECK: store i32 [[OLD]] int cmp = 0; int desired = 1; return __atomic_compare_exchange(i, &cmp, &desired, 0, memory_order_acquire, memory_order_acquire); } _Bool fi4b(int *i) { // CHECK-LABEL: @fi4b( // CHECK: [[PAIR:%[.0-9A-Z_a-z]+]] = cmpxchg weak i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 0 // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = extractvalue { i32, i1 } [[PAIR]], 1 // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] // CHECK: store i32 [[OLD]] int cmp = 0; return __atomic_compare_exchange_n(i, &cmp, 1, 1, memory_order_acquire, memory_order_acquire); } _Bool fi4c(atomic_int *i) { // CHECK-LABEL: @fi4c // CHECK: cmpxchg i32* int cmp = 0; return atomic_compare_exchange_strong(i, &cmp, 1); } #define _AS1 __attribute__((address_space(1))) _Bool fi4d(_Atomic(int) *i, int _AS1 *ptr2) { // CHECK-LABEL: @fi4d( // CHECK: [[EXPECTED:%[.0-9A-Z_a-z]+]] = load i32, i32 addrspace(1)* %{{[0-9]+}} // CHECK: cmpxchg i32* %{{[0-9]+}}, i32 [[EXPECTED]], i32 %{{[0-9]+}} acquire acquire return __c11_atomic_compare_exchange_strong(i, ptr2, 1, memory_order_acquire, memory_order_acquire); } float ff1(_Atomic(float) *d) { // CHECK-LABEL: @ff1 // CHECK: load atomic i32, i32* {{.*}} monotonic return __c11_atomic_load(d, memory_order_relaxed); } void ff2(_Atomic(float) *d) { // CHECK-LABEL: @ff2 // CHECK: store atomic i32 {{.*}} release __c11_atomic_store(d, 1, memory_order_release); } float ff3(_Atomic(float) *d) { return __c11_atomic_exchange(d, 2, memory_order_seq_cst); } struct S { double x; }; void implicit_store(_Atomic(struct S) *a, struct S s) { // CHECK-LABEL: @implicit_store( // CHECK: store atomic i64 %{{.*}}, i64* %{{.*}} seq_cst, align 8 *a = s; } struct S implicit_load(_Atomic(struct S) *a) { // CHECK-LABEL: @implicit_load( // CHECK: load atomic i64, i64* %{{.*}} seq_cst, align 8 return *a; } struct S fd1(struct S *a) { // CHECK-LABEL: @fd1 // CHECK: [[RETVAL:%.*]] = alloca %struct.S, align 4 // CHECK: [[A:%.*]] = bitcast %struct.S* {{.*}} to i64* // CHECK: [[CAST:%.*]] = bitcast %struct.S* [[RETVAL]] to i64* // CHECK: [[SRC:%.*]] = bitcast i64* [[A]] to i8* // CHECK: [[DEST:%.*]] = bitcast i64* [[CAST]] to i8* // CHECK: call void @__atomic_load(i32 8, i8* [[SRC]], i8* [[DEST]], i32 5) // CHECK: ret struct S ret; __atomic_load(a, &ret, memory_order_seq_cst); return ret; } void fd2(struct S *a, struct S *b) { // CHECK-LABEL: @fd2 // CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S*, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S*, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: [[COERCED_A_TMP:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i64* // CHECK-NEXT: [[COERCED_B:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i64* // CHECK-NEXT: [[COERCED_A:%.*]] = bitcast i64* [[COERCED_A_TMP]] to i8* // CHECK-NEXT: [[CAST_B:%.*]] = bitcast i64* [[COERCED_B]] to i8* // CHECK-NEXT: call void @__atomic_store(i32 8, i8* [[COERCED_A]], i8* [[CAST_B]], // CHECK-NEXT: ret void __atomic_store(a, b, memory_order_seq_cst); } void fd3(struct S *a, struct S *b, struct S *c) { // CHECK-LABEL: @fd3 // CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: [[C_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: store %struct.S* %c, %struct.S** [[C_ADDR]], align 4 // CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S*, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S*, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: [[LOAD_C_PTR:%.*]] = load %struct.S*, %struct.S** [[C_ADDR]], align 4 // CHECK-NEXT: [[COERCED_A_TMP:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i64* // CHECK-NEXT: [[COERCED_B:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i64* // CHECK-NEXT: [[COERCED_C:%.*]] = bitcast %struct.S* [[LOAD_C_PTR]] to i64* // CHECK-NEXT: [[COERCED_A:%.*]] = bitcast i64* [[COERCED_A_TMP]] to i8* // CHECK-NEXT: [[CAST_B:%.*]] = bitcast i64* [[COERCED_B]] to i8* // CHECK-NEXT: [[CAST_C:%.*]] = bitcast i64* [[COERCED_C]] to i8* // CHECK-NEXT: call void @__atomic_exchange(i32 8, i8* [[COERCED_A]], i8* [[CAST_B]], i8* [[CAST_C]], __atomic_exchange(a, b, c, memory_order_seq_cst); } _Bool fd4(struct S *a, struct S *b, struct S *c) { // CHECK-LABEL: @fd4 // CHECK: [[A_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: [[B_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK-NEXT: [[C_ADDR:%.*]] = alloca %struct.S*, align 4 // CHECK: store %struct.S* %a, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: store %struct.S* %b, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: store %struct.S* %c, %struct.S** [[C_ADDR]], align 4 // CHECK-NEXT: [[LOAD_A_PTR:%.*]] = load %struct.S*, %struct.S** [[A_ADDR]], align 4 // CHECK-NEXT: [[LOAD_B_PTR:%.*]] = load %struct.S*, %struct.S** [[B_ADDR]], align 4 // CHECK-NEXT: [[LOAD_C_PTR:%.*]] = load %struct.S*, %struct.S** [[C_ADDR]], align 4 // CHECK-NEXT: [[COERCED_A_TMP:%.*]] = bitcast %struct.S* [[LOAD_A_PTR]] to i64* // CHECK-NEXT: [[COERCED_B_TMP:%.*]] = bitcast %struct.S* [[LOAD_B_PTR]] to i64* // CHECK-NEXT: [[COERCED_C:%.*]] = bitcast %struct.S* [[LOAD_C_PTR]] to i64* // CHECK-NEXT: [[COERCED_A:%.*]] = bitcast i64* [[COERCED_A_TMP]] to i8* // CHECK-NEXT: [[COERCED_B:%.*]] = bitcast i64* [[COERCED_B_TMP]] to i8* // CHECK-NEXT: [[CAST_C:%.*]] = bitcast i64* [[COERCED_C]] to i8* // CHECK-NEXT: [[CALL:%.*]] = call zeroext i1 @__atomic_compare_exchange(i32 8, i8* [[COERCED_A]], i8* [[COERCED_B]], i8* [[CAST_C]], // CHECK-NEXT: ret i1 [[CALL]] return __atomic_compare_exchange(a, b, c, 1, 5, 5); } int* fp1(_Atomic(int*) *p) { // CHECK-LABEL: @fp1 // CHECK: load atomic i32, i32* {{.*}} seq_cst return __c11_atomic_load(p, memory_order_seq_cst); } int* fp2(_Atomic(int*) *p) { // CHECK-LABEL: @fp2 // CHECK: store i32 4 // CHECK: atomicrmw add {{.*}} monotonic return __c11_atomic_fetch_add(p, 1, memory_order_relaxed); } int *fp2a(int **p) { // CHECK-LABEL: @fp2a // CHECK: store i32 4 // CHECK: atomicrmw sub {{.*}} monotonic // Note, the GNU builtins do not multiply by sizeof(T)! return __atomic_fetch_sub(p, 4, memory_order_relaxed); } _Complex float fc(_Atomic(_Complex float) *c) { // CHECK-LABEL: @fc // CHECK: atomicrmw xchg i64* return __c11_atomic_exchange(c, 2, memory_order_seq_cst); } typedef struct X { int x; } X; X fs(_Atomic(X) *c) { // CHECK-LABEL: @fs // CHECK: atomicrmw xchg i32* return __c11_atomic_exchange(c, (X){2}, memory_order_seq_cst); } X fsa(X *c, X *d) { // CHECK-LABEL: @fsa // CHECK: atomicrmw xchg i32* X ret; __atomic_exchange(c, d, &ret, memory_order_seq_cst); return ret; } _Bool fsb(_Bool *c) { // CHECK-LABEL: @fsb // CHECK: atomicrmw xchg i8* return __atomic_exchange_n(c, 1, memory_order_seq_cst); } char flag1; volatile char flag2; void test_and_set() { // CHECK: atomicrmw xchg i8* @flag1, i8 1 seq_cst __atomic_test_and_set(&flag1, memory_order_seq_cst); // CHECK: atomicrmw volatile xchg i8* @flag2, i8 1 acquire __atomic_test_and_set(&flag2, memory_order_acquire); // CHECK: store atomic volatile i8 0, i8* @flag2 release __atomic_clear(&flag2, memory_order_release); // CHECK: store atomic i8 0, i8* @flag1 seq_cst __atomic_clear(&flag1, memory_order_seq_cst); } struct Sixteen { char c[16]; } sixteen; struct Seventeen { char c[17]; } seventeen; struct Incomplete; int lock_free(struct Incomplete *incomplete) { // CHECK-LABEL: @lock_free // CHECK: call zeroext i1 @__atomic_is_lock_free(i32 3, i8* null) __c11_atomic_is_lock_free(3); // CHECK: call zeroext i1 @__atomic_is_lock_free(i32 16, i8* {{.*}}@sixteen{{.*}}) __atomic_is_lock_free(16, &sixteen); // CHECK: call zeroext i1 @__atomic_is_lock_free(i32 17, i8* {{.*}}@seventeen{{.*}}) __atomic_is_lock_free(17, &seventeen); // CHECK: call zeroext i1 @__atomic_is_lock_free(i32 4, {{.*}}) __atomic_is_lock_free(4, incomplete); char cs[20]; // CHECK: call zeroext i1 @__atomic_is_lock_free(i32 4, {{.*}}) __atomic_is_lock_free(4, cs+1); // CHECK-NOT: call __atomic_always_lock_free(3, 0); __atomic_always_lock_free(16, 0); __atomic_always_lock_free(17, 0); __atomic_always_lock_free(16, &sixteen); __atomic_always_lock_free(17, &seventeen); int n; __atomic_is_lock_free(4, &n); // CHECK: ret i32 1 return __c11_atomic_is_lock_free(sizeof(_Atomic(int))); } // Tests for atomic operations on big values. These should call the functions // defined here: // http://gcc.gnu.org/wiki/Atomic/GCCMM/LIbrary#The_Library_interface struct foo { int big[128]; }; struct bar { char c[3]; }; struct bar smallThing, thing1, thing2; struct foo bigThing; _Atomic(struct foo) bigAtomic; void structAtomicStore() { // CHECK-LABEL: @structAtomicStore struct foo f = {0}; struct bar b = {0}; __atomic_store(&smallThing, &b, 5); // CHECK: call void @__atomic_store(i32 3, i8* {{.*}} @smallThing __atomic_store(&bigThing, &f, 5); // CHECK: call void @__atomic_store(i32 512, i8* {{.*}} @bigThing } void structAtomicLoad() { // CHECK-LABEL: @structAtomicLoad struct bar b; __atomic_load(&smallThing, &b, 5); // CHECK: call void @__atomic_load(i32 3, i8* {{.*}} @smallThing struct foo f = {0}; __atomic_load(&bigThing, &f, 5); // CHECK: call void @__atomic_load(i32 512, i8* {{.*}} @bigThing } struct foo structAtomicExchange() { // CHECK-LABEL: @structAtomicExchange struct foo f = {0}; struct foo old; __atomic_exchange(&f, &bigThing, &old, 5); // CHECK: call void @__atomic_exchange(i32 512, {{.*}}, i8* bitcast ({{.*}} @bigThing to i8*), return __c11_atomic_exchange(&bigAtomic, f, 5); // CHECK: call void @__atomic_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*), } int structAtomicCmpExchange() { // CHECK-LABEL: @structAtomicCmpExchange // CHECK: %[[x_mem:.*]] = alloca i8 _Bool x = __atomic_compare_exchange(&smallThing, &thing1, &thing2, 1, 5, 5); // CHECK: %[[call1:.*]] = call zeroext i1 @__atomic_compare_exchange(i32 3, {{.*}} @smallThing{{.*}} @thing1{{.*}} @thing2 // CHECK: %[[zext1:.*]] = zext i1 %[[call1]] to i8 // CHECK: store i8 %[[zext1]], i8* %[[x_mem]], align 1 // CHECK: %[[x:.*]] = load i8, i8* %[[x_mem]] // CHECK: %[[x_bool:.*]] = trunc i8 %[[x]] to i1 // CHECK: %[[conv1:.*]] = zext i1 %[[x_bool]] to i32 struct foo f = {0}; struct foo g = {0}; g.big[12] = 12; return x & __c11_atomic_compare_exchange_strong(&bigAtomic, &f, g, 5, 5); // CHECK: %[[call2:.*]] = call zeroext i1 @__atomic_compare_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*), // CHECK: %[[conv2:.*]] = zext i1 %[[call2]] to i32 // CHECK: %[[and:.*]] = and i32 %[[conv1]], %[[conv2]] // CHECK: ret i32 %[[and]] } // Check that no atomic operations are used in any initialisation of _Atomic // types. _Atomic(int) atomic_init_i = 42; // CHECK-LABEL: @atomic_init_foo void atomic_init_foo() { // CHECK-NOT: } // CHECK-NOT: atomic // CHECK: store _Atomic(int) j = 12; // CHECK-NOT: } // CHECK-NOT: atomic // CHECK: store __c11_atomic_init(&j, 42); // CHECK-NOT: atomic // CHECK: } } // CHECK-LABEL: @failureOrder void failureOrder(_Atomic(int) *ptr, int *ptr2) { __c11_atomic_compare_exchange_strong(ptr, ptr2, 43, memory_order_acquire, memory_order_relaxed); // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acquire monotonic __c11_atomic_compare_exchange_weak(ptr, ptr2, 43, memory_order_seq_cst, memory_order_acquire); // CHECK: cmpxchg weak i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst acquire // Unknown ordering: conservatively pick strongest valid option (for now!). __atomic_compare_exchange(ptr2, ptr2, ptr2, 0, memory_order_acq_rel, *ptr2); // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acq_rel acquire // Undefined behaviour: don't really care what that last ordering is so leave // it out: __atomic_compare_exchange_n(ptr2, ptr2, 43, 1, memory_order_seq_cst, 42); // CHECK: cmpxchg weak i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst } // CHECK-LABEL: @generalFailureOrder void generalFailureOrder(_Atomic(int) *ptr, int *ptr2, int success, int fail) { __c11_atomic_compare_exchange_strong(ptr, ptr2, 42, success, fail); // CHECK: switch i32 {{.*}}, label %[[MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: i32 1, label %[[ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 2, label %[[ACQUIRE]] // CHECK-NEXT: i32 3, label %[[RELEASE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 4, label %[[ACQREL:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 5, label %[[SEQCST:[0-9a-zA-Z._]+]] // CHECK: [[MONOTONIC]] // CHECK: switch {{.*}}, label %[[MONOTONIC_MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: ] // CHECK: [[ACQUIRE]] // CHECK: switch {{.*}}, label %[[ACQUIRE_MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: i32 1, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 2, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: ] // CHECK: [[RELEASE]] // CHECK: switch {{.*}}, label %[[RELEASE_MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: ] // CHECK: [[ACQREL]] // CHECK: switch {{.*}}, label %[[ACQREL_MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: i32 1, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 2, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: ] // CHECK: [[SEQCST]] // CHECK: switch {{.*}}, label %[[SEQCST_MONOTONIC:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: i32 1, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 2, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]] // CHECK-NEXT: i32 5, label %[[SEQCST_SEQCST:[0-9a-zA-Z._]+]] // CHECK-NEXT: ] // CHECK: [[MONOTONIC_MONOTONIC]] // CHECK: cmpxchg {{.*}} monotonic monotonic // CHECK: br // CHECK: [[ACQUIRE_MONOTONIC]] // CHECK: cmpxchg {{.*}} acquire monotonic // CHECK: br // CHECK: [[ACQUIRE_ACQUIRE]] // CHECK: cmpxchg {{.*}} acquire acquire // CHECK: br // CHECK: [[ACQREL_MONOTONIC]] // CHECK: cmpxchg {{.*}} acq_rel monotonic // CHECK: br // CHECK: [[ACQREL_ACQUIRE]] // CHECK: cmpxchg {{.*}} acq_rel acquire // CHECK: br // CHECK: [[SEQCST_MONOTONIC]] // CHECK: cmpxchg {{.*}} seq_cst monotonic // CHECK: br // CHECK: [[SEQCST_ACQUIRE]] // CHECK: cmpxchg {{.*}} seq_cst acquire // CHECK: br // CHECK: [[SEQCST_SEQCST]] // CHECK: cmpxchg {{.*}} seq_cst seq_cst // CHECK: br } void generalWeakness(int *ptr, int *ptr2, _Bool weak) { __atomic_compare_exchange_n(ptr, ptr2, 42, weak, memory_order_seq_cst, memory_order_seq_cst); // CHECK: switch i1 {{.*}}, label %[[WEAK:[0-9a-zA-Z._]+]] [ // CHECK-NEXT: i1 false, label %[[STRONG:[0-9a-zA-Z._]+]] // CHECK: [[STRONG]] // CHECK-NOT: br // CHECK: cmpxchg {{.*}} seq_cst seq_cst // CHECK: br // CHECK: [[WEAK]] // CHECK-NOT: br // CHECK: cmpxchg weak {{.*}} seq_cst seq_cst // CHECK: br } // Having checked the flow in the previous two cases, we'll trust clang to // combine them sanely. void EMIT_ALL_THE_THINGS(int *ptr, int *ptr2, int new, _Bool weak, int success, int fail) { __atomic_compare_exchange(ptr, ptr2, &new, weak, success, fail); // CHECK: = cmpxchg {{.*}} monotonic monotonic // CHECK: = cmpxchg weak {{.*}} monotonic monotonic // CHECK: = cmpxchg {{.*}} acquire monotonic // CHECK: = cmpxchg {{.*}} acquire acquire // CHECK: = cmpxchg weak {{.*}} acquire monotonic // CHECK: = cmpxchg weak {{.*}} acquire acquire // CHECK: = cmpxchg {{.*}} release monotonic // CHECK: = cmpxchg weak {{.*}} release monotonic // CHECK: = cmpxchg {{.*}} acq_rel monotonic // CHECK: = cmpxchg {{.*}} acq_rel acquire // CHECK: = cmpxchg weak {{.*}} acq_rel monotonic // CHECK: = cmpxchg weak {{.*}} acq_rel acquire // CHECK: = cmpxchg {{.*}} seq_cst monotonic // CHECK: = cmpxchg {{.*}} seq_cst acquire // CHECK: = cmpxchg {{.*}} seq_cst seq_cst // CHECK: = cmpxchg weak {{.*}} seq_cst monotonic // CHECK: = cmpxchg weak {{.*}} seq_cst acquire // CHECK: = cmpxchg weak {{.*}} seq_cst seq_cst } int PR21643() { return __atomic_or_fetch((int __attribute__((address_space(257))) *)0x308, 1, __ATOMIC_RELAXED); // CHECK: %[[atomictmp:.*]] = alloca i32, align 4 // CHECK: %[[atomicdst:.*]] = alloca i32, align 4 // CHECK: store i32 1, i32* %[[atomictmp]] // CHECK: %[[one:.*]] = load i32, i32* %[[atomictmp]], align 4 // CHECK: %[[old:.*]] = atomicrmw or i32 addrspace(257)* inttoptr (i32 776 to i32 addrspace(257)*), i32 %[[one]] monotonic // CHECK: %[[new:.*]] = or i32 %[[old]], %[[one]] // CHECK: store i32 %[[new]], i32* %[[atomicdst]], align 4 // CHECK: %[[ret:.*]] = load i32, i32* %[[atomicdst]], align 4 // CHECK: ret i32 %[[ret]] } int PR17306_1(volatile _Atomic(int) *i) { // CHECK-LABEL: @PR17306_1 // CHECK: %[[i_addr:.*]] = alloca i32 // CHECK-NEXT: %[[atomicdst:.*]] = alloca i32 // CHECK-NEXT: store i32* %i, i32** %[[i_addr]] // CHECK-NEXT: %[[addr:.*]] = load i32*, i32** %[[i_addr]] // CHECK-NEXT: %[[res:.*]] = load atomic volatile i32, i32* %[[addr]] seq_cst // CHECK-NEXT: store i32 %[[res]], i32* %[[atomicdst]] // CHECK-NEXT: %[[retval:.*]] = load i32, i32* %[[atomicdst]] // CHECK-NEXT: ret i32 %[[retval]] return __c11_atomic_load(i, memory_order_seq_cst); } int PR17306_2(volatile int *i, int value) { // CHECK-LABEL: @PR17306_2 // CHECK: %[[i_addr:.*]] = alloca i32* // CHECK-NEXT: %[[value_addr:.*]] = alloca i32 // CHECK-NEXT: %[[atomictmp:.*]] = alloca i32 // CHECK-NEXT: %[[atomicdst:.*]] = alloca i32 // CHECK-NEXT: store i32* %i, i32** %[[i_addr]] // CHECK-NEXT: store i32 %value, i32* %[[value_addr]] // CHECK-NEXT: %[[i_lval:.*]] = load i32*, i32** %[[i_addr]] // CHECK-NEXT: %[[value:.*]] = load i32, i32* %[[value_addr]] // CHECK-NEXT: store i32 %[[value]], i32* %[[atomictmp]] // CHECK-NEXT: %[[value_lval:.*]] = load i32, i32* %[[atomictmp]] // CHECK-NEXT: %[[old_val:.*]] = atomicrmw volatile add i32* %[[i_lval]], i32 %[[value_lval]] seq_cst // CHECK-NEXT: %[[new_val:.*]] = add i32 %[[old_val]], %[[value_lval]] // CHECK-NEXT: store i32 %[[new_val]], i32* %[[atomicdst]] // CHECK-NEXT: %[[retval:.*]] = load i32, i32* %[[atomicdst]] // CHECK-NEXT: ret i32 %[[retval]] return __atomic_add_fetch(i, value, memory_order_seq_cst); } void test_underaligned() { // CHECK-LABEL: @test_underaligned struct Underaligned { char c[8]; } underaligned_a, underaligned_b, underaligned_c; // CHECK: call void @__atomic_load(i32 8, __atomic_load(&underaligned_a, &underaligned_b, memory_order_seq_cst); // CHECK: call void @__atomic_store(i32 8, __atomic_store(&underaligned_a, &underaligned_b, memory_order_seq_cst); // CHECK: call void @__atomic_exchange(i32 8, __atomic_exchange(&underaligned_a, &underaligned_b, &underaligned_c, memory_order_seq_cst); // CHECK: call {{.*}} @__atomic_compare_exchange(i32 8, __atomic_compare_exchange(&underaligned_a, &underaligned_b, &underaligned_c, 1, memory_order_seq_cst, memory_order_seq_cst); __attribute__((aligned)) struct Underaligned aligned_a, aligned_b, aligned_c; // CHECK: load atomic __atomic_load(&aligned_a, &aligned_b, memory_order_seq_cst); // CHECK: store atomic __atomic_store(&aligned_a, &aligned_b, memory_order_seq_cst); // CHECK: atomicrmw xchg __atomic_exchange(&aligned_a, &aligned_b, &aligned_c, memory_order_seq_cst); // CHECK: cmpxchg weak __atomic_compare_exchange(&aligned_a, &aligned_b, &aligned_c, 1, memory_order_seq_cst, memory_order_seq_cst); } void test_c11_minmax(_Atomic(int) * si, _Atomic(unsigned) * ui, _Atomic(short) * ss, _Atomic(unsigned char) * uc, _Atomic(long long) * sll) { // CHECK-LABEL: @test_c11_minmax // CHECK: atomicrmw max i32 *si = __c11_atomic_fetch_max(si, 42, memory_order_acquire); // CHECK: atomicrmw min i32 *si = __c11_atomic_fetch_min(si, 42, memory_order_acquire); // CHECK: atomicrmw umax i32 *ui = __c11_atomic_fetch_max(ui, 42, memory_order_acquire); // CHECK: atomicrmw umin i32 *ui = __c11_atomic_fetch_min(ui, 42, memory_order_acquire); // CHECK: atomicrmw max i16 *ss = __c11_atomic_fetch_max(ss, 42, memory_order_acquire); // CHECK: atomicrmw min i16 *ss = __c11_atomic_fetch_min(ss, 42, memory_order_acquire); // CHECK: atomicrmw umax i8 *uc = __c11_atomic_fetch_max(uc, 42, memory_order_acquire); // CHECK: atomicrmw umin i8 *uc = __c11_atomic_fetch_min(uc, 42, memory_order_acquire); // CHECK: atomicrmw max i64 *sll = __c11_atomic_fetch_max(sll, 42, memory_order_acquire); // CHECK: atomicrmw min i64 *sll = __c11_atomic_fetch_min(sll, 42, memory_order_acquire); } void test_minmax_postop(int *si, unsigned *ui, unsigned short *us, signed char *sc, unsigned long long *ull) { int val = 42; // CHECK-LABEL: @test_minmax_postop // CHECK: [[OLD:%.*]] = atomicrmw max i32* [[PTR:%.*]], i32 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp sgt i32 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i32 [[OLD]], i32 [[RHS]] // CHECK: store i32 [[NEW]], i32* *si = __atomic_max_fetch(si, 42, memory_order_release); // CHECK: [[OLD:%.*]] = atomicrmw min i32* [[PTR:%.*]], i32 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp slt i32 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i32 [[OLD]], i32 [[RHS]] // CHECK: store i32 [[NEW]], i32* *si = __atomic_min_fetch(si, 42, memory_order_release); // CHECK: [[OLD:%.*]] = atomicrmw umax i32* [[PTR:%.*]], i32 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp ugt i32 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i32 [[OLD]], i32 [[RHS]] // CHECK: store i32 [[NEW]], i32* *ui = __atomic_max_fetch(ui, 42, memory_order_release); // CHECK: [[OLD:%.*]] = atomicrmw umin i32* [[PTR:%.*]], i32 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp ult i32 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i32 [[OLD]], i32 [[RHS]] // CHECK: store i32 [[NEW]], i32* *ui = __atomic_min_fetch(ui, 42, memory_order_release); // CHECK: [[OLD:%.*]] = atomicrmw umin i16* [[PTR:%.*]], i16 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp ult i16 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i16 [[OLD]], i16 [[RHS]] // CHECK: store i16 [[NEW]], i16* *us = __atomic_min_fetch(us, 42, memory_order_release); // CHECK: [[OLD:%.*]] = atomicrmw min i8* [[PTR:%.*]], i8 [[RHS:%.*]] release // CHECK: [[TST:%.*]] = icmp slt i8 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i8 [[OLD]], i8 [[RHS]] // CHECK: store i8 [[NEW]], i8* *sc = __atomic_min_fetch(sc, 42, memory_order_release); // CHECK: [[OLD:%.*]] = call i64 @__atomic_fetch_umin_8(i8* {{%.*}}, i64 [[RHS:%.*]], // CHECK: [[TST:%.*]] = icmp ult i64 [[OLD]], [[RHS]] // CHECK: [[NEW:%.*]] = select i1 [[TST]], i64 [[OLD]], i64 [[RHS]] // CHECK: store i64 [[NEW]], i64* *ull = __atomic_min_fetch(ull, 42, memory_order_release); } #endif