| // RUN: %clang_cc1 %s -emit-llvm -o - -triple=i686-apple-darwin9 | FileCheck %s |
| |
| // Also test serialization of atomic operations here, to avoid duplicating the |
| // test. |
| // RUN: %clang_cc1 %s -emit-pch -o %t -triple=i686-apple-darwin9 |
| // RUN: %clang_cc1 %s -include-pch %t -triple=i686-apple-darwin9 -emit-llvm -o - | FileCheck %s |
| #ifndef ALREADY_INCLUDED |
| #define ALREADY_INCLUDED |
| |
| // Basic IRGen tests for __c11_atomic_* and GNU __atomic_* |
| |
| typedef enum memory_order { |
| memory_order_relaxed, memory_order_consume, memory_order_acquire, |
| memory_order_release, memory_order_acq_rel, memory_order_seq_cst |
| } memory_order; |
| |
| int fi1(_Atomic(int) *i) { |
| // CHECK: @fi1 |
| // CHECK: load atomic i32* {{.*}} seq_cst |
| return __c11_atomic_load(i, memory_order_seq_cst); |
| } |
| |
| int fi1a(int *i) { |
| // CHECK: @fi1a |
| // CHECK: load atomic i32* {{.*}} seq_cst |
| int v; |
| __atomic_load(i, &v, memory_order_seq_cst); |
| return v; |
| } |
| |
| int fi1b(int *i) { |
| // CHECK: @fi1b |
| // CHECK: load atomic i32* {{.*}} seq_cst |
| return __atomic_load_n(i, memory_order_seq_cst); |
| } |
| |
| void fi2(_Atomic(int) *i) { |
| // CHECK: @fi2 |
| // CHECK: store atomic i32 {{.*}} seq_cst |
| __c11_atomic_store(i, 1, memory_order_seq_cst); |
| } |
| |
| void fi2a(int *i) { |
| // CHECK: @fi2a |
| // CHECK: store atomic i32 {{.*}} seq_cst |
| int v = 1; |
| __atomic_store(i, &v, memory_order_seq_cst); |
| } |
| |
| void fi2b(int *i) { |
| // CHECK: @fi2b |
| // CHECK: store atomic i32 {{.*}} seq_cst |
| __atomic_store_n(i, 1, memory_order_seq_cst); |
| } |
| |
| int fi3(_Atomic(int) *i) { |
| // CHECK: @fi3 |
| // CHECK: atomicrmw and |
| // CHECK-NOT: and |
| return __c11_atomic_fetch_and(i, 1, memory_order_seq_cst); |
| } |
| |
| int fi3a(int *i) { |
| // CHECK: @fi3a |
| // CHECK: atomicrmw xor |
| // CHECK-NOT: xor |
| return __atomic_fetch_xor(i, 1, memory_order_seq_cst); |
| } |
| |
| int fi3b(int *i) { |
| // CHECK: @fi3b |
| // CHECK: atomicrmw add |
| // CHECK: add |
| return __atomic_add_fetch(i, 1, memory_order_seq_cst); |
| } |
| |
| int fi3c(int *i) { |
| // CHECK: @fi3c |
| // CHECK: atomicrmw nand |
| // CHECK-NOT: and |
| return __atomic_fetch_nand(i, 1, memory_order_seq_cst); |
| } |
| |
| int fi3d(int *i) { |
| // CHECK: @fi3d |
| // CHECK: atomicrmw nand |
| // CHECK: and |
| // CHECK: xor |
| return __atomic_nand_fetch(i, 1, memory_order_seq_cst); |
| } |
| |
| _Bool fi4(_Atomic(int) *i) { |
| // CHECK: @fi4 |
| // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] |
| // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = icmp eq i32 [[OLD]], [[EXPECTED]] |
| // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] |
| // CHECK: store i32 [[OLD]] |
| int cmp = 0; |
| return __c11_atomic_compare_exchange_strong(i, &cmp, 1, memory_order_acquire, memory_order_acquire); |
| } |
| |
| _Bool fi4a(int *i) { |
| // CHECK: @fi4 |
| // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] |
| // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = icmp eq i32 [[OLD]], [[EXPECTED]] |
| // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] |
| // CHECK: store i32 [[OLD]] |
| int cmp = 0; |
| int desired = 1; |
| return __atomic_compare_exchange(i, &cmp, &desired, 0, memory_order_acquire, memory_order_acquire); |
| } |
| |
| _Bool fi4b(int *i) { |
| // CHECK: @fi4 |
| // CHECK: [[OLD:%[.0-9A-Z_a-z]+]] = cmpxchg i32* [[PTR:%[.0-9A-Z_a-z]+]], i32 [[EXPECTED:%[.0-9A-Z_a-z]+]], i32 [[DESIRED:%[.0-9A-Z_a-z]+]] |
| // CHECK: [[CMP:%[.0-9A-Z_a-z]+]] = icmp eq i32 [[OLD]], [[EXPECTED]] |
| // CHECK: br i1 [[CMP]], label %[[STORE_EXPECTED:[.0-9A-Z_a-z]+]], label %[[CONTINUE:[.0-9A-Z_a-z]+]] |
| // CHECK: store i32 [[OLD]] |
| int cmp = 0; |
| return __atomic_compare_exchange_n(i, &cmp, 1, 1, memory_order_acquire, memory_order_acquire); |
| } |
| |
| float ff1(_Atomic(float) *d) { |
| // CHECK: @ff1 |
| // CHECK: load atomic i32* {{.*}} monotonic |
| return __c11_atomic_load(d, memory_order_relaxed); |
| } |
| |
| void ff2(_Atomic(float) *d) { |
| // CHECK: @ff2 |
| // CHECK: store atomic i32 {{.*}} release |
| __c11_atomic_store(d, 1, memory_order_release); |
| } |
| |
| float ff3(_Atomic(float) *d) { |
| return __c11_atomic_exchange(d, 2, memory_order_seq_cst); |
| } |
| |
| int* fp1(_Atomic(int*) *p) { |
| // CHECK: @fp1 |
| // CHECK: load atomic i32* {{.*}} seq_cst |
| return __c11_atomic_load(p, memory_order_seq_cst); |
| } |
| |
| int* fp2(_Atomic(int*) *p) { |
| // CHECK: @fp2 |
| // CHECK: store i32 4 |
| // CHECK: atomicrmw add {{.*}} monotonic |
| return __c11_atomic_fetch_add(p, 1, memory_order_relaxed); |
| } |
| |
| int *fp2a(int **p) { |
| // CHECK: @fp2a |
| // CHECK: store i32 4 |
| // CHECK: atomicrmw sub {{.*}} monotonic |
| // Note, the GNU builtins do not multiply by sizeof(T)! |
| return __atomic_fetch_sub(p, 4, memory_order_relaxed); |
| } |
| |
| _Complex float fc(_Atomic(_Complex float) *c) { |
| // CHECK: @fc |
| // CHECK: atomicrmw xchg i64* |
| return __c11_atomic_exchange(c, 2, memory_order_seq_cst); |
| } |
| |
| typedef struct X { int x; } X; |
| X fs(_Atomic(X) *c) { |
| // CHECK: @fs |
| // CHECK: atomicrmw xchg i32* |
| return __c11_atomic_exchange(c, (X){2}, memory_order_seq_cst); |
| } |
| |
| X fsa(X *c, X *d) { |
| // CHECK: @fsa |
| // CHECK: atomicrmw xchg i32* |
| X ret; |
| __atomic_exchange(c, d, &ret, memory_order_seq_cst); |
| return ret; |
| } |
| |
| _Bool fsb(_Bool *c) { |
| // CHECK: @fsb |
| // CHECK: atomicrmw xchg i8* |
| return __atomic_exchange_n(c, 1, memory_order_seq_cst); |
| } |
| |
| char flag1; |
| volatile char flag2; |
| void test_and_set() { |
| // CHECK: atomicrmw xchg i8* @flag1, i8 1 seq_cst |
| __atomic_test_and_set(&flag1, memory_order_seq_cst); |
| // CHECK: atomicrmw volatile xchg i8* @flag2, i8 1 acquire |
| __atomic_test_and_set(&flag2, memory_order_acquire); |
| // CHECK: store atomic volatile i8 0, i8* @flag2 release |
| __atomic_clear(&flag2, memory_order_release); |
| // CHECK: store atomic i8 0, i8* @flag1 seq_cst |
| __atomic_clear(&flag1, memory_order_seq_cst); |
| } |
| |
| struct Sixteen { |
| char c[16]; |
| } sixteen; |
| struct Seventeen { |
| char c[17]; |
| } seventeen; |
| |
| int lock_free(struct Incomplete *incomplete) { |
| // CHECK: @lock_free |
| |
| // CHECK: call i32 @__atomic_is_lock_free(i32 3, i8* null) |
| __c11_atomic_is_lock_free(3); |
| |
| // CHECK: call i32 @__atomic_is_lock_free(i32 16, i8* {{.*}}@sixteen{{.*}}) |
| __atomic_is_lock_free(16, &sixteen); |
| |
| // CHECK: call i32 @__atomic_is_lock_free(i32 17, i8* {{.*}}@seventeen{{.*}}) |
| __atomic_is_lock_free(17, &seventeen); |
| |
| // CHECK: call i32 @__atomic_is_lock_free(i32 4, {{.*}}) |
| __atomic_is_lock_free(4, incomplete); |
| |
| char cs[20]; |
| // CHECK: call i32 @__atomic_is_lock_free(i32 4, {{.*}}) |
| __atomic_is_lock_free(4, cs+1); |
| |
| // CHECK-NOT: call |
| __atomic_always_lock_free(3, 0); |
| __atomic_always_lock_free(16, 0); |
| __atomic_always_lock_free(17, 0); |
| __atomic_always_lock_free(16, &sixteen); |
| __atomic_always_lock_free(17, &seventeen); |
| |
| int n; |
| __atomic_is_lock_free(4, &n); |
| |
| // CHECK: ret i32 1 |
| return __c11_atomic_is_lock_free(sizeof(_Atomic(int))); |
| } |
| |
| // Tests for atomic operations on big values. These should call the functions |
| // defined here: |
| // http://gcc.gnu.org/wiki/Atomic/GCCMM/LIbrary#The_Library_interface |
| |
| struct foo { |
| int big[128]; |
| }; |
| struct bar { |
| char c[3]; |
| }; |
| |
| struct bar smallThing, thing1, thing2; |
| struct foo bigThing; |
| _Atomic(struct foo) bigAtomic; |
| |
| void structAtomicStore() { |
| // CHECK: @structAtomicStore |
| struct foo f = {0}; |
| struct bar b = {0}; |
| __atomic_store(&smallThing, &b, 5); |
| // CHECK: call void @__atomic_store(i32 3, i8* {{.*}} @smallThing |
| |
| __atomic_store(&bigThing, &f, 5); |
| // CHECK: call void @__atomic_store(i32 512, i8* {{.*}} @bigThing |
| } |
| void structAtomicLoad() { |
| // CHECK: @structAtomicLoad |
| struct bar b; |
| __atomic_load(&smallThing, &b, 5); |
| // CHECK: call void @__atomic_load(i32 3, i8* {{.*}} @smallThing |
| |
| struct foo f = {0}; |
| __atomic_load(&bigThing, &f, 5); |
| // CHECK: call void @__atomic_load(i32 512, i8* {{.*}} @bigThing |
| } |
| struct foo structAtomicExchange() { |
| // CHECK: @structAtomicExchange |
| struct foo f = {0}; |
| struct foo old; |
| __atomic_exchange(&f, &bigThing, &old, 5); |
| // CHECK: call void @__atomic_exchange(i32 512, {{.*}}, i8* bitcast ({{.*}} @bigThing to i8*), |
| |
| return __c11_atomic_exchange(&bigAtomic, f, 5); |
| // CHECK: call void @__atomic_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*), |
| } |
| int structAtomicCmpExchange() { |
| // CHECK: @structAtomicCmpExchange |
| _Bool x = __atomic_compare_exchange(&smallThing, &thing1, &thing2, 1, 5, 5); |
| // CHECK: call zeroext i1 @__atomic_compare_exchange(i32 3, {{.*}} @smallThing{{.*}} @thing1{{.*}} @thing2 |
| |
| struct foo f = {0}; |
| struct foo g = {0}; |
| g.big[12] = 12; |
| return x & __c11_atomic_compare_exchange_strong(&bigAtomic, &f, g, 5, 5); |
| // CHECK: call zeroext i1 @__atomic_compare_exchange(i32 512, i8* bitcast ({{.*}} @bigAtomic to i8*), |
| } |
| |
| // Check that no atomic operations are used in any initialisation of _Atomic |
| // types. |
| _Atomic(int) atomic_init_i = 42; |
| |
| // CHECK: @atomic_init_foo |
| void atomic_init_foo() |
| { |
| // CHECK-NOT: } |
| // CHECK-NOT: atomic |
| // CHECK: store |
| _Atomic(int) j = 12; |
| |
| // CHECK-NOT: } |
| // CHECK-NOT: atomic |
| // CHECK: store |
| __c11_atomic_init(&j, 42); |
| |
| // CHECK-NOT: atomic |
| // CHECK: } |
| } |
| |
| // CHECK-LABEL: @failureOrder |
| void failureOrder(_Atomic(int) *ptr, int *ptr2) { |
| __c11_atomic_compare_exchange_strong(ptr, ptr2, 43, memory_order_acquire, memory_order_relaxed); |
| // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acquire monotonic |
| |
| __c11_atomic_compare_exchange_weak(ptr, ptr2, 43, memory_order_seq_cst, memory_order_acquire); |
| // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst acquire |
| |
| // Unknown ordering: conservatively pick strongest valid option (for now!). |
| __atomic_compare_exchange(ptr2, ptr2, ptr2, 0, memory_order_acq_rel, *ptr2); |
| // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} acq_rel acquire |
| |
| // Undefined behaviour: don't really care what that last ordering is so leave |
| // it out: |
| __atomic_compare_exchange_n(ptr2, ptr2, 43, 1, memory_order_seq_cst, 42); |
| // CHECK: cmpxchg i32* {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z._]+}}, i32 {{%[0-9A-Za-z_.]+}} seq_cst |
| } |
| |
| // CHECK-LABEL: @generalFailureOrder |
| void generalFailureOrder(_Atomic(int) *ptr, int *ptr2, int success, int fail) { |
| __c11_atomic_compare_exchange_strong(ptr, ptr2, 42, success, fail); |
| // CHECK: switch i32 {{.*}}, label %[[MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: i32 1, label %[[ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 2, label %[[ACQUIRE]] |
| // CHECK-NEXT: i32 3, label %[[RELEASE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 4, label %[[ACQREL:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 5, label %[[SEQCST:[0-9a-zA-Z._]+]] |
| |
| // CHECK: [[MONOTONIC]] |
| // CHECK: switch {{.*}}, label %[[MONOTONIC_MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: ] |
| |
| // CHECK: [[ACQUIRE]] |
| // CHECK: switch {{.*}}, label %[[ACQUIRE_MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: i32 1, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 2, label %[[ACQUIRE_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: ] |
| |
| // CHECK: [[RELEASE]] |
| // CHECK: switch {{.*}}, label %[[RELEASE_MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: ] |
| |
| // CHECK: [[ACQREL]] |
| // CHECK: switch {{.*}}, label %[[ACQREL_MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: i32 1, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 2, label %[[ACQREL_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: ] |
| |
| // CHECK: [[SEQCST]] |
| // CHECK: switch {{.*}}, label %[[SEQCST_MONOTONIC:[0-9a-zA-Z._]+]] [ |
| // CHECK-NEXT: i32 1, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 2, label %[[SEQCST_ACQUIRE:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: i32 5, label %[[SEQCST_SEQCST:[0-9a-zA-Z._]+]] |
| // CHECK-NEXT: ] |
| |
| // CHECK: [[MONOTONIC_MONOTONIC]] |
| // CHECK: cmpxchg {{.*}} monotonic monotonic |
| // CHECK: br |
| |
| // CHECK: [[ACQUIRE_MONOTONIC]] |
| // CHECK: cmpxchg {{.*}} acquire monotonic |
| // CHECK: br |
| |
| // CHECK: [[ACQUIRE_ACQUIRE]] |
| // CHECK: cmpxchg {{.*}} acquire acquire |
| // CHECK: br |
| |
| // CHECK: [[ACQREL_MONOTONIC]] |
| // CHECK: cmpxchg {{.*}} acq_rel monotonic |
| // CHECK: br |
| |
| // CHECK: [[ACQREL_ACQUIRE]] |
| // CHECK: cmpxchg {{.*}} acq_rel acquire |
| // CHECK: br |
| |
| // CHECK: [[SEQCST_MONOTONIC]] |
| // CHECK: cmpxchg {{.*}} seq_cst monotonic |
| // CHECK: br |
| |
| // CHECK: [[SEQCST_ACQUIRE]] |
| // CHECK: cmpxchg {{.*}} seq_cst acquire |
| // CHECK: br |
| |
| // CHECK: [[SEQCST_SEQCST]] |
| // CHECK: cmpxchg {{.*}} seq_cst seq_cst |
| // CHECK: br |
| } |
| |
| #endif |