Home | History | Annotate | Download | only in CodeGen
      1 // RUN: %clang_cc1 < %s -triple armv5e-none-linux-gnueabi -emit-llvm -O1 | FileCheck %s
      2 
      3 enum memory_order {
      4   memory_order_relaxed, memory_order_consume, memory_order_acquire,
      5   memory_order_release, memory_order_acq_rel, memory_order_seq_cst
      6 };
      7 
      8 int *test_c11_atomic_fetch_add_int_ptr(_Atomic(int *) *p) {
      9   // CHECK: test_c11_atomic_fetch_add_int_ptr
     10   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 12, i32 5)
     11   return __c11_atomic_fetch_add(p, 3, memory_order_seq_cst);
     12 }
     13 
     14 int *test_c11_atomic_fetch_sub_int_ptr(_Atomic(int *) *p) {
     15   // CHECK: test_c11_atomic_fetch_sub_int_ptr
     16   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 20, i32 5)
     17   return __c11_atomic_fetch_sub(p, 5, memory_order_seq_cst);
     18 }
     19 
     20 int test_c11_atomic_fetch_add_int(_Atomic(int) *p) {
     21   // CHECK: test_c11_atomic_fetch_add_int
     22   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 3, i32 5)
     23   return __c11_atomic_fetch_add(p, 3, memory_order_seq_cst);
     24 }
     25 
     26 int test_c11_atomic_fetch_sub_int(_Atomic(int) *p) {
     27   // CHECK: test_c11_atomic_fetch_sub_int
     28   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 5, i32 5)
     29   return __c11_atomic_fetch_sub(p, 5, memory_order_seq_cst);
     30 }
     31 
     32 int *fp2a(int **p) {
     33   // CHECK: @fp2a
     34   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 4, i32 0)
     35   // Note, the GNU builtins do not multiply by sizeof(T)!
     36   return __atomic_fetch_sub(p, 4, memory_order_relaxed);
     37 }
     38 
     39 int test_atomic_fetch_add(int *p) {
     40   // CHECK: test_atomic_fetch_add
     41   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     42   return __atomic_fetch_add(p, 55, memory_order_seq_cst);
     43 }
     44 
     45 int test_atomic_fetch_sub(int *p) {
     46   // CHECK: test_atomic_fetch_sub
     47   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     48   return __atomic_fetch_sub(p, 55, memory_order_seq_cst);
     49 }
     50 
     51 int test_atomic_fetch_and(int *p) {
     52   // CHECK: test_atomic_fetch_and
     53   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_and_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     54   return __atomic_fetch_and(p, 55, memory_order_seq_cst);
     55 }
     56 
     57 int test_atomic_fetch_or(int *p) {
     58   // CHECK: test_atomic_fetch_or
     59   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_or_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     60   return __atomic_fetch_or(p, 55, memory_order_seq_cst);
     61 }
     62 
     63 int test_atomic_fetch_xor(int *p) {
     64   // CHECK: test_atomic_fetch_xor
     65   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_xor_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     66   return __atomic_fetch_xor(p, 55, memory_order_seq_cst);
     67 }
     68 
     69 int test_atomic_fetch_nand(int *p) {
     70   // CHECK: test_atomic_fetch_nand
     71   // CHECK: {{%[^ ]*}} = tail call i32 @__atomic_fetch_nand_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     72   return __atomic_fetch_nand(p, 55, memory_order_seq_cst);
     73 }
     74 
     75 int test_atomic_add_fetch(int *p) {
     76   // CHECK: test_atomic_add_fetch
     77   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_add_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     78   // CHECK: {{%[^ ]*}} = add i32 [[CALL]], 55
     79   return __atomic_add_fetch(p, 55, memory_order_seq_cst);
     80 }
     81 
     82 int test_atomic_sub_fetch(int *p) {
     83   // CHECK: test_atomic_sub_fetch
     84   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_sub_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     85   // CHECK: {{%[^ ]*}} = add i32 [[CALL]], -55
     86   return __atomic_sub_fetch(p, 55, memory_order_seq_cst);
     87 }
     88 
     89 int test_atomic_and_fetch(int *p) {
     90   // CHECK: test_atomic_and_fetch
     91   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_and_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     92   // CHECK: {{%[^ ]*}} = and i32 [[CALL]], 55
     93   return __atomic_and_fetch(p, 55, memory_order_seq_cst);
     94 }
     95 
     96 int test_atomic_or_fetch(int *p) {
     97   // CHECK: test_atomic_or_fetch
     98   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_or_4(i8* {{%[0-9]+}}, i32 55, i32 5)
     99   // CHECK: {{%[^ ]*}} = or i32 [[CALL]], 55
    100   return __atomic_or_fetch(p, 55, memory_order_seq_cst);
    101 }
    102 
    103 int test_atomic_xor_fetch(int *p) {
    104   // CHECK: test_atomic_xor_fetch
    105   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_xor_4(i8* {{%[0-9]+}}, i32 55, i32 5)
    106   // CHECK: {{%[^ ]*}} = xor i32 [[CALL]], 55
    107   return __atomic_xor_fetch(p, 55, memory_order_seq_cst);
    108 }
    109 
    110 int test_atomic_nand_fetch(int *p) {
    111   // CHECK: test_atomic_nand_fetch
    112   // CHECK: [[CALL:%[^ ]*]] = tail call i32 @__atomic_fetch_nand_4(i8* {{%[0-9]+}}, i32 55, i32 5)
    113   // CHECK: [[OR:%[^ ]*]] = or i32 [[CALL]], -56
    114   // CHECK: {{%[^ ]*}} = xor i32 [[OR]], 55
    115   return __atomic_nand_fetch(p, 55, memory_order_seq_cst);
    116 }
    117