Home | History | Annotate | Download | only in AArch64
      1 ; RUN: llc < %s -march=arm64 -asm-verbose=false -verify-machineinstrs -mcpu=cyclone | FileCheck %s
      2 
      3 define i32 @val_compare_and_swap(i32* %p, i32 %cmp, i32 %new) #0 {
      4 ; CHECK-LABEL: val_compare_and_swap:
      5 ; CHECK-NEXT: mov    x[[ADDR:[0-9]+]], x0
      6 ; CHECK-NEXT: [[TRYBB:.?LBB[0-9_]+]]:
      7 ; CHECK-NEXT: ldaxr  [[RESULT:w[0-9]+]], [x[[ADDR]]]
      8 ; CHECK-NEXT: cmp    [[RESULT]], w1
      9 ; CHECK-NEXT: b.ne   [[FAILBB:.?LBB[0-9_]+]]
     10 ; CHECK-NEXT: stxr   [[SCRATCH_REG:w[0-9]+]], w2, [x[[ADDR]]]
     11 ; CHECK-NEXT: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     12 ; CHECK-NEXT: b      [[EXITBB:.?LBB[0-9_]+]]
     13 ; CHECK-NEXT: [[FAILBB]]:
     14 ; CHECK-NEXT: clrex
     15 ; CHECK-NEXT: [[EXITBB]]:
     16   %pair = cmpxchg i32* %p, i32 %cmp, i32 %new acquire acquire
     17   %val = extractvalue { i32, i1 } %pair, 0
     18   ret i32 %val
     19 }
     20 
     21 define i32 @val_compare_and_swap_from_load(i32* %p, i32 %cmp, i32* %pnew) #0 {
     22 ; CHECK-LABEL: val_compare_and_swap_from_load:
     23 ; CHECK-NEXT: ldr    [[NEW:w[0-9]+]], [x2]
     24 ; CHECK-NEXT: [[TRYBB:.?LBB[0-9_]+]]:
     25 ; CHECK-NEXT: ldaxr  [[RESULT:w[0-9]+]], [x0]
     26 ; CHECK-NEXT: cmp    [[RESULT]], w1
     27 ; CHECK-NEXT: b.ne   [[FAILBB:.?LBB[0-9_]+]]
     28 ; CHECK-NEXT: stxr   [[SCRATCH_REG:w[0-9]+]], [[NEW]], [x0]
     29 ; CHECK-NEXT: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     30 ; CHECK-NEXT: b      [[EXITBB:.?LBB[0-9_]+]]
     31 ; CHECK-NEXT: [[FAILBB]]:
     32 ; CHECK-NEXT: clrex
     33 ; CHECK-NEXT: [[EXITBB]]:
     34   %new = load i32, i32* %pnew
     35   %pair = cmpxchg i32* %p, i32 %cmp, i32 %new acquire acquire
     36   %val = extractvalue { i32, i1 } %pair, 0
     37   ret i32 %val
     38 }
     39 
     40 define i32 @val_compare_and_swap_rel(i32* %p, i32 %cmp, i32 %new) #0 {
     41 ; CHECK-LABEL: val_compare_and_swap_rel:
     42 ; CHECK-NEXT: mov    x[[ADDR:[0-9]+]], x0
     43 ; CHECK-NEXT: [[TRYBB:.?LBB[0-9_]+]]:
     44 ; CHECK-NEXT: ldaxr  [[RESULT:w[0-9]+]], [x[[ADDR]]
     45 ; CHECK-NEXT: cmp    [[RESULT]], w1
     46 ; CHECK-NEXT: b.ne   [[FAILBB:.?LBB[0-9_]+]]
     47 ; CHECK-NEXT: stlxr  [[SCRATCH_REG:w[0-9]+]], w2, [x[[ADDR]]
     48 ; CHECK-NEXT: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     49 ; CHECK-NEXT: b      [[EXITBB:.?LBB[0-9_]+]]
     50 ; CHECK-NEXT: [[FAILBB]]:
     51 ; CHECK-NEXT: clrex
     52 ; CHECK-NEXT: [[EXITBB]]:
     53   %pair = cmpxchg i32* %p, i32 %cmp, i32 %new acq_rel monotonic
     54   %val = extractvalue { i32, i1 } %pair, 0
     55   ret i32 %val
     56 }
     57 
     58 define i64 @val_compare_and_swap_64(i64* %p, i64 %cmp, i64 %new) #0 {
     59 ; CHECK-LABEL: val_compare_and_swap_64:
     60 ; CHECK-NEXT: mov    x[[ADDR:[0-9]+]], x0
     61 ; CHECK-NEXT: [[TRYBB:.?LBB[0-9_]+]]:
     62 ; CHECK-NEXT: ldxr   [[RESULT:x[0-9]+]], [x[[ADDR]]]
     63 ; CHECK-NEXT: cmp    [[RESULT]], x1
     64 ; CHECK-NEXT: b.ne   [[FAILBB:.?LBB[0-9_]+]]
     65 ; CHECK-NEXT: stxr   [[SCRATCH_REG:w[0-9]+]], x2, [x[[ADDR]]]
     66 ; CHECK-NEXT: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     67 ; CHECK-NEXT: b      [[EXITBB:.?LBB[0-9_]+]]
     68 ; CHECK-NEXT: [[FAILBB]]:
     69 ; CHECK-NEXT: clrex
     70 ; CHECK-NEXT: [[EXITBB]]:
     71   %pair = cmpxchg i64* %p, i64 %cmp, i64 %new monotonic monotonic
     72   %val = extractvalue { i64, i1 } %pair, 0
     73   ret i64 %val
     74 }
     75 
     76 define i32 @fetch_and_nand(i32* %p) #0 {
     77 ; CHECK-LABEL: fetch_and_nand:
     78 ; CHECK: [[TRYBB:.?LBB[0-9_]+]]:
     79 ; CHECK: ldxr   w[[DEST_REG:[0-9]+]], [x0]
     80 ; CHECK: mvn    [[TMP_REG:w[0-9]+]], w[[DEST_REG]]
     81 ; CHECK: orr    [[SCRATCH2_REG:w[0-9]+]], [[TMP_REG]], #0xfffffff8
     82 ; CHECK-NOT: stlxr [[SCRATCH2_REG]], [[SCRATCH2_REG]]
     83 ; CHECK: stlxr   [[SCRATCH_REG:w[0-9]+]], [[SCRATCH2_REG]], [x0]
     84 ; CHECK: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     85 ; CHECK: mov    x0, x[[DEST_REG]]
     86   %val = atomicrmw nand i32* %p, i32 7 release
     87   ret i32 %val
     88 }
     89 
     90 define i64 @fetch_and_nand_64(i64* %p) #0 {
     91 ; CHECK-LABEL: fetch_and_nand_64:
     92 ; CHECK: mov    x[[ADDR:[0-9]+]], x0
     93 ; CHECK: [[TRYBB:.?LBB[0-9_]+]]:
     94 ; CHECK: ldaxr   x[[DEST_REG:[0-9]+]], [x[[ADDR]]]
     95 ; CHECK: mvn    w[[TMP_REG:[0-9]+]], w[[DEST_REG]]
     96 ; CHECK: orr    [[SCRATCH2_REG:x[0-9]+]], x[[TMP_REG]], #0xfffffffffffffff8
     97 ; CHECK: stlxr   [[SCRATCH_REG:w[0-9]+]], [[SCRATCH2_REG]], [x[[ADDR]]]
     98 ; CHECK: cbnz   [[SCRATCH_REG]], [[TRYBB]]
     99 
    100   %val = atomicrmw nand i64* %p, i64 7 acq_rel
    101   ret i64 %val
    102 }
    103 
    104 define i32 @fetch_and_or(i32* %p) #0 {
    105 ; CHECK-LABEL: fetch_and_or:
    106 ; CHECK: movz   [[OLDVAL_REG:w[0-9]+]], #0x5
    107 ; CHECK: [[TRYBB:.?LBB[0-9_]+]]:
    108 ; CHECK: ldaxr   w[[DEST_REG:[0-9]+]], [x0]
    109 ; CHECK: orr    [[SCRATCH2_REG:w[0-9]+]], w[[DEST_REG]], [[OLDVAL_REG]]
    110 ; CHECK-NOT: stlxr [[SCRATCH2_REG]], [[SCRATCH2_REG]]
    111 ; CHECK: stlxr [[SCRATCH_REG:w[0-9]+]], [[SCRATCH2_REG]], [x0]
    112 ; CHECK: cbnz   [[SCRATCH_REG]], [[TRYBB]]
    113 ; CHECK: mov    x0, x[[DEST_REG]]
    114   %val = atomicrmw or i32* %p, i32 5 seq_cst
    115   ret i32 %val
    116 }
    117 
    118 define i64 @fetch_and_or_64(i64* %p) #0 {
    119 ; CHECK: fetch_and_or_64:
    120 ; CHECK: mov    x[[ADDR:[0-9]+]], x0
    121 ; CHECK: [[TRYBB:.?LBB[0-9_]+]]:
    122 ; CHECK: ldxr   [[DEST_REG:x[0-9]+]], [x[[ADDR]]]
    123 ; CHECK: orr    [[SCRATCH2_REG:x[0-9]+]], [[DEST_REG]], #0x7
    124 ; CHECK: stxr   [[SCRATCH_REG:w[0-9]+]], [[SCRATCH2_REG]], [x[[ADDR]]]
    125 ; CHECK: cbnz   [[SCRATCH_REG]], [[TRYBB]]
    126   %val = atomicrmw or i64* %p, i64 7 monotonic
    127   ret i64 %val
    128 }
    129 
    130 define void @acquire_fence() #0 {
    131    fence acquire
    132    ret void
    133    ; CHECK-LABEL: acquire_fence:
    134    ; CHECK: dmb ishld
    135 }
    136 
    137 define void @release_fence() #0 {
    138    fence release
    139    ret void
    140    ; CHECK-LABEL: release_fence:
    141    ; CHECK: dmb ish{{$}}
    142 }
    143 
    144 define void @seq_cst_fence() #0 {
    145    fence seq_cst
    146    ret void
    147    ; CHECK-LABEL: seq_cst_fence:
    148    ; CHECK: dmb ish{{$}}
    149 }
    150 
    151 define i32 @atomic_load(i32* %p) #0 {
    152    %r = load atomic i32, i32* %p seq_cst, align 4
    153    ret i32 %r
    154    ; CHECK-LABEL: atomic_load:
    155    ; CHECK: ldar
    156 }
    157 
    158 define i8 @atomic_load_relaxed_8(i8* %p, i32 %off32) #0 {
    159 ; CHECK-LABEL: atomic_load_relaxed_8:
    160   %ptr_unsigned = getelementptr i8, i8* %p, i32 4095
    161   %val_unsigned = load atomic i8, i8* %ptr_unsigned monotonic, align 1
    162 ; CHECK: ldrb {{w[0-9]+}}, [x0, #4095]
    163 
    164   %ptr_regoff = getelementptr i8, i8* %p, i32 %off32
    165   %val_regoff = load atomic i8, i8* %ptr_regoff unordered, align 1
    166   %tot1 = add i8 %val_unsigned, %val_regoff
    167 ; CHECK: ldrb {{w[0-9]+}}, [x0, w1, sxtw]
    168 
    169   %ptr_unscaled = getelementptr i8, i8* %p, i32 -256
    170   %val_unscaled = load atomic i8, i8* %ptr_unscaled monotonic, align 1
    171   %tot2 = add i8 %tot1, %val_unscaled
    172 ; CHECK: ldurb {{w[0-9]+}}, [x0, #-256]
    173 
    174   %ptr_random = getelementptr i8, i8* %p, i32 1191936 ; 0x123000 (i.e. ADD imm)
    175   %val_random = load atomic i8, i8* %ptr_random unordered, align 1
    176   %tot3 = add i8 %tot2, %val_random
    177 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    178 ; CHECK: ldrb {{w[0-9]+}}, [x[[ADDR]]]
    179 
    180   ret i8 %tot3
    181 }
    182 
    183 define i16 @atomic_load_relaxed_16(i16* %p, i32 %off32) #0 {
    184 ; CHECK-LABEL: atomic_load_relaxed_16:
    185   %ptr_unsigned = getelementptr i16, i16* %p, i32 4095
    186   %val_unsigned = load atomic i16, i16* %ptr_unsigned monotonic, align 2
    187 ; CHECK: ldrh {{w[0-9]+}}, [x0, #8190]
    188 
    189   %ptr_regoff = getelementptr i16, i16* %p, i32 %off32
    190   %val_regoff = load atomic i16, i16* %ptr_regoff unordered, align 2
    191   %tot1 = add i16 %val_unsigned, %val_regoff
    192 ; CHECK: ldrh {{w[0-9]+}}, [x0, w1, sxtw #1]
    193 
    194   %ptr_unscaled = getelementptr i16, i16* %p, i32 -128
    195   %val_unscaled = load atomic i16, i16* %ptr_unscaled monotonic, align 2
    196   %tot2 = add i16 %tot1, %val_unscaled
    197 ; CHECK: ldurh {{w[0-9]+}}, [x0, #-256]
    198 
    199   %ptr_random = getelementptr i16, i16* %p, i32 595968 ; 0x123000/2 (i.e. ADD imm)
    200   %val_random = load atomic i16, i16* %ptr_random unordered, align 2
    201   %tot3 = add i16 %tot2, %val_random
    202 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    203 ; CHECK: ldrh {{w[0-9]+}}, [x[[ADDR]]]
    204 
    205   ret i16 %tot3
    206 }
    207 
    208 define i32 @atomic_load_relaxed_32(i32* %p, i32 %off32) #0 {
    209 ; CHECK-LABEL: atomic_load_relaxed_32:
    210   %ptr_unsigned = getelementptr i32, i32* %p, i32 4095
    211   %val_unsigned = load atomic i32, i32* %ptr_unsigned monotonic, align 4
    212 ; CHECK: ldr {{w[0-9]+}}, [x0, #16380]
    213 
    214   %ptr_regoff = getelementptr i32, i32* %p, i32 %off32
    215   %val_regoff = load atomic i32, i32* %ptr_regoff unordered, align 4
    216   %tot1 = add i32 %val_unsigned, %val_regoff
    217 ; CHECK: ldr {{w[0-9]+}}, [x0, w1, sxtw #2]
    218 
    219   %ptr_unscaled = getelementptr i32, i32* %p, i32 -64
    220   %val_unscaled = load atomic i32, i32* %ptr_unscaled monotonic, align 4
    221   %tot2 = add i32 %tot1, %val_unscaled
    222 ; CHECK: ldur {{w[0-9]+}}, [x0, #-256]
    223 
    224   %ptr_random = getelementptr i32, i32* %p, i32 297984 ; 0x123000/4 (i.e. ADD imm)
    225   %val_random = load atomic i32, i32* %ptr_random unordered, align 4
    226   %tot3 = add i32 %tot2, %val_random
    227 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    228 ; CHECK: ldr {{w[0-9]+}}, [x[[ADDR]]]
    229 
    230   ret i32 %tot3
    231 }
    232 
    233 define i64 @atomic_load_relaxed_64(i64* %p, i32 %off32) #0 {
    234 ; CHECK-LABEL: atomic_load_relaxed_64:
    235   %ptr_unsigned = getelementptr i64, i64* %p, i32 4095
    236   %val_unsigned = load atomic i64, i64* %ptr_unsigned monotonic, align 8
    237 ; CHECK: ldr {{x[0-9]+}}, [x0, #32760]
    238 
    239   %ptr_regoff = getelementptr i64, i64* %p, i32 %off32
    240   %val_regoff = load atomic i64, i64* %ptr_regoff unordered, align 8
    241   %tot1 = add i64 %val_unsigned, %val_regoff
    242 ; CHECK: ldr {{x[0-9]+}}, [x0, w1, sxtw #3]
    243 
    244   %ptr_unscaled = getelementptr i64, i64* %p, i32 -32
    245   %val_unscaled = load atomic i64, i64* %ptr_unscaled monotonic, align 8
    246   %tot2 = add i64 %tot1, %val_unscaled
    247 ; CHECK: ldur {{x[0-9]+}}, [x0, #-256]
    248 
    249   %ptr_random = getelementptr i64, i64* %p, i32 148992 ; 0x123000/8 (i.e. ADD imm)
    250   %val_random = load atomic i64, i64* %ptr_random unordered, align 8
    251   %tot3 = add i64 %tot2, %val_random
    252 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    253 ; CHECK: ldr {{x[0-9]+}}, [x[[ADDR]]]
    254 
    255   ret i64 %tot3
    256 }
    257 
    258 
    259 define void @atomc_store(i32* %p) #0 {
    260    store atomic i32 4, i32* %p seq_cst, align 4
    261    ret void
    262    ; CHECK-LABEL: atomc_store:
    263    ; CHECK: stlr
    264 }
    265 
    266 define void @atomic_store_relaxed_8(i8* %p, i32 %off32, i8 %val) #0 {
    267 ; CHECK-LABEL: atomic_store_relaxed_8:
    268   %ptr_unsigned = getelementptr i8, i8* %p, i32 4095
    269   store atomic i8 %val, i8* %ptr_unsigned monotonic, align 1
    270 ; CHECK: strb {{w[0-9]+}}, [x0, #4095]
    271 
    272   %ptr_regoff = getelementptr i8, i8* %p, i32 %off32
    273   store atomic i8 %val, i8* %ptr_regoff unordered, align 1
    274 ; CHECK: strb {{w[0-9]+}}, [x0, w1, sxtw]
    275 
    276   %ptr_unscaled = getelementptr i8, i8* %p, i32 -256
    277   store atomic i8 %val, i8* %ptr_unscaled monotonic, align 1
    278 ; CHECK: sturb {{w[0-9]+}}, [x0, #-256]
    279 
    280   %ptr_random = getelementptr i8, i8* %p, i32 1191936 ; 0x123000 (i.e. ADD imm)
    281   store atomic i8 %val, i8* %ptr_random unordered, align 1
    282 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    283 ; CHECK: strb {{w[0-9]+}}, [x[[ADDR]]]
    284 
    285   ret void
    286 }
    287 
    288 define void @atomic_store_relaxed_16(i16* %p, i32 %off32, i16 %val) #0 {
    289 ; CHECK-LABEL: atomic_store_relaxed_16:
    290   %ptr_unsigned = getelementptr i16, i16* %p, i32 4095
    291   store atomic i16 %val, i16* %ptr_unsigned monotonic, align 2
    292 ; CHECK: strh {{w[0-9]+}}, [x0, #8190]
    293 
    294   %ptr_regoff = getelementptr i16, i16* %p, i32 %off32
    295   store atomic i16 %val, i16* %ptr_regoff unordered, align 2
    296 ; CHECK: strh {{w[0-9]+}}, [x0, w1, sxtw #1]
    297 
    298   %ptr_unscaled = getelementptr i16, i16* %p, i32 -128
    299   store atomic i16 %val, i16* %ptr_unscaled monotonic, align 2
    300 ; CHECK: sturh {{w[0-9]+}}, [x0, #-256]
    301 
    302   %ptr_random = getelementptr i16, i16* %p, i32 595968 ; 0x123000/2 (i.e. ADD imm)
    303   store atomic i16 %val, i16* %ptr_random unordered, align 2
    304 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    305 ; CHECK: strh {{w[0-9]+}}, [x[[ADDR]]]
    306 
    307   ret void
    308 }
    309 
    310 define void @atomic_store_relaxed_32(i32* %p, i32 %off32, i32 %val) #0 {
    311 ; CHECK-LABEL: atomic_store_relaxed_32:
    312   %ptr_unsigned = getelementptr i32, i32* %p, i32 4095
    313   store atomic i32 %val, i32* %ptr_unsigned monotonic, align 4
    314 ; CHECK: str {{w[0-9]+}}, [x0, #16380]
    315 
    316   %ptr_regoff = getelementptr i32, i32* %p, i32 %off32
    317   store atomic i32 %val, i32* %ptr_regoff unordered, align 4
    318 ; CHECK: str {{w[0-9]+}}, [x0, w1, sxtw #2]
    319 
    320   %ptr_unscaled = getelementptr i32, i32* %p, i32 -64
    321   store atomic i32 %val, i32* %ptr_unscaled monotonic, align 4
    322 ; CHECK: stur {{w[0-9]+}}, [x0, #-256]
    323 
    324   %ptr_random = getelementptr i32, i32* %p, i32 297984 ; 0x123000/4 (i.e. ADD imm)
    325   store atomic i32 %val, i32* %ptr_random unordered, align 4
    326 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    327 ; CHECK: str {{w[0-9]+}}, [x[[ADDR]]]
    328 
    329   ret void
    330 }
    331 
    332 define void @atomic_store_relaxed_64(i64* %p, i32 %off32, i64 %val) #0 {
    333 ; CHECK-LABEL: atomic_store_relaxed_64:
    334   %ptr_unsigned = getelementptr i64, i64* %p, i32 4095
    335   store atomic i64 %val, i64* %ptr_unsigned monotonic, align 8
    336 ; CHECK: str {{x[0-9]+}}, [x0, #32760]
    337 
    338   %ptr_regoff = getelementptr i64, i64* %p, i32 %off32
    339   store atomic i64 %val, i64* %ptr_regoff unordered, align 8
    340 ; CHECK: str {{x[0-9]+}}, [x0, w1, sxtw #3]
    341 
    342   %ptr_unscaled = getelementptr i64, i64* %p, i32 -32
    343   store atomic i64 %val, i64* %ptr_unscaled monotonic, align 8
    344 ; CHECK: stur {{x[0-9]+}}, [x0, #-256]
    345 
    346   %ptr_random = getelementptr i64, i64* %p, i32 148992 ; 0x123000/8 (i.e. ADD imm)
    347   store atomic i64 %val, i64* %ptr_random unordered, align 8
    348 ; CHECK: add x[[ADDR:[0-9]+]], x0, #291, lsl #12
    349 ; CHECK: str {{x[0-9]+}}, [x[[ADDR]]]
    350 
    351   ret void
    352 }
    353 
    354 ; rdar://11531169
    355 ; rdar://11531308
    356 
    357 %"class.X::Atomic" = type { %struct.x_atomic_t }
    358 %struct.x_atomic_t = type { i32 }
    359 
    360 @counter = external hidden global %"class.X::Atomic", align 4
    361 
    362 define i32 @next_id() nounwind optsize ssp align 2 {
    363 entry:
    364   %0 = atomicrmw add i32* getelementptr inbounds (%"class.X::Atomic", %"class.X::Atomic"* @counter, i64 0, i32 0, i32 0), i32 1 seq_cst
    365   %add.i = add i32 %0, 1
    366   %tobool = icmp eq i32 %add.i, 0
    367   br i1 %tobool, label %if.else, label %return
    368 
    369 if.else:                                          ; preds = %entry
    370   %1 = atomicrmw add i32* getelementptr inbounds (%"class.X::Atomic", %"class.X::Atomic"* @counter, i64 0, i32 0, i32 0), i32 1 seq_cst
    371   %add.i2 = add i32 %1, 1
    372   br label %return
    373 
    374 return:                                           ; preds = %if.else, %entry
    375   %retval.0 = phi i32 [ %add.i2, %if.else ], [ %add.i, %entry ]
    376   ret i32 %retval.0
    377 }
    378 
    379 attributes #0 = { nounwind }
    380