/external/compiler-rt/test/tsan/ |
atomic_free3.cc | 15 while (__atomic_load_n(a, __ATOMIC_RELAXED) == 0)
|
bench_acquire_only.cc | 13 __atomic_load_n(&x, __ATOMIC_ACQUIRE);
|
stack_sync_reuse.cc | 25 while (__atomic_load_n(&syncp, __ATOMIC_ACQUIRE) == 0) 39 while (__atomic_load_n(&syncp, __ATOMIC_RELAXED) != 0) 52 sink = __atomic_load_n(&s, __ATOMIC_ACQUIRE);
|
bench_single_writer.cc | 17 __atomic_load_n(&x, __ATOMIC_ACQUIRE);
|
bench_release_only.cc | 13 __atomic_load_n(&x[tid * kStride], __ATOMIC_ACQUIRE);
|
ignore_lib_lib.h | 11 while ((p = __atomic_load_n(&mem, __ATOMIC_ACQUIRE)) == 0)
|
signal_sync2.cc | 26 if (!__atomic_load_n(&process_signals, __ATOMIC_RELAXED)) 38 while (!__atomic_load_n(&done, __ATOMIC_RELAXED)) { 63 while (__atomic_load_n(&ready[pos], __ATOMIC_ACQUIRE) == 0) {
|
ignore_malloc.cc | 15 while ((p = __atomic_load_n(&g, __ATOMIC_RELAXED)) == 0)
|
atomic_norace.cc | 22 sink = __atomic_load_n(p, __ATOMIC_SEQ_CST); 27 sink = __atomic_load_n(p, __ATOMIC_SEQ_CST);
|
java_volatile.cc | 8 while (__atomic_load_n((int*)lockaddr, __ATOMIC_RELAXED) == 0)
|
atomic_race.cc | 22 sink = __atomic_load_n(p, __ATOMIC_SEQ_CST);
|
bench.h | 32 __atomic_load_n(&grow_clock_var, __ATOMIC_ACQUIRE);
|
signal_reset.cc | 22 while (__atomic_load_n(&stop, __ATOMIC_RELAXED) == 0) {
|
/external/protobuf/src/google/protobuf/stubs/ |
atomicops_internals_generic_gcc.h | 94 return __atomic_load_n(ptr, __ATOMIC_RELAXED); 98 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE); 102 return __atomic_load_n(ptr, __ATOMIC_SEQ_CST); 112 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE);
|
/external/libvpx/libvpx/vpx_util/ |
vpx_atomics.h | 33 #if (__has_builtin(__atomic_load_n)) || \ 94 return __atomic_load_n(&atomic->value, __ATOMIC_ACQUIRE);
|
/external/webrtc/webrtc/base/ |
atomicops.h | 63 return __atomic_load_n(i, __ATOMIC_ACQUIRE); 74 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE);
|
/external/v8/src/base/ |
atomicops_internals_portable.h | 110 return __atomic_load_n(ptr, __ATOMIC_RELAXED); 114 return __atomic_load_n(ptr, __ATOMIC_RELAXED); 118 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE); 168 return __atomic_load_n(ptr, __ATOMIC_RELAXED); 172 return __atomic_load_n(ptr, __ATOMIC_ACQUIRE);
|
/external/grpc-grpc/include/grpc/impl/codegen/ |
atm_gcc_atomic.h | 50 #define gpr_atm_acq_load(p) (__atomic_load_n((p), __ATOMIC_ACQUIRE)) 51 #define gpr_atm_no_barrier_load(p) (__atomic_load_n((p), __ATOMIC_RELAXED))
|
/external/libcxx/src/include/ |
atomic_support.h | 16 #if defined(__clang__) && __has_builtin(__atomic_load_n) \ 75 return __atomic_load_n(__val, __order);
|
/external/libcxxabi/src/include/ |
atomic_support.h | 20 #if defined(__clang__) && __has_builtin(__atomic_load_n) \ 79 return __atomic_load_n(__val, __order);
|
/external/compiler-rt/test/msan/ |
fork.cc | 34 if (__atomic_load_n(&done, __ATOMIC_RELAXED))
|
/external/libvpx/libvpx/third_party/libyuv/include/libyuv/ |
cpu_id.h | 63 int cpu_info = __atomic_load_n(&cpu_info_, __ATOMIC_RELAXED);
|
/external/syzkaller/executor/ |
common_fuchsia.h | 38 if (__atomic_load_n(&skip_segv, __ATOMIC_RELAXED)) { 144 while (!__atomic_load_n(&ev->state, __ATOMIC_ACQUIRE)) 150 return __atomic_load_n(&ev->state, __ATOMIC_ACQUIRE); 157 if (__atomic_load_n(&ev->state, __ATOMIC_RELAXED))
|
/external/syzkaller/tools/kcovtrace/ |
kcovtrace.c | 55 n = __atomic_load_n(&cover[0], __ATOMIC_RELAXED);
|
/external/harfbuzz_ng/src/ |
hb-atomic.hh | 61 #define hb_atomic_int_impl_get_relaxed(AI) __atomic_load_n ((AI), __ATOMIC_RELAXED) 62 #define hb_atomic_int_impl_get(AI) __atomic_load_n ((AI), __ATOMIC_ACQUIRE) 65 #define hb_atomic_ptr_impl_get_relaxed(P) __atomic_load_n ((P), __ATOMIC_RELAXED) 66 #define hb_atomic_ptr_impl_get(P) __atomic_load_n ((P), __ATOMIC_ACQUIRE)
|