/ndk/sources/cxx-stl/llvm-libc++/test/utilities/memory/util.smartptr/util.smartptr.shared.atomic/ |
atomic_load.pass.cpp | 16 // atomic_load(const shared_ptr<T>* p) 26 std::shared_ptr<int> q = std::atomic_load(&p);
|
/ndk/sources/cxx-stl/llvm-libc++/test/atomics/atomics.types.operations/atomics.types.operations.req/ |
atomic_load.pass.cpp | 14 // atomic_load(const volatile atomic<T>* obj); 18 // atomic_load(const atomic<T>* obj); 31 assert(std::atomic_load(&t) == T(1)); 34 assert(std::atomic_load(&vt) == T(2));
|
/external/compiler-rt/lib/sanitizer_common/ |
sanitizer_lfstack.h | 33 return (atomic_load(&head_, memory_order_relaxed) & kPtrMask) == 0; 37 u64 cmp = atomic_load(&head_, memory_order_relaxed); 49 u64 cmp = atomic_load(&head_, memory_order_acquire);
|
sanitizer_stackdepot.cc | 74 uptr cmp = atomic_load(&depot.region_pos, memory_order_acquire); 75 uptr end = atomic_load(&depot.region_end, memory_order_acquire); 127 uptr cmp = atomic_load(p, memory_order_relaxed); 149 uptr v = atomic_load(p, memory_order_consume); 191 uptr v = atomic_load(p, memory_order_consume);
|
sanitizer_mutex.h | 52 if (atomic_load(&state_, memory_order_relaxed) == 0
|
sanitizer_atomic_clang.h | 38 INLINE typename T::Type atomic_load( function in namespace:__sanitizer
|
sanitizer_atomic_msvc.h | 69 INLINE typename T::Type atomic_load( function in namespace:__sanitizer
|
sanitizer_quarantine.h | 122 return atomic_load(&size_, memory_order_relaxed);
|
sanitizer_linux.cc | 570 CHECK_NE(MtxUnlocked, atomic_load(m, memory_order_relaxed));
|
sanitizer_allocator.h | 215 v += atomic_load(&stats_[i], memory_order_relaxed); 224 return atomic_load(&stats_[i], memory_order_relaxed); [all...] |
/external/compiler-rt/lib/tsan/rtl/ |
tsan_mutex.cc | 208 CHECK_EQ(atomic_load(&state_, memory_order_relaxed), kUnlocked); 220 if (atomic_load(&state_, memory_order_relaxed) == kUnlocked) { 250 prev = atomic_load(&state_, memory_order_acquire); 271 CHECK_NE(atomic_load(&state_, memory_order_relaxed), 0);
|
tsan_fd.cc | 52 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) 58 if (s && atomic_load(&s->rc, memory_order_relaxed) != (u64)-1) { 74 uptr l1 = atomic_load(pl1, memory_order_consume); 122 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed); 134 FdDesc *tab = (FdDesc*)atomic_load(&fdctx.tab[l1], memory_order_relaxed);
|
tsan_mman.cc | 36 uptr v = atomic_load(a, memory_order_relaxed); 43 v = atomic_load(a, memory_order_relaxed); 53 uptr v = atomic_load(a, memory_order_relaxed);
|
tsan_interceptors.cc | 653 u32 cmp = atomic_load(g, memory_order_acquire); 717 while ((tid = atomic_load(&p->tid, memory_order_acquire)) == 0) 758 while (atomic_load(&p.tid, memory_order_acquire) != 0) 1042 u32 v = atomic_load(a, memory_order_acquire); 1055 v = atomic_load(a, memory_order_acquire); [all...] |
tsan_rtl.cc | 323 u64 raw = atomic_load((atomic_uint64_t*)p, memory_order_relaxed);
|
/prebuilts/gcc/linux-x86/host/i686-linux-glibc2.7-4.4.3/i686-linux/include/c++/4.4.3/bits/ |
atomicfwd_c.h | 125 #define atomic_load(__a) \ macro
|
/prebuilts/ndk/5/sources/cxx-stl/gnu-libstdc++/include/bits/ |
atomicfwd_c.h | 125 #define atomic_load(__a) \ macro
|
/prebuilts/ndk/6/sources/cxx-stl/gnu-libstdc++/include/bits/ |
atomicfwd_c.h | 125 #define atomic_load(__a) \ macro
|
/prebuilts/ndk/7/sources/cxx-stl/gnu-libstdc++/include/bits/ |
atomicfwd_c.h | 125 #define atomic_load(__a) \ macro
|
/prebuilts/ndk/8/sources/cxx-stl/gnu-libstdc++/4.4.3/include/bits/ |
atomicfwd_c.h | 125 #define atomic_load(__a) \ macro
|
/external/compiler-rt/lib/tsan/tests/rtl/ |
tsan_mutex.cc | 169 int *val = (int *)atomic_load(singleton, memory_order_acquire); 215 uptr v = atomic_load(&flag, memory_order_acquire);
|
tsan_test_util_linux.cc | 323 Event* ev = (Event*)atomic_load(&impl->event, memory_order_acquire); 343 CHECK_EQ(atomic_load(&event, memory_order_relaxed), 0); 345 while (atomic_load(&event, memory_order_acquire) != 0)
|
/external/compiler-rt/lib/asan/ |
asan_allocator.cc | 332 int n = atomic_load(&n_page_groups_, memory_order_relaxed);
|