HomeSort by relevance Sort by last modified time
    Searched full:atomic_load_explicit (Results 1 - 25 of 113) sorted by null

1 2 3 4 5

  /external/libcxx/test/std/atomics/atomics.types.operations/atomics.types.operations.req/
atomic_load_explicit.pass.cpp 17 // atomic_load_explicit(const volatile atomic<T>* obj, memory_order m);
21 // atomic_load_explicit(const atomic<T>* obj, memory_order m);
35 assert(std::atomic_load_explicit(&t, std::memory_order_seq_cst) == T(1));
38 assert(std::atomic_load_explicit(&vt, std::memory_order_seq_cst) == T(2));
  /prebuilts/ndk/r13/sources/cxx-stl/llvm-libc++/test/std/atomics/atomics.types.operations/atomics.types.operations.req/
atomic_load_explicit.pass.cpp 17 // atomic_load_explicit(const volatile atomic<T>* obj, memory_order m);
21 // atomic_load_explicit(const atomic<T>* obj, memory_order m);
35 assert(std::atomic_load_explicit(&t, std::memory_order_seq_cst) == T(1));
38 assert(std::atomic_load_explicit(&vt, std::memory_order_seq_cst) == T(2));
  /prebuilts/ndk/r11/sources/cxx-stl/llvm-libc++/libcxx/test/atomics/atomics.types.operations/atomics.types.operations.req/
atomic_load_explicit.pass.cpp 14 // atomic_load_explicit(const volatile atomic<T>* obj, memory_order m);
18 // atomic_load_explicit(const atomic<T>* obj, memory_order m);
31 assert(std::atomic_load_explicit(&t, std::memory_order_seq_cst) == T(1));
34 assert(std::atomic_load_explicit(&vt, std::memory_order_seq_cst) == T(2));
  /cts/apps/CtsVerifier/jni/audio_loopback/audio_utils/
atomic.c 24 return atomic_load_explicit(a, memory_order_acquire);
  /external/libcxx/test/std/utilities/memory/util.smartptr/util.smartptr.shared.atomic/
atomic_load_explicit.pass.cpp 23 // atomic_load_explicit(const shared_ptr<T>* p, memory_order mo)
36 std::shared_ptr<int> q = std::atomic_load_explicit(&p, std::memory_order_relaxed);
  /prebuilts/ndk/r11/sources/cxx-stl/llvm-libc++/libcxx/test/utilities/memory/util.smartptr/util.smartptr.shared.atomic/
atomic_load_explicit.pass.cpp 21 // atomic_load_explicit(const shared_ptr<T>* p, memory_order mo)
31 std::shared_ptr<int> q = std::atomic_load_explicit(&p, std::memory_order_relaxed);
  /prebuilts/ndk/r13/sources/cxx-stl/llvm-libc++/test/std/utilities/memory/util.smartptr/util.smartptr.shared.atomic/
atomic_load_explicit.pass.cpp 23 // atomic_load_explicit(const shared_ptr<T>* p, memory_order mo)
33 std::shared_ptr<int> q = std::atomic_load_explicit(&p, std::memory_order_relaxed);
  /bionic/libc/bionic/
pthread_key.cpp 80 uintptr_t seq = atomic_load_explicit(&key_map[i].seq, memory_order_relaxed);
90 atomic_load_explicit(&key_map[i].key_destructor, memory_order_relaxed));
95 if (atomic_load_explicit(&key_map[i].seq, memory_order_relaxed) != seq) {
120 uintptr_t seq = atomic_load_explicit(&key_map[i].seq, memory_order_relaxed);
142 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed);
156 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed);
174 uintptr_t seq = atomic_load_explicit(&key_map[key].seq, memory_order_relaxed);
pthread_barrier.cpp 120 while(atomic_load_explicit(&barrier->state, memory_order_acquire) == RELEASE) {
124 uint32_t prev_wait_count = atomic_load_explicit(&barrier->wait_count, memory_order_relaxed);
154 while (atomic_load_explicit(&barrier->state, memory_order_acquire) == WAIT) {
175 while (atomic_load_explicit(&barrier->state, memory_order_acquire) == RELEASE) {
178 if (atomic_load_explicit(&barrier->wait_count, memory_order_relaxed) != 0) {
pthread_rwlock.cpp 257 if (atomic_load_explicit(&rwlock->state, memory_order_relaxed) != 0) {
273 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed);
292 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) == __get_thread()->tid) {
306 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed);
349 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed);
365 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) == __get_thread()->tid) {
378 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed);
453 int old_state = atomic_load_explicit(&rwlock->state, memory_order_relaxed);
455 if (atomic_load_explicit(&rwlock->writer_tid, memory_order_relaxed) != __get_thread()->tid) {
pthread_once.cpp 54 int old_value = atomic_load_explicit(once_control_ptr, memory_order_acquire);
83 old_value = atomic_load_explicit(once_control_ptr, memory_order_acquire);
pthread_mutex.cpp 415 uint32_t owner_tid = atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed);
424 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed);
435 if (tid == atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed)) {
499 old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed);
515 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed);
539 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed);
551 if ( tid != atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed) ) {
583 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed);
597 if (tid == atomic_load_explicit(&mutex->owner_tid, memory_order_relaxed)) {
642 uint16_t old_state = atomic_load_explicit(&mutex->state, memory_order_relaxed)
    [all...]
pthread_cond.cpp 111 return COND_IS_SHARED(atomic_load_explicit(&state, memory_order_relaxed));
115 return COND_GET_CLOCK(atomic_load_explicit(&state, memory_order_relaxed)) == CLOCK_REALTIME;
180 unsigned int old_state = atomic_load_explicit(&cond->state, memory_order_relaxed);
semaphore.cpp 101 return (atomic_load_explicit(sem_count_ptr, memory_order_relaxed) & SEMCOUNT_SHARED_MASK);
145 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
165 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
189 unsigned int old_value = atomic_load_explicit(sem_count_ptr, memory_order_relaxed);
system_properties.cpp 350 uint_least32_t off = atomic_load_explicit(off_p, memory_order_consume);
355 uint_least32_t off = atomic_load_explicit(off_p, memory_order_consume);
386 uint_least32_t left_offset = atomic_load_explicit(&current->left, memory_order_relaxed);
402 uint_least32_t right_offset = atomic_load_explicit(&current->right, memory_order_relaxed);
438 uint_least32_t children_offset = atomic_load_explicit(&current->children, memory_order_relaxed);
463 uint_least32_t prop_offset = atomic_load_explicit(&current->prop, memory_order_relaxed);
658 uint_least32_t left_offset = atomic_load_explicit(&trie->left, memory_order_relaxed);
663 uint_least32_t prop_offset = atomic_load_explicit(&trie->prop, memory_order_relaxed);
669 uint_least32_t children_offset = atomic_load_explicit(&trie->children, memory_order_relaxed);
674 uint_least32_t right_offset = atomic_load_explicit(&trie->right, memory_order_relaxed)
    [all...]
__cxa_guard.cpp 82 int old_value = atomic_load_explicit(&gv->state, memory_order_acquire);
108 old_value = atomic_load_explicit(&gv->state, memory_order_acquire);
  /bionic/tests/
stdatomic_test.cpp 97 ASSERT_EQ(123, atomic_load_explicit(&i, memory_order_relaxed));
218 yval = atomic_load_explicit(&a->y, memory_order_acquire);
219 zval = atomic_load_explicit(&a->z, memory_order_relaxed);
220 xval = atomic_load_explicit(&a->x, memory_order_relaxed);
258 EXPECT_EQ(atomic_load_explicit(&a.x, memory_order_consume), BIG + 1);
259 EXPECT_EQ(atomic_load_explicit(&a.y, memory_order_seq_cst), BIG + 1);
  /external/libcxx/test/libcxx/atomics/
diagnose_invalid_memory_order.fail.cpp 37 std::atomic_load_explicit(&x, std::memory_order_release); // expected-warning {{memory order argument to atomic operation is invalid}}
38 std::atomic_load_explicit(&x, std::memory_order_acq_rel); // expected-warning {{memory order argument to atomic operation is invalid}}
39 std::atomic_load_explicit(&vx, std::memory_order_release); // expected-warning {{memory order argument to atomic operation is invalid}}
40 std::atomic_load_explicit(&vx, std::memory_order_acq_rel); // expected-warning {{memory order argument to atomic operation is invalid}}
42 std::atomic_load_explicit(&x, std::memory_order_relaxed);
43 std::atomic_load_explicit(&x, std::memory_order_consume);
44 std::atomic_load_explicit(&x, std::memory_order_acquire);
45 std::atomic_load_explicit(&x, std::memory_order_seq_cst);
  /external/drrickorang/LoopbackApp/app/src/main/jni/audio_utils/
atomic.c 24 return atomic_load_explicit(a, memory_order_acquire);
  /system/media/audio_utils/
fifo_index.cpp 27 return atomic_load_explicit(&mIndex, std::memory_order_acquire);
48 return atomic_load_explicit(&mIndex, std::memory_order_consume);
  /frameworks/native/libs/vr/libbroadcastring/include/libbroadcastring/
broadcast_ring.h 205 record_size = std::atomic_load_explicit(&ring.header_mmap()->record_size,
207 record_count = std::atomic_load_explicit(
270 return std::atomic_load_explicit(&header_mmap()->head,
279 return std::atomic_load_explicit(&header_mmap()->tail,
316 uint32_t tail = std::atomic_load_explicit(&header_mmap()->tail,
318 uint32_t head = std::atomic_load_explicit(&header_mmap()->head,
340 uint32_t final_head = std::atomic_load_explicit(
528 std::atomic_load_explicit(&in->data[i], std::memory_order_relaxed);
562 std::atomic_load_explicit(&header_mmap()->head,
564 std::atomic_load_explicit(&header_mmap()->tail
    [all...]
  /frameworks/av/media/libstagefright/include/media/stagefright/
MediaBuffer.h 197 return std::atomic_load_explicit(&mRemoteRefcount, std::memory_order_seq_cst);
206 return (std::atomic_load_explicit(
  /prebuilts/gcc/darwin-x86/aarch64/aarch64-linux-android-4.9/lib/gcc/aarch64-linux-android/4.9.x/include/
stdatomic.h 147 #define atomic_load_explicit(PTR, MO) \ macro
156 #define atomic_load(PTR) atomic_load_explicit (PTR, __ATOMIC_SEQ_CST)
  /prebuilts/gcc/darwin-x86/arm/arm-linux-androideabi-4.9/lib/gcc/arm-linux-androideabi/4.9.x/include/
stdatomic.h 147 #define atomic_load_explicit(PTR, MO) \ macro
156 #define atomic_load(PTR) atomic_load_explicit (PTR, __ATOMIC_SEQ_CST)
  /prebuilts/gcc/darwin-x86/mips/mips64el-linux-android-4.9/lib/gcc/mips64el-linux-android/4.9.x/include/
stdatomic.h 147 #define atomic_load_explicit(PTR, MO) \ macro
156 #define atomic_load(PTR) atomic_load_explicit (PTR, __ATOMIC_SEQ_CST)

Completed in 1063 milliseconds

1 2 3 4 5