/external/skia/include/private/ |
SkOncePtr.h | 70 &fState, &state, (uintptr_t)1, sk_memory_order_relaxed, sk_memory_order_relaxed)) {
|
SkAtomics.h | 18 sk_memory_order_relaxed, enumerator in enum:sk_memory_order 96 SkASSERT(mo == sk_memory_order_relaxed || 106 SkASSERT(mo == sk_memory_order_relaxed || 132 SkASSERT(failure == sk_memory_order_relaxed ||
|
SkOnce.h | 75 if (!sk_atomic_load(done, sk_memory_order_relaxed)) {
|
/external/skia/src/core/ |
SkTaskGroup.cpp | 59 SkASSERT(pending->load(sk_memory_order_relaxed) == 0); 127 pending->fetch_add(+1, sk_memory_order_relaxed); // No barrier needed. 136 pending->fetch_add(+N, sk_memory_order_relaxed); // No barrier needed.
|
SkPicture.cpp | 34 uint32_t id = sk_atomic_load(&fUniqueID, sk_memory_order_relaxed); 43 uint32_t id = sk_atomic_load(&fUniqueID, sk_memory_order_relaxed); 47 sk_memory_order_relaxed, 48 sk_memory_order_relaxed)) {
|
SkSharedMutex.cpp | 276 int32_t oldQueueCounts = fQueueCounts.load(sk_memory_order_relaxed); 302 sk_memory_order_release, sk_memory_order_relaxed)); 314 int32_t oldQueueCounts = fQueueCounts.load(sk_memory_order_relaxed); 325 sk_memory_order_acquire, sk_memory_order_relaxed));
|
SkTraceEvent.h | 109 #define TRACE_EVENT_API_ATOMIC_LOAD(var) sk_atomic_load(&var, sk_memory_order_relaxed) 111 sk_atomic_store(&var, value, sk_memory_order_relaxed)
|
/external/skia/include/core/ |
SkRefCnt.h | 70 (void)sk_atomic_fetch_add(&fRefCnt, +1, sk_memory_order_relaxed); // No barrier required. 241 void ref() const { (void)sk_atomic_fetch_add(&fRefCnt, +1, sk_memory_order_relaxed); }
|
SkPath.h | [all...] |
/external/skia/src/gpu/vk/ |
GrVkResource.h | 88 (void)sk_atomic_fetch_add(&fRefCnt, +1, sk_memory_order_relaxed); // No barrier required.
|