/external/skia/src/core/ |
SkSpinlock.cpp | 12 while(sk_atomic_exchange(&fLocked, true, sk_memory_order_acquire)) { /*spin*/ }
|
SkSharedMutex.cpp | 263 sk_memory_order_acquire); 325 sk_memory_order_acquire, sk_memory_order_relaxed));
|
SkTaskGroup.cpp | 63 while (pending->load(sk_memory_order_acquire) > 0) {
|
/external/skia/src/utils/ |
SkEventTracer.cpp | 47 SkASSERT(nullptr == sk_atomic_load(&gUserTracer, sk_memory_order_acquire)); 50 atexit([]() { delete sk_atomic_load(&gUserTracer, sk_memory_order_acquire); }); 54 if (SkEventTracer* tracer = sk_atomic_load(&gUserTracer, sk_memory_order_acquire)) {
|
/external/skia/include/private/ |
SkSpinlock.h | 22 if (sk_atomic_exchange(&fLocked, true, sk_memory_order_acquire)) {
|
SkSemaphore.h | 37 if (sk_atomic_fetch_sub(&fCount, 1, sk_memory_order_acquire) <= 0) {
|
SkOncePtr.h | 64 uintptr_t state = sk_atomic_load(&fState, sk_memory_order_acquire); 85 state = sk_atomic_load(&fState, sk_memory_order_acquire); 95 auto state = sk_atomic_load(&fState, sk_memory_order_acquire);
|
SkAtomics.h | 20 sk_memory_order_acquire, enumerator in enum:sk_memory_order 98 mo == sk_memory_order_acquire || 134 failure == sk_memory_order_acquire || 182 T sk_acquire_load(T* ptr) { return sk_atomic_load(ptr, sk_memory_order_acquire); }
|
SkOnce.h | 113 if (!sk_atomic_load(done, sk_memory_order_acquire)) {
|
/external/skia/include/core/ |
SkRefCnt.h | 49 if (1 == sk_atomic_load(&fRefCnt, sk_memory_order_acquire)) { 240 bool unique() const { return 1 == sk_atomic_load(&fRefCnt, sk_memory_order_acquire); }
|
/external/skia/src/gpu/vk/ |
GrVkResource.h | 74 if (1 == sk_atomic_load(&fRefCnt, sk_memory_order_acquire)) {
|