/external/skia/src/gpu/ |
GrNonAtomicRef.h | 20 GrNonAtomicRef() : fRefCnt(1) {} 24 // fRefCnt can be one when a subclass is created statically 25 SkASSERT((0 == fRefCnt || 1 == fRefCnt)); 27 fRefCnt = -10; 31 bool unique() const { return 1 == fRefCnt; } 35 SkASSERT(fRefCnt > 0); 36 ++fRefCnt; 40 SkASSERT(fRefCnt > 0); 41 --fRefCnt; [all...] |
GrProgramElement.h | 43 // fRefCnt can be one when an effect is created statically using GR_CREATE_STATIC_EFFECT 44 SkASSERT((0 == fRefCnt || 1 == fRefCnt) && 0 == fPendingExecutions); 46 SkDEBUGCODE(fRefCnt = fPendingExecutions = -10;) 52 SkASSERT(fRefCnt > 0); 53 ++fRefCnt; 59 --fRefCnt; 60 if (0 == fRefCnt) { 74 SkASSERT(fRefCnt >= 0); 76 SkASSERT(fRefCnt + fPendingExecutions > 0) [all...] |
/external/skqp/src/gpu/ |
GrNonAtomicRef.h | 20 GrNonAtomicRef() : fRefCnt(1) {} 24 // fRefCnt can be one when a subclass is created statically 25 SkASSERT((0 == fRefCnt || 1 == fRefCnt)); 27 fRefCnt = -10; 31 bool unique() const { return 1 == fRefCnt; } 35 SkASSERT(fRefCnt > 0); 36 ++fRefCnt; 40 SkASSERT(fRefCnt > 0); 41 --fRefCnt; [all...] |
GrProgramElement.h | 43 // fRefCnt can be one when an effect is created statically using GR_CREATE_STATIC_EFFECT 44 SkASSERT((0 == fRefCnt || 1 == fRefCnt) && 0 == fPendingExecutions); 46 SkDEBUGCODE(fRefCnt = fPendingExecutions = -10;) 52 SkASSERT(fRefCnt > 0); 53 ++fRefCnt; 59 --fRefCnt; 60 if (0 == fRefCnt) { 74 SkASSERT(fRefCnt >= 0); 76 SkASSERT(fRefCnt + fPendingExecutions > 0) [all...] |
/external/skia/src/gpu/vk/ |
GrVkResource.h | 69 GrVkResource() : fRefCnt(1) { 80 SkASSERTF(fRefCnt == 1, "fRefCnt was %d", fRefCnt); 81 fRefCnt = 0; // illegal value, to catch us if we reuse after delete 87 int32_t getRefCnt() const { return fRefCnt; } 94 if (1 == sk_atomic_load(&fRefCnt, sk_memory_order_acquire)) { 107 SkASSERT(fRefCnt > 0); 108 (void)sk_atomic_fetch_add(&fRefCnt, +1, sk_memory_order_relaxed); // No barrier required. 117 SkASSERT(fRefCnt > 0) [all...] |
/external/skqp/src/gpu/vk/ |
GrVkResource.h | 69 GrVkResource() : fRefCnt(1) { 80 SkASSERTF(fRefCnt == 1, "fRefCnt was %d", fRefCnt); 81 fRefCnt = 0; // illegal value, to catch us if we reuse after delete 87 int32_t getRefCnt() const { return fRefCnt; } 94 if (1 == sk_atomic_load(&fRefCnt, sk_memory_order_acquire)) { 107 SkASSERT(fRefCnt > 0); 108 (void)sk_atomic_fetch_add(&fRefCnt, +1, sk_memory_order_relaxed); // No barrier required. 117 SkASSERT(fRefCnt > 0) [all...] |
/external/skia/src/core/ |
SkCachedData.cpp | 34 , fRefCnt(1) 46 , fRefCnt(1) 99 if ((1 == fRefCnt) && fInCache) { 103 fRefCnt += 1; 111 switch (--fRefCnt) { 136 return 0 == fRefCnt; 185 SkASSERT((fInCache && fRefCnt > 1) || !fInCache); 195 SkASSERT((fInCache && 1 == fRefCnt) || (0 == fRefCnt));
|
SkCachedData.h | 30 int testing_only_getRefCnt() const { return fRefCnt; } 56 int fRefCnt; // low-bit means we're owned by the cache
|
SkRegionPriv.h | 45 int32_t fRefCnt; 79 head->fRefCnt = 1; 102 SkASSERT(fRefCnt == 1); 112 if (fRefCnt > 1) { 123 if (sk_atomic_dec(&fRefCnt) == 1) {
|
/external/skqp/src/core/ |
SkCachedData.cpp | 34 , fRefCnt(1) 46 , fRefCnt(1) 99 if ((1 == fRefCnt) && fInCache) { 103 fRefCnt += 1; 111 switch (--fRefCnt) { 136 return 0 == fRefCnt; 185 SkASSERT((fInCache && fRefCnt > 1) || !fInCache); 195 SkASSERT((fInCache && 1 == fRefCnt) || (0 == fRefCnt));
|
SkCachedData.h | 30 int testing_only_getRefCnt() const { return fRefCnt; } 56 int fRefCnt; // low-bit means we're owned by the cache
|
SkRegionPriv.h | 45 int32_t fRefCnt; 79 head->fRefCnt = 1; 102 SkASSERT(fRefCnt == 1); 112 if (fRefCnt > 1) { 123 if (sk_atomic_dec(&fRefCnt) == 1) {
|
/external/skia/include/private/ |
GrSurfaceProxy.h | 36 ++fRefCnt; 49 --fRefCnt; 55 SkASSERT(fRefCnt >= 0 && fPendingWrites >= 0 && fPendingReads >= 0); 56 return 1 == fRefCnt + fPendingWrites + fPendingReads; 61 SkASSERT(1 == fRefCnt); 73 SkASSERT(fRefCnt >= 0); 76 SkASSERT(fRefCnt + fPendingReads + fPendingWrites >= 1); 82 SkASSERT(fTarget->fRefCnt >= fRefCnt); 135 GrIORefProxy() : fTarget(nullptr), fRefCnt(1), fPendingReads(0), fPendingWrites(0) { [all...] |
SkWeakRefCnt.h | 83 /** If fRefCnt is 0, returns 0. 84 * Otherwise increments fRefCnt, acquires, and returns the old value. 87 int32_t prev = fRefCnt.load(std::memory_order_relaxed); 92 } while(!fRefCnt.compare_exchange_weak(prev, prev+1, std::memory_order_acquire, 147 return fRefCnt.load(std::memory_order_relaxed) == 0; 169 /* Invariant: fWeakCnt = #weak + (fRefCnt > 0 ? 1 : 0) */
|
/external/skqp/include/private/ |
GrSurfaceProxy.h | 36 ++fRefCnt; 49 --fRefCnt; 55 SkASSERT(fRefCnt >= 0 && fPendingWrites >= 0 && fPendingReads >= 0); 56 return 1 == fRefCnt + fPendingWrites + fPendingReads; 62 SkASSERT(fRefCnt >= 0); 65 SkASSERT(fRefCnt + fPendingReads + fPendingWrites >= 1); 71 SkASSERT(fTarget->fRefCnt >= fRefCnt); 124 GrIORefProxy() : fTarget(nullptr), fRefCnt(1), fPendingReads(0), fPendingWrites(0) {} 125 GrIORefProxy(sk_sp<GrSurface> surface) : fRefCnt(1), fPendingReads(0), fPendingWrites(0) [all...] |
SkWeakRefCnt.h | 83 /** If fRefCnt is 0, returns 0. 84 * Otherwise increments fRefCnt, acquires, and returns the old value. 87 int32_t prev = fRefCnt.load(std::memory_order_relaxed); 92 } while(!fRefCnt.compare_exchange_weak(prev, prev+1, std::memory_order_acquire, 147 return fRefCnt.load(std::memory_order_relaxed) == 0; 169 /* Invariant: fWeakCnt = #weak + (fRefCnt > 0 ? 1 : 0) */
|
/external/skia/include/gpu/ |
GrGpuResource.h | 54 ++fRefCnt; 60 if (!(--fRefCnt)) { 71 SkASSERT(fRefCnt >= 0); 74 SkASSERT(fRefCnt + fPendingReads + fPendingWrites >= 0); 79 GrIORef() : fRefCnt(1), fPendingReads(0), fPendingWrites(0) { } 93 bool internalHasRef() const { return SkToBool(fRefCnt); } 94 bool internalHasUniqueRef() const { return fRefCnt == 1; } 126 if (0 == fPendingReads && 0 == fPendingWrites && 0 == fRefCnt) { 131 mutable int32_t fRefCnt;
|
/external/skqp/include/gpu/ |
GrGpuResource.h | 54 ++fRefCnt; 60 if (!(--fRefCnt)) { 71 SkASSERT(fRefCnt >= 0); 74 SkASSERT(fRefCnt + fPendingReads + fPendingWrites >= 0); 79 GrIORef() : fRefCnt(1), fPendingReads(0), fPendingWrites(0) { } 93 bool internalHasRef() const { return SkToBool(fRefCnt); } 94 bool internalHasUniqueRef() const { return fRefCnt == 1; } 126 if (0 == fPendingReads && 0 == fPendingWrites && 0 == fRefCnt) { 131 mutable int32_t fRefCnt;
|
/external/skia/include/core/ |
SkRefCnt.h | 33 SkRefCntBase() : fRefCnt(1) {} 39 SkASSERTF(getRefCnt() == 1, "fRefCnt was %d", getRefCnt()); 41 fRefCnt.store(0, std::memory_order_relaxed); 48 return fRefCnt.load(std::memory_order_relaxed); 60 if (1 == fRefCnt.load(std::memory_order_acquire)) { 74 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); 84 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { 94 * so they can reset fRefCnt before the destructor is called or if they 99 fRefCnt.store(1, std::memory_order_relaxed); 115 mutable std::atomic<int32_t> fRefCnt; [all...] |
/external/skqp/include/core/ |
SkRefCnt.h | 33 SkRefCntBase() : fRefCnt(1) {} 39 SkASSERTF(getRefCnt() == 1, "fRefCnt was %d", getRefCnt()); 41 fRefCnt.store(0, std::memory_order_relaxed); 48 return fRefCnt.load(std::memory_order_relaxed); 60 if (1 == fRefCnt.load(std::memory_order_acquire)) { 74 (void)fRefCnt.fetch_add(+1, std::memory_order_relaxed); 84 if (1 == fRefCnt.fetch_add(-1, std::memory_order_acq_rel)) { 94 * so they can reset fRefCnt before the destructor is called or if they 99 fRefCnt.store(1, std::memory_order_relaxed); 115 mutable std::atomic<int32_t> fRefCnt; [all...] |
/external/skia/tests/ |
RefCntTest.cpp | 91 Effect() : fRefCnt(1) { 96 int fRefCnt; 100 fRefCnt += 1; 105 SkASSERT(fRefCnt > 0); 106 if (0 == --fRefCnt) { 160 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 1); 193 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 2); 198 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 3);
|
ProxyRefTest.cpp | 23 return fRefCnt; 28 return fTarget->fRefCnt; 31 return fRefCnt;
|
/external/skqp/tests/ |
RefCntTest.cpp | 91 Effect() : fRefCnt(1) { 96 int fRefCnt; 100 fRefCnt += 1; 105 SkASSERT(fRefCnt > 0); 106 if (0 == --fRefCnt) { 160 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 1); 193 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 2); 198 REPORTER_ASSERT(reporter, paint.fEffect.get()->fRefCnt == 3);
|
ProxyRefTest.cpp | 23 return fRefCnt; 28 return fTarget->fRefCnt; 31 return fRefCnt;
|
/external/skia/src/gpu/ops/ |
GrDrawPathOp.h | 129 instanceData->fRefCnt = 1; 153 void ref() const { ++fRefCnt; } 156 if (0 == --fRefCnt) { 171 mutable int fRefCnt;
|