/art/runtime/gc/space/ |
bump_pointer_space-inl.h | 58 bytes_allocated_.StoreRelaxed(bytes_allocated_.LoadRelaxed() + num_bytes); 85 bytes_allocated_.FetchAndAddSequentiallyConsistent(num_bytes);
|
bump_pointer_space.cc | 50 objects_allocated_(0), bytes_allocated_(0), 60 objects_allocated_(0), bytes_allocated_(0), 76 bytes_allocated_.StoreRelaxed(0); 163 uint64_t total = static_cast<uint64_t>(bytes_allocated_.LoadRelaxed()); 199 bytes_allocated_.FetchAndAddSequentiallyConsistent(thread->GetThreadLocalBytesAllocated());
|
bump_pointer_space.h | 159 bytes_allocated_.FetchAndSubSequentiallyConsistent(bytes); 182 AtomicInteger bytes_allocated_; // Accumulated from revoked thread local regions. member in class:art::gc::space::FINAL
|
/art/runtime/base/ |
arena_allocator.cc | 138 // may not have the bytes_allocated_ updated correctly. 182 Arena::Arena() : bytes_allocated_(0), memory_(nullptr), size_(0), next_(nullptr) { 262 if (bytes_allocated_ > 0) { 264 bytes_allocated_ = 0; 269 if (bytes_allocated_ > 0) { 270 memset(Begin(), 0, bytes_allocated_); local 271 bytes_allocated_ = 0; 347 MEMORY_TOOL_MAKE_UNDEFINED(arena->memory_, arena->bytes_allocated_); 400 arena_head_->bytes_allocated_ = ptr_ - begin_; 461 new_arena->bytes_allocated_ = bytes; // UpdateBytesAllocated() on the new_arena [all...] |
scoped_arena_allocator.cc | 87 if (top_arena_->bytes_allocated_ < allocated) { 88 top_arena_->bytes_allocated_ = allocated;
|
arena_allocator.h | 218 return Size() - bytes_allocated_; 222 return bytes_allocated_; 227 return memory_ <= ptr && ptr < memory_ + bytes_allocated_; 231 size_t bytes_allocated_; member in class:art::Arena
|
/external/v8/src/heap/ |
incremental-marking.cc | 470 bytes_allocated_ = 0; [all...] |
incremental-marking.h | 288 size_t bytes_allocated_; member in class:v8::internal::IncrementalMarking
|