/external/v8/src/heap/ |
incremental-marking.cc | 29 : heap_(heap), 71 heap_->mark_compact_collector()->RecordSlot(obj, slot, value); 108 RelocInfo rinfo(heap_->isolate(), pc, RelocInfo::CODE_TARGET, 0, host); 116 Code* host = heap_->isolate() 119 RelocInfo rinfo(heap_->isolate(), pc, RelocInfo::CODE_TARGET, 0, host); 130 heap_->mark_compact_collector()->RecordCodeEntrySlot( 139 heap_->mark_compact_collector()->RecordRelocSlot(host, rinfo, value); 146 heap_->mark_compact_collector()->marking_deque()->Push(obj); 297 : heap_(incremental_marking->heap()) {} 310 IncrementalMarking::MarkObject(heap_, HeapObject::cast(obj)) 313 Heap* heap_; member in class:v8::internal::IncrementalMarkingRootMarkingVisitor [all...] |
gc-tracer.cc | 26 start_time_ = tracer_->heap_->MonotonicallyIncreasingTimeInMs(); 29 RuntimeCallStats::Enter(tracer_->heap_->isolate(), &timer_, 38 tracer_->heap_->MonotonicallyIncreasingTimeInMs() - start_time_; 41 RuntimeCallStats::Leave(tracer_->heap_->isolate(), &timer_); 114 : heap_(heap), 134 current_.end_time = heap_->MonotonicallyIncreasingTimeInMs(); 165 double start_time = heap_->MonotonicallyIncreasingTimeInMs(); 166 SampleAllocation(start_time, heap_->NewSpaceAllocationCounter(), 167 heap_->OldGenerationAllocationCounter()); 174 if (heap_->incremental_marking()->WasActivated()) [all...] |
scavenger.h | 19 explicit Scavenger(Heap* heap) : heap_(heap) {} 40 Heap* heap() { return heap_; } 43 Heap* heap_; member in class:v8::internal::Scavenger 52 explicit ScavengeVisitor(Heap* heap) : heap_(heap) {} 60 Heap* heap_; member in class:v8::internal::ScavengeVisitor
|
store-buffer.h | 36 Heap* heap_; member in class:v8::internal::StoreBuffer
|
memory-reducer.h | 110 : heap_(heap), 129 Heap* heap() { return heap_; } 154 Heap* heap_; member in class:v8::internal::MemoryReducer
|
page-parallel-job.h | 44 : heap_(heap), 91 Task* task = new Task(heap_, items_, num_items_, start_index, 114 JobTraits::FinalizePageSequentially(heap_, item->chunk, success, 140 heap_(heap), 161 heap_, data_, current->chunk, current->data); 173 Heap* heap_; member in class:v8::internal::PageParallelJob::Task 182 Heap* heap_; member in class:v8::internal::PageParallelJob
|
store-buffer.cc | 19 : heap_(heap), 69 DCHECK(!heap_->code_space()->Contains(*current)); 71 Page* page = Page::FromAnyPointerAddress(heap_, addr);
|
array-buffer-tracker.h | 68 explicit LocalArrayBufferTracker(Heap* heap) : heap_(heap) {} 95 Heap* heap_; member in class:v8::internal::LocalArrayBufferTracker
|
object-stats.h | 17 explicit ObjectStats(Heap* heap) : heap_(heap) {} 72 Heap* heap() { return heap_; } 75 Heap* heap_; member in class:v8::internal::ObjectStats
|
array-buffer-tracker.cc | 23 heap_->isolate()->array_buffer_allocator()->Free(it->second.first, 32 heap_->update_external_memory_concurrently_freed( 62 heap_->isolate()->array_buffer_allocator()->Free(it->second.first, 71 heap_->update_external_memory_concurrently_freed(
|
/system/core/libmemunreachable/tests/ |
LeakFolding_test.cpp | 26 LeakFoldingTest() : disable_malloc_(), heap_() {} 29 ASSERT_TRUE(heap_.empty()); 37 Heap heap_; member in class:LeakFoldingTest 48 HeapWalker heap_walker(heap_); 52 LeakFolding folding(heap_, heap_walker); 56 allocator::vector<LeakFolding::Leak> leaked(heap_); 72 HeapWalker heap_walker(heap_); 77 LeakFolding folding(heap_, heap_walker); 81 allocator::vector<LeakFolding::Leak> leaked(heap_); 101 HeapWalker heap_walker(heap_); [all...] |
HeapWalker_test.cpp | 28 HeapWalkerTest() : disable_malloc_(), heap_() {} 31 ASSERT_TRUE(heap_.empty()); 39 Heap heap_; member in class:HeapWalkerTest 43 HeapWalker heap_walker(heap_); 52 HeapWalker heap_walker(heap_); 66 HeapWalker heap_walker(heap_); 83 HeapWalker heap_walker(heap_); 88 allocator::vector<Range> leaked(heap_); 110 HeapWalker heap_walker(heap_); 116 allocator::vector<Range> leaked(heap_); [all...] |
/art/runtime/gc/accounting/ |
remembered_set.h | 51 : name_(name), heap_(heap), space_(space) {} 68 return heap_; 77 Heap* const heap_; member in class:art::gc::accounting::RememberedSet
|
mod_union_table.h | 53 heap_(heap), 88 return heap_; 97 Heap* const heap_; member in class:art::gc::accounting::ModUnionTable
|
heap_bitmap.h | 67 explicit HeapBitmap(Heap* heap) : heap_(heap) {} 70 const Heap* const heap_; member in class:art::gc::accounting::HeapBitmap
|
/external/v8/src/ |
identity-map.cc | 21 heap_->UnregisterStrongRoots(keys_); 43 CHECK_NE(address, heap_->not_mapped_symbol()); 51 Object* not_mapped = heap_->not_mapped_symbol(); 65 Object* not_mapped = heap_->not_mapped_symbol(); 94 gc_counter_ = heap_->gc_count(); 97 Object* not_mapped = heap_->not_mapped_symbol(); 102 heap_->RegisterStrongRoots(keys_, keys_ + size_); 109 if (gc_counter_ != heap_->gc_count()) Rehash(); 125 if (result == nullptr && gc_counter_ != heap_->gc_count()) { 135 gc_counter_ = heap_->gc_count() [all...] |
isolate-inl.h | 36 thread_local_top_.pending_exception_ = heap_.the_hole_value(); 47 thread_local_top_.pending_message_obj_ = heap_.the_hole_value(); 60 return thread_local_top_.scheduled_exception_ != heap_.the_hole_value(); 66 thread_local_top_.scheduled_exception_ = heap_.the_hole_value();
|
identity-map.h | 29 : heap_(heap), 53 Heap* heap_; member in class:v8::internal::IdentityMapBase
|
/system/core/libmemunreachable/ |
Allocator.h | 113 heap_(heap) { 119 heap_(other.heap_) { 126 return reinterpret_cast<T*>(heap_.allocate(n * sizeof(T))); 130 heap_.deallocate(ptr); 135 return heap_ == other.heap_; 146 Heap heap_; 172 using STLAllocator<T>::heap_; 178 heap_.deallocate(ptr) [all...] |
/external/v8/src/profiler/ |
sampling-heap-profiler.h | 122 Heap* heap() const { return heap_; } 143 Heap* const heap_; member in class:v8::internal::SamplingHeapProfiler 163 heap_(heap), 170 USE(heap_); variable 171 DCHECK(heap_->gc_state() == Heap::NOT_IN_GC); 181 Heap* const heap_; member in class:v8::internal::SamplingAllocationObserver
|
/art/runtime/gc/collector/ |
garbage_collector.cc | 60 : heap_(heap), 124 heap_->GetLiveBitmap()->ReplaceBitmap(live_bitmap, mark_bitmap); 125 heap_->GetMarkBitmap()->ReplaceBitmap(mark_bitmap, live_bitmap); 135 heap_->GetLiveBitmap()->ReplaceLargeObjectBitmap(live_set, mark_set); 136 heap_->GetMarkBitmap()->ReplaceLargeObjectBitmap(mark_set, live_set); 169 return heap_->GetCurrentGcIteration(); 172 return heap_->GetCurrentGcIteration(); 177 heap_->RecordFree(freed.objects, freed.bytes); 181 heap_->RecordFree(freed.objects, freed.bytes);
|
mark_compact.cc | 107 mark_stack_ = heap_->GetMarkStack(); 113 mark_bitmap_ = heap_->GetMarkBitmap(); 119 heap_->GetReferenceProcessor()->ProcessReferences( 169 heap_->ProcessCards(GetTimings(), false, false, true); 173 heap_->GetCardTable()->ClearCardTable(); 178 heap_->RevokeAllThreadLocalAllocationStacks(self); 181 heap_->SwapStacks(); 201 // heap_->PreSweepingGcVerification(this); 206 for (auto& space : heap_->GetContinuousSpaces()) { 209 accounting::ModUnionTable* table = heap_->FindModUnionTableFromSpace(space) [all...] |
mark_sweep-inl.h | 35 DCHECK(IsMarked(obj)) << "Scanning unmarked object " << obj << "\n" << heap_->DumpSpaces();
|
concurrent_copying-inl.h | 35 DCHECK(heap_->collector_type_ == kCollectorTypeCC); 66 DCHECK(region_space_->IsInToSpace(to_ref) || heap_->non_moving_space_->HasAddress(to_ref))
|
/art/patchoat/ |
patchoat.h | 68 : oat_file_(oat_file), image_(nullptr), bitmap_(nullptr), heap_(nullptr), delta_(delta), 72 : image_(image), bitmap_(bitmap), heap_(heap), 77 : oat_file_(oat_file), image_(image), bitmap_(bitmap), heap_(heap), 138 DCHECK_GT(reinterpret_cast<uintptr_t>(obj), reinterpret_cast<uintptr_t>(heap_->Begin())); 139 DCHECK_LT(reinterpret_cast<uintptr_t>(obj), reinterpret_cast<uintptr_t>(heap_->End())); 141 reinterpret_cast<uintptr_t>(obj) - reinterpret_cast<uintptr_t>(heap_->Begin()); 218 const MemMap* const heap_; member in class:art::PatchOat
|