/external/v8/src/crankshaft/ |
lithium-codegen.h | 34 Heap* heap() const { return isolate()->heap(); } function in class:v8::internal::BASE_EMBEDDED
|
/external/v8/src/ic/ |
access-compiler.h | 52 Heap* heap() const { return isolate()->heap(); } function in class:v8::internal::BASE_EMBEDDED
|
/art/runtime/gc/space/ |
space.h | 38 class Heap; 188 friend class art::gc::Heap; 213 // This is what is to be added to Heap::num_bytes_allocated_. 237 // from Heap::num_bytes_allocated_ or zero if unnecessary. 243 // from Heap::num_bytes_allocated_ or zero if unnecessary. 374 // for the allocation space we support starting with a small heap and then extending it. 405 // Used by the heap compaction interface to enable copying from one type of alloc space to another. 453 friend class gc::Heap;
|
/external/v8/src/snapshot/ |
deserializer.cc | 11 #include "src/heap/heap-inl.h" 38 for (Page* p : *isolate_->heap()->code_space()) { 48 isolate_->heap()->RecordWritesIntoCode(code); 63 if (!isolate_->heap()->ReserveSpace(reservations_, &allocated_maps_)) 93 isolate_->heap()->IterateStrongRoots(this, VISIT_ONLY_STRONG_ROOT_LIST); 94 isolate_->heap()->IterateSmiRoots(this); 95 isolate_->heap()->IterateStrongRoots(this, VISIT_ONLY_STRONG); 96 isolate_->heap()->RepairFreeListsAfterDeserialization(); 97 isolate_->heap()->IterateWeakRoots(this, VISIT_ALL) 344 Heap* heap = isolate->heap(); local 404 Heap* heap = isolate_->heap(); local [all...] |
/external/v8/src/builtins/mips/ |
builtins-mips.cc | 118 __ LoadRoot(a2, Heap::kUndefinedValueRootIndex); 133 Heap::RootListIndex const root_index = 134 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 135 : Heap::kMinusInfinityValueRootIndex; 160 __ JumpIfRoot(t0, Heap::kHeapNumberMapRootIndex, &convert_number); 227 __ LoadRoot(t2, Heap::kNanValueRootIndex); 400 __ LoadRoot(v0, Heap::kempty_stringRootIndex); 457 __ LoadRoot(a0, Heap::kempty_stringRootIndex); 552 __ LoadRoot(t0, Heap::kStackLimitRootIndex); 613 __ PushRoot(Heap::kTheHoleValueRootIndex) [all...] |
/external/v8/src/builtins/x64/ |
builtins-x64.cc | 108 __ CompareRoot(rsp, Heap::kStackLimitRootIndex); 165 __ PushRoot(Heap::kTheHoleValueRootIndex); 191 masm->isolate()->heap()->SetConstructStubInvokeDeoptPCOffset( 256 masm->isolate()->heap()->SetConstructStubCreateDeoptPCOffset( 318 __ LoadRoot(kScratchRegister, Heap::kRealStackLimitRootIndex); 546 __ PushRoot(Heap::kTheHoleValueRootIndex); 701 __ CompareRoot(rdx, Heap::kRealStackLimitRootIndex); 709 __ LoadRoot(rdx, Heap::kUndefinedValueRootIndex); 721 __ LoadRoot(kInterpreterAccumulatorRegister, Heap::kUndefinedValueRootIndex); 732 masm->isolate()->heap()->SetInterpreterEntryReturnPCOffset(masm->pc_offset()) [all...] |
/external/v8/src/builtins/arm/ |
builtins-arm.cc | 114 __ LoadRoot(r2, Heap::kUndefinedValueRootIndex); 131 Heap::RootListIndex const root_index = 132 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 133 : Heap::kMinusInfinityValueRootIndex; 158 __ JumpIfRoot(r3, Heap::kHeapNumberMapRootIndex, &convert_number); 212 __ LoadRoot(r5, Heap::kNanValueRootIndex); 382 __ LoadRoot(r0, Heap::kempty_stringRootIndex); 438 __ LoadRoot(r2, Heap::kempty_stringRootIndex); 540 __ LoadRoot(ip, Heap::kStackLimitRootIndex); 603 __ PushRoot(Heap::kTheHoleValueRootIndex) [all...] |
/external/v8/src/builtins/ppc/ |
builtins-ppc.cc | 114 __ LoadRoot(r5, Heap::kUndefinedValueRootIndex); 130 Heap::RootListIndex const root_index = 131 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 132 : Heap::kMinusInfinityValueRootIndex; 164 __ JumpIfRoot(r6, Heap::kHeapNumberMapRootIndex, &convert_number); 217 __ LoadRoot(r8, Heap::kNanValueRootIndex); 393 __ LoadRoot(r3, Heap::kempty_stringRootIndex); 451 __ LoadRoot(r5, Heap::kempty_stringRootIndex); 548 __ LoadRoot(ip, Heap::kStackLimitRootIndex); 584 __ PushRoot(Heap::kTheHoleValueRootIndex) [all...] |
/external/v8/src/builtins/s390/ |
builtins-s390.cc | 114 __ LoadRoot(r4, Heap::kUndefinedValueRootIndex); 130 Heap::RootListIndex const root_index = 131 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 132 : Heap::kMinusInfinityValueRootIndex; 163 __ JumpIfRoot(r5, Heap::kHeapNumberMapRootIndex, &convert_number); 217 __ LoadRoot(r7, Heap::kNanValueRootIndex); 391 __ LoadRoot(r2, Heap::kempty_stringRootIndex); 448 __ LoadRoot(r4, Heap::kempty_stringRootIndex); 545 __ CmpLogicalP(sp, RootMemOperand(Heap::kStackLimitRootIndex)); 581 __ PushRoot(Heap::kTheHoleValueRootIndex) [all...] |
/external/v8/src/heap/ |
spaces.h | 20 #include "src/heap/heap.h" 21 #include "src/heap/marking.h" 48 // Heap structures: 50 // A JS heap consists of a young generation, an old generation, and a large 67 // references. See heap/store-buffer.h. 70 // buffer overflow) iterate intergenerational pointers without decoding heap 74 // the Heap::InNewSpace() predicate must be a pointer to a live heap object in 165 void RepairFreeList(Heap* heap) 438 inline Heap* heap() const { return heap_; } function in class:v8::internal::MemoryChunk 842 Heap* heap() const { return heap_; } function in class:v8::internal::Space [all...] |
scavenger.cc | 5 #include "src/heap/scavenger.h" 8 #include "src/heap/heap-inl.h" 9 #include "src/heap/incremental-marking.h" 10 #include "src/heap/objects-visiting-inl.h" 11 #include "src/heap/scavenger-inl.h" 14 #include "src/profiler/heap-profiler.h" 104 static void RecordCopiedObject(Heap* heap, HeapObject* obj) { 111 if (heap->new_space()->Contains(obj)) 159 Heap* heap = map->GetHeap(); local 188 Heap* heap = map->GetHeap(); local 219 Heap* heap = map->GetHeap(); local 316 Heap* heap = map->GetHeap(); local [all...] |
/external/v8/src/builtins/mips64/ |
builtins-mips64.cc | 118 __ LoadRoot(a2, Heap::kUndefinedValueRootIndex); 133 Heap::RootListIndex const root_index = 134 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 135 : Heap::kMinusInfinityValueRootIndex; 160 __ JumpIfRoot(a4, Heap::kHeapNumberMapRootIndex, &convert_number); 224 __ LoadRoot(t1, Heap::kNanValueRootIndex); 399 __ LoadRoot(v0, Heap::kempty_stringRootIndex); 456 __ LoadRoot(a0, Heap::kempty_stringRootIndex); 550 __ LoadRoot(a4, Heap::kStackLimitRootIndex); 607 __ PushRoot(Heap::kTheHoleValueRootIndex) [all...] |
/external/v8/src/ia32/ |
macro-assembler-ia32.h | 74 void LoadRoot(Register destination, Heap::RootListIndex index); 75 void StoreRoot(Register source, Register scratch, Heap::RootListIndex index); 76 void CompareRoot(Register with, Register scratch, Heap::RootListIndex index); 79 void CompareRoot(Register with, Heap::RootListIndex index); 80 void CompareRoot(const Operand& with, Heap::RootListIndex index); 81 void PushRoot(Heap::RootListIndex index); 84 void JumpIfRoot(Register with, Heap::RootListIndex index, Label* if_equal, 89 void JumpIfRoot(const Operand& with, Heap::RootListIndex index, 97 void JumpIfNotRoot(Register with, Heap::RootListIndex index, 103 void JumpIfNotRoot(const Operand& with, Heap::RootListIndex index [all...] |
/external/v8/src/x87/ |
macro-assembler-x87.h | 77 void LoadRoot(Register destination, Heap::RootListIndex index); 78 void StoreRoot(Register source, Register scratch, Heap::RootListIndex index); 79 void CompareRoot(Register with, Register scratch, Heap::RootListIndex index); 82 void CompareRoot(Register with, Heap::RootListIndex index); 83 void CompareRoot(const Operand& with, Heap::RootListIndex index); 84 void PushRoot(Heap::RootListIndex index); 87 void JumpIfRoot(Register with, Heap::RootListIndex index, Label* if_equal, 92 void JumpIfRoot(const Operand& with, Heap::RootListIndex index, 100 void JumpIfNotRoot(Register with, Heap::RootListIndex index, 106 void JumpIfNotRoot(const Operand& with, Heap::RootListIndex index [all...] |
/external/v8/src/ |
code-stub-assembler.cc | 148 return LoadRoot(Heap::k##rootName##RootIndex); \ 161 return LoadAndUntagToWord32Root(Heap::kHashSeedRootIndex); 165 return LoadRoot(Heap::kStaleRegisterRootIndex); 618 Node* empty_elements = LoadRoot(Heap::kEmptyFixedArrayRootIndex); 782 LoadRoot(Heap::kOnePointerFillerMapRootIndex)); 957 Heap::RootListIndex root_index) { [all...] |
layout-descriptor.h | 79 LayoutDescriptor* Trim(Heap* heap, Map* map, DescriptorArray* descriptors,
|
objects-body-descriptors-inl.h | 58 void BodyDescriptorBase::IterateBodyImpl(Heap* heap, HeapObject* obj, 61 IteratePointers<StaticVisitor>(heap, obj, start_offset, end_offset); 72 IteratePointers<StaticVisitor>(heap, obj, offset, end_of_region_offset); 91 void BodyDescriptorBase::IteratePointers(Heap* heap, HeapObject* obj, 93 StaticVisitor::VisitPointers(heap, obj, 107 void BodyDescriptorBase::IteratePointer(Heap* heap, HeapObject* obj, 109 StaticVisitor::VisitPointer(heap, obj, HeapObject::RawField(obj, offset)) 143 Heap* heap = obj->GetHeap(); local 185 Heap* heap = obj->GetHeap(); local 214 Heap* heap = obj->GetHeap(); local 240 Heap* heap = obj->GetHeap(); local 273 Heap* heap = obj->GetHeap(); local 411 Heap* heap = obj->GetHeap(); local [all...] |
/external/v8/src/builtins/arm64/ |
builtins-arm64.cc | 114 __ LoadRoot(x2, Heap::kUndefinedValueRootIndex); 132 Heap::RootListIndex const root_index = 133 (kind == MathMaxMinKind::kMin) ? Heap::kInfinityValueRootIndex 134 : Heap::kMinusInfinityValueRootIndex; 373 __ LoadRoot(x0, Heap::kempty_stringRootIndex); 432 __ LoadRoot(x2, Heap::kempty_stringRootIndex); 529 __ CompareRoot(masm->StackPointer(), Heap::kStackLimitRootIndex); 599 __ PushRoot(Heap::kTheHoleValueRootIndex); 643 masm->isolate()->heap()->SetConstructStubInvokeDeoptPCOffset( 712 masm->isolate()->heap()->SetConstructStubCreateDeoptPCOffset [all...] |
/external/v8/src/builtins/ia32/ |
builtins-ia32.cc | 165 __ PushRoot(Heap::kTheHoleValueRootIndex); 191 masm->isolate()->heap()->SetConstructStubInvokeDeoptPCOffset( 254 masm->isolate()->heap()->SetConstructStubCreateDeoptPCOffset( 469 __ PushRoot(Heap::kTheHoleValueRootIndex); 641 __ LoadRoot(kInterpreterAccumulatorRegister, Heap::kUndefinedValueRootIndex); 654 masm->isolate()->heap()->SetInterpreterEntryReturnPCOffset(masm->pc_offset()); [all...] |
/external/v8/src/profiler/ |
heap-snapshot-generator.h | 78 // HeapEntry instances represent an entity from the heap (or a special 148 // HeapSnapshot represents a single heap snapshot. It is stored in 150 // HeapSnapshots. All HeapSnapshots share strings copied from JS heap 218 explicit HeapObjectsMap(Heap* heap); 220 Heap* heap() const { return heap_; } function in class:v8::internal::HeapObjectsMap 267 Heap* heap_; 274 // in any kind of heap memory. 287 // real heap objects and their representations in heap snapshots [all...] |
/art/runtime/gc/accounting/ |
card_table.h | 39 class Heap; 46 // non-null values to heap addresses should go through an entry in 87 // Returns a value that when added to a heap address >> GC_CARD_SHIFT will address the appropriate 127 // Returns the first address in the heap which maps to this card. 130 // Returns the address of the relevant byte in the card table, given an address on the heap.
|
/art/runtime/gc/ |
reference_processor.h | 43 class Heap;
|
/external/v8/src/builtins/ |
builtins-symbol.cc | 45 return *isolate->SymbolFor(Heap::kPublicSymbolTableRootIndex, key, false); 63 result = isolate->heap()->undefined_value(); 65 DCHECK_EQ(isolate->heap()->public_symbol_table()->SlowReverseLookup(*symbol),
|
/external/v8/src/builtins/x87/ |
builtins-x87.cc | 163 __ PushRoot(Heap::kTheHoleValueRootIndex); 186 masm->isolate()->heap()->SetConstructStubDeoptPCOffset(masm->pc_offset()); 435 __ PushRoot(Heap::kTheHoleValueRootIndex); 607 __ LoadRoot(kInterpreterAccumulatorRegister, Heap::kUndefinedValueRootIndex); 620 masm->isolate()->heap()->SetInterpreterEntryReturnPCOffset(masm->pc_offset()); [all...] |
/frameworks/base/libs/hwui/hwui/ |
Bitmap.cpp | 177 , mPixelStorageType(PixelStorageType::Heap) { 178 mPixelStorage.heap.address = address; 179 mPixelStorage.heap.size = size; 229 case PixelStorageType::Heap: 230 free(mPixelStorage.heap.address); 257 case PixelStorageType::Heap: 258 return mPixelStorage.heap.address; 275 case PixelStorageType::Heap: 276 return mPixelStorage.heap.size;
|