Lines Matching refs:page
42 Page* PageIterator::next() {
123 void MemoryAllocator::ProtectChunkFromPage(Page* page) {
124 int id = GetChunkId(page);
129 void MemoryAllocator::UnprotectChunkFromPage(Page* page) {
130 int id = GetChunkId(page);
140 Page* Page::Initialize(Heap* heap,
144 Page* page = reinterpret_cast<Page*>(chunk);
145 ASSERT(page->area_size() <= kMaxRegularHeapObjectSize);
147 owner->IncreaseCapacity(page->area_size());
148 owner->Free(page->area_start(), page->area_size());
152 return page;
157 Page* p = Page::FromAddress(addr);
177 OffsetFrom(addr) & ~Page::kPageAlignmentMask);
199 // to another chunk. See the comment to Page::FromAllocationTop.
215 Page* Page::next_page() {
217 return static_cast<Page*>(next_chunk());
221 Page* Page::prev_page() {
223 return static_cast<Page*>(prev_chunk());
227 void Page::set_next_page(Page* page) {
228 ASSERT(page->owner() == owner());
229 set_next_chunk(page);
233 void Page::set_prev_page(Page* page) {
234 ASSERT(page->owner() == owner());
235 set_prev_chunk(page);
239 // Try linear allocation in the page of alloc_info's allocation top. Does
240 // not contain slow case logic (e.g. move to the next page or try free list