Lines Matching refs:mapping
185 * mapping from the currently active vm_flags protection bits (the
233 unsigned long private; /* Mapping-private opaque data:
240 struct address_space *mapping; /* If low bit clear, points to
252 pgoff_t index; /* Our offset within mapping. */
285 * - cache mapping (page->mapping)
287 * - page mapped in a task's page tables, each mapping
365 * A page may belong to an inode's memory mapping. In this case,
366 * page->mapping is the pointer to the inode, and page->index is the
380 * There is also a per-mapping radix tree mapping index to the page
381 * in memory if present. The tree is rooted at mapping->root.
431 * We are going to use the flags for the page to node mapping if its in
561 * page->mapping points to its anon_vma, not to a struct address_space;
573 struct address_space *mapping = page->mapping;
576 mapping = &swapper_space;
577 else if (unlikely((unsigned long)mapping & PAGE_MAPPING_ANON))
578 mapping = NULL;
579 return mapping;
584 return ((unsigned long)page->mapping & PAGE_MAPPING_ANON) != 0;
704 struct address_space *check_mapping; /* Check page->mapping if set */
726 void unmap_mapping_range(struct address_space *mapping,
729 static inline void unmap_shared_mapping_range(struct address_space *mapping,
732 unmap_mapping_range(mapping, holebegin, holelen, 0);
835 * When freeing, reset page->mapping so free_pages_check won't complain.
841 #define pte_lock_deinit(page) ((page)->mapping = NULL)
976 int do_page_cache_readahead(struct address_space *mapping, struct file *filp,
978 int force_page_cache_readahead(struct address_space *mapping, struct file *filp,
980 unsigned long page_cache_readahead(struct address_space *mapping,
985 void handle_ra_miss(struct address_space *mapping,