Lines Matching full:page
2 * virtual page mapping and translated block handling
91 /* any access to the tbs or the page table must use this lock */
145 /* list of TBs intersecting this ram page */
148 of lookups we do to a given page to use a bitmap */
157 /* offset in host memory of the page + io_index in the low bits */
688 printf("ERROR page flags: PC=%08lx size=%04x f1=%x f2=%x\n",
798 /* remove the TB from the page list */
888 /* NOTE: tb_end may be after the end of the page, but
934 /* check next page if needed */
944 /* invalidate all TBs which intersect with the target physical page
946 the same physical page. 'is_cpu_write_access' should be true if called
985 /* NOTE: tb_end may be after the end of the page, but
1143 /* add the tb in the target page and protect it if necessary */
1165 /* force the host page as non writable (writes will have a
1166 page fault + mprotect overhead) */
1182 printf("protecting code page: 0x" TARGET_FMT_lx "\n",
1189 allocated in a physical page */
1228 /* add a new TB and link it to the physical page tables. phys_page2 is
1229 (-1) to indicate that only one page contains the TB. */
1245 /* add in the page list */
1772 overlap the flushed page. */
1854 /* update the TLBs so that writes to code in the virtual page 'addr'
1863 /* update the TLB so that writes in physical page 'phys_addr' are no longer
1992 /* update the TLB corresponding to virtual page vaddr
2006 is permitted. Return 0 if OK or 2 if the page could not be mapped
2108 * If we have memchecker running, we need to make sure that page, cached
2113 * We need to check with memory checker if we should invalidate this page
2116 * - Page that's been cached belongs to the user space.
2117 * - Request to cache this page didn't come from softmmu. We're covered
2118 * there, because after page was cached here we will invalidate it in
2120 * - Cached page belongs to RAM, not I/O area.
2121 * - Page is cached for read, or write access.
2232 /* modify the flags of a page and invalidate the code if
2287 /* unprotect the page if it was put read-only because it
2300 page. Return TRUE if the fault was successfully handled. */
2326 /* if the page was really writable, then we change its
2381 page size. If (phys_offset & ~TARGET_PAGE_MASK) != 0, then it is an
2382 io memory page. The address used when calling the IO function is
2384 start_addr and region_offset are rounded down to a page boundary
3137 target_ulong page;
3141 page = addr & TARGET_PAGE_MASK;
3142 l = (page + TARGET_PAGE_SIZE) - addr;
3145 flags = page_get_flags(page);
3180 target_phys_addr_t page;
3185 page = addr & TARGET_PAGE_MASK;
3186 l = (page + TARGET_PAGE_SIZE) - addr;
3189 p = phys_page_find(page >> TARGET_PAGE_BITS);
3277 target_phys_addr_t page;
3282 page = addr & TARGET_PAGE_MASK;
3283 l = (page + TARGET_PAGE_SIZE) - addr;
3286 p = phys_page_find(page >> TARGET_PAGE_BITS);
3371 target_phys_addr_t page;
3377 page = addr & TARGET_PAGE_MASK;
3378 l = (page + TARGET_PAGE_SIZE) - addr;
3381 p = phys_page_find(page >> TARGET_PAGE_BITS);
3539 /* warning: addr must be aligned. The ram page is not masked as dirty
3675 target_ulong page;
3678 page = addr & TARGET_PAGE_MASK;
3679 phys_addr = cpu_get_phys_page_debug(env, page);
3680 /* if no physical page mapped, return an error */
3683 l = (page + TARGET_PAGE_SIZE) - addr;
3797 cpu_fprintf(f, "cross page TB count %d (%d%%)\n",