/bionic/libc/kernel/arch-mips/asm/ |
tlb.h | 21 #define tlb_start_vma(tlb, vma) do { if (!tlb->fullmm) flush_cache_range(vma, vma->vm_start, vma->vm_end); } while (0) 22 #define tlb_end_vma(tlb, vma) do { } while (0)
|
tlbflush.h | 25 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 27 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/development/ndk/platforms/android-9/arch-mips/include/asm/ |
tlb.h | 21 #define tlb_start_vma(tlb, vma) do { if (!tlb->fullmm) flush_cache_range(vma, vma->vm_start, vma->vm_end); } while (0) 22 #define tlb_end_vma(tlb, vma) do { } while (0)
|
tlbflush.h | 25 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 27 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/external/kernel-headers/original/asm-mips/ |
tlb.h | 5 * MIPS doesn't need any special per-pte or per-vma handling, except 8 #define tlb_start_vma(tlb, vma) \ 11 flush_cache_range(vma, vma->vm_start, vma->vm_end); \ 13 #define tlb_end_vma(tlb, vma) do { } while (0)
|
tlbflush.h | 11 * - flush_tlb_page(vma, vmaddr) flushes one page 12 * - flush_tlb_range(vma, start, end) flushes a range of pages 17 extern void local_flush_tlb_range(struct vm_area_struct *vma, 21 extern void local_flush_tlb_page(struct vm_area_struct *vma, 29 extern void flush_tlb_range(struct vm_area_struct *vma, unsigned long, 39 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 42 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
cacheflush.h | 22 * - flush_cache_range(vma, start, end) flushes a range of pages 36 extern void (*flush_cache_range)(struct vm_area_struct *vma, 38 extern void (*flush_cache_page)(struct vm_area_struct *vma, unsigned long page, unsigned long pfn); 53 static inline void flush_anon_page(struct vm_area_struct *vma, 60 static inline void flush_icache_page(struct vm_area_struct *vma, 84 extern void copy_to_user_page(struct vm_area_struct *vma, 88 extern void copy_from_user_page(struct vm_area_struct *vma,
|
/prebuilts/gcc/darwin-x86/mips/mipsel-linux-android-4.4.3/sysroot/usr/include/asm/ |
tlb.h | 21 #define tlb_start_vma(tlb, vma) do { if (!tlb->fullmm) flush_cache_range(vma, vma->vm_start, vma->vm_end); } while (0) 22 #define tlb_end_vma(tlb, vma) do { } while (0)
|
tlbflush.h | 25 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 27 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/prebuilts/gcc/linux-x86/mips/mipsel-linux-android-4.4.3/sysroot/usr/include/asm/ |
tlb.h | 21 #define tlb_start_vma(tlb, vma) do { if (!tlb->fullmm) flush_cache_range(vma, vma->vm_start, vma->vm_end); } while (0) 22 #define tlb_end_vma(tlb, vma) do { } while (0)
|
tlbflush.h | 25 #define flush_tlb_range(vma, vmaddr, end) local_flush_tlb_range(vma, vmaddr, end) 27 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
|
/bionic/libc/kernel/arch-x86/asm/ |
cacheflush.h | 26 #define flush_cache_range(vma, start, end) do { } while (0) 27 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 34 #define flush_icache_page(vma,pg) do { } while (0) 35 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 39 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 40 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/development/ndk/platforms/android-9/arch-x86/include/asm/ |
cacheflush.h | 26 #define flush_cache_range(vma, start, end) do { } while (0) 27 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 34 #define flush_icache_page(vma,pg) do { } while (0) 35 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 39 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 40 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/prebuilts/ndk/android-ndk-r4/platforms/android-5/arch-x86/usr/include/asm/ |
cacheflush.h | 20 #define flush_cache_range(vma, start, end) do { } while (0) 21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 26 #define flush_icache_page(vma,pg) do { } while (0) 27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/prebuilts/ndk/android-ndk-r4/platforms/android-8/arch-x86/usr/include/asm/ |
cacheflush.h | 20 #define flush_cache_range(vma, start, end) do { } while (0) 21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 26 #define flush_icache_page(vma,pg) do { } while (0) 27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/prebuilts/ndk/android-ndk-r6/platforms/android-9/arch-x86/usr/include/asm/ |
cacheflush.h | 20 #define flush_cache_range(vma, start, end) do { } while (0) 21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 26 #define flush_icache_page(vma,pg) do { } while (0) 27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/prebuilts/ndk/android-ndk-r7/platforms/android-14/arch-x86/usr/include/asm/ |
cacheflush.h | 20 #define flush_cache_range(vma, start, end) do { } while (0) 21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 26 #define flush_icache_page(vma,pg) do { } while (0) 27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/prebuilts/ndk/android-ndk-r7/platforms/android-9/arch-x86/usr/include/asm/ |
cacheflush.h | 20 #define flush_cache_range(vma, start, end) do { } while (0) 21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 26 #define flush_icache_page(vma,pg) do { } while (0) 27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len) 32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
|
/external/oprofile/libopagent/ |
opagent.h | 38 unsigned long vma; member in struct:debug_line_info 76 * vma: The virtual memory address of the executable code. 79 * that given by the vma argument. For some JIT compilers, 90 uint64_t vma, void const * code, 117 * vma: The virtual memory address of the compiled code being unloaded. 118 * An op_write_native_code() with the same vma should have 124 int op_unload_native_code(op_agent_t hdl, uint64_t vma);
|
/external/oprofile/libpp/ |
symbol_functors.cpp | 26 if (lhs.sample.vma != rhs.sample.vma) 27 return lhs.sample.vma < rhs.sample.vma;
|
/external/kernel-headers/original/asm-x86/ |
cacheflush.h | 11 #define flush_cache_range(vma, start, end) do { } while (0) 12 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0) 17 #define flush_icache_page(vma,pg) do { } while (0) 18 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) 22 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 24 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
|
tlbflush_32.h | 78 * - flush_tlb_page(vma, vmaddr) flushes one page 79 * - flush_tlb_range(vma, start, end) flushes a range of pages 104 static inline void flush_tlb_page(struct vm_area_struct *vma, 107 if (vma->vm_mm == current->active_mm) 111 static inline void flush_tlb_range(struct vm_area_struct *vma, 114 if (vma->vm_mm == current->active_mm) 137 static inline void flush_tlb_range(struct vm_area_struct * vma, unsigned long start, unsigned long end) 139 flush_tlb_mm(vma->vm_mm);
|
/external/kernel-headers/original/linux/ |
mempolicy.h | 21 #define MPOL_F_ADDR (1<<1) /* look up vma using address */ 45 * A mempolicy can be either associated with a process or with a VMA. 46 * For VMA related allocations the VMA policy is preferred, otherwise 52 * its own state. All vma manipulation is somewhat protected by a down_read on 94 #define vma_policy(vma) ((vma)->vm_policy) 95 #define vma_set_policy(vma, pol) ((vma)->vm_policy = (pol)) 114 #define mpol_set_vma_default(vma) ((vma)->vm_policy = NULL [all...] |
/external/kernel-headers/original/asm-arm/ |
cacheflush.h | 128 * - flags - vma->vm_flags field 248 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ 251 flush_ptrace_access(vma, page, vaddr, dst, len, 1);\ 254 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ 271 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end) 273 if (cpu_isset(smp_processor_id(), vma->vm_mm->cpu_vm_mask)) 275 vma->vm_flags); 279 flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn) 281 if (cpu_isset(smp_processor_id(), vma->vm_mm->cpu_vm_mask)) { 283 __cpuc_flush_user_range(addr, addr + PAGE_SIZE, vma->vm_flags) [all...] |
/external/oprofile/opjitconv/ |
jitsymbol.c | 80 if (a0->vma < b0->vma) 82 if (a0->vma == b0->vma) 98 if (entries_address_ascending[i]->vma) 171 * writing a 0 address as magic vma and sorting 177 e->vma, e->symbol_name); 178 e->vma = 0; 245 unsigned long long start_addr_keep = keep->vma; 246 unsigned long long end_addr_keep = keep->vma + keep->code_size [all...] |