HomeSort by relevance Sort by last modified time
    Searched refs:vma (Results 1 - 25 of 135) sorted by null

1 2 3 4 5 6

  /bionic/libc/kernel/arch-sh/asm/
tlb.h 17 #define tlb_start_vma(tlb, vma) flush_cache_range(vma, vma->vm_start, vma->vm_end)
19 #define tlb_end_vma(tlb, vma) flush_tlb_range(vma, vma->vm_start, vma->vm_end)
tlbflush.h 17 #define flush_tlb_page(vma, page) local_flush_tlb_page(vma, page)
20 #define flush_tlb_range(vma, start, end) local_flush_tlb_range(vma, start, end)
  /bionic/libc/kernel/arch-x86/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /development/ndk/platforms/android-9/arch-x86/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /prebuilt/ndk/android-ndk-r4/platforms/android-5/arch-x86/usr/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /prebuilt/ndk/android-ndk-r4/platforms/android-8/arch-x86/usr/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /prebuilt/ndk/android-ndk-r6/platforms/android-9/arch-x86/usr/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /prebuilt/ndk/android-ndk-r7/platforms/android-14/arch-x86/usr/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /prebuilt/ndk/android-ndk-r7/platforms/android-9/arch-x86/usr/include/asm/
cacheflush.h 20 #define flush_cache_range(vma, start, end) do { } while (0)
21 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
26 #define flush_icache_page(vma,pg) do { } while (0)
27 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
31 #define copy_to_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
32 #define copy_from_user_page(vma, page, vaddr, dst, src, len) memcpy(dst, src, len)
pgtable_32.h 144 #define ptep_set_access_flags(vma, address, ptep, entry, dirty) ({ int __changed = !pte_same(*(ptep), entry); if (__changed && dirty) { (ptep)->pte_low = (entry).pte_low; pte_update_defer((vma)->vm_mm, (address), (ptep)); flush_tlb_page(vma, address); } __changed; })
146 #define ptep_test_and_clear_young(vma, addr, ptep) ({ int __ret = 0; if (pte_young(*(ptep))) __ret = test_and_clear_bit(_PAGE_BIT_ACCESSED, &(ptep)->pte_low); if (__ret) pte_update((vma)->vm_mm, addr, ptep); __ret; })
148 #define ptep_clear_flush_young(vma, address, ptep) ({ int __young; __young = ptep_test_and_clear_young((vma), (address), (ptep)); if (__young) flush_tlb_page(vma, address); __young; })
170 #define update_mmu_cache(vma,address,pte) do { } while (0)
173 #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) remap_pfn_range(vma, vaddr, pfn, size, prot
    [all...]
  /external/oprofile/libopagent/
opagent.h 38 unsigned long vma; member in struct:debug_line_info
76 * vma: The virtual memory address of the executable code.
79 * that given by the vma argument. For some JIT compilers,
90 uint64_t vma, void const * code,
117 * vma: The virtual memory address of the compiled code being unloaded.
118 * An op_write_native_code() with the same vma should have
124 int op_unload_native_code(op_agent_t hdl, uint64_t vma);
  /external/oprofile/libpp/
symbol_functors.cpp 26 if (lhs.sample.vma != rhs.sample.vma)
27 return lhs.sample.vma < rhs.sample.vma;
symbol_sort.h 29 vma, enumerator in enum:sort_options::sort_order
  /external/kernel-headers/original/asm-x86/
cacheflush.h 11 #define flush_cache_range(vma, start, end) do { } while (0)
12 #define flush_cache_page(vma, vmaddr, pfn) do { } while (0)
17 #define flush_icache_page(vma,pg) do { } while (0)
18 #define flush_icache_user_range(vma,pg,adr,len) do { } while (0)
22 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
24 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
tlbflush_32.h 78 * - flush_tlb_page(vma, vmaddr) flushes one page
79 * - flush_tlb_range(vma, start, end) flushes a range of pages
104 static inline void flush_tlb_page(struct vm_area_struct *vma,
107 if (vma->vm_mm == current->active_mm)
111 static inline void flush_tlb_range(struct vm_area_struct *vma,
114 if (vma->vm_mm == current->active_mm)
137 static inline void flush_tlb_range(struct vm_area_struct * vma, unsigned long start, unsigned long end)
139 flush_tlb_mm(vma->vm_mm);
  /external/kernel-headers/original/linux/
mempolicy.h 21 #define MPOL_F_ADDR (1<<1) /* look up vma using address */
45 * A mempolicy can be either associated with a process or with a VMA.
46 * For VMA related allocations the VMA policy is preferred, otherwise
52 * its own state. All vma manipulation is somewhat protected by a down_read on
94 #define vma_policy(vma) ((vma)->vm_policy)
95 #define vma_set_policy(vma, pol) ((vma)->vm_policy = (pol))
114 #define mpol_set_vma_default(vma) ((vma)->vm_policy = NULL
    [all...]
  /external/kernel-headers/original/asm-arm/
cacheflush.h 128 * - flags - vma->vm_flags field
248 #define copy_to_user_page(vma, page, vaddr, dst, src, len) \
251 flush_ptrace_access(vma, page, vaddr, dst, len, 1);\
254 #define copy_from_user_page(vma, page, vaddr, dst, src, len) \
271 flush_cache_range(struct vm_area_struct *vma, unsigned long start, unsigned long end)
273 if (cpu_isset(smp_processor_id(), vma->vm_mm->cpu_vm_mask))
275 vma->vm_flags);
279 flush_cache_page(struct vm_area_struct *vma, unsigned long user_addr, unsigned long pfn)
281 if (cpu_isset(smp_processor_id(), vma->vm_mm->cpu_vm_mask)) {
283 __cpuc_flush_user_range(addr, addr + PAGE_SIZE, vma->vm_flags)
    [all...]
  /external/oprofile/opjitconv/
jitsymbol.c 80 if (a0->vma < b0->vma)
82 if (a0->vma == b0->vma)
98 if (entries_address_ascending[i]->vma)
171 * writing a 0 address as magic vma and sorting
177 e->vma, e->symbol_name);
178 e->vma = 0;
245 unsigned long long start_addr_keep = keep->vma;
246 unsigned long long end_addr_keep = keep->vma + keep->code_size
    [all...]
create_bfd.c 58 s->value = e->vma - section->vma;
76 size_t size, bfd_vma vma, flagword flags)
86 if (bfd_set_section_vma(abfd, section, vma) == FALSE) {
113 entries_address_ascending[start_idx]->vma;
115 unsigned long long vma_end = ee->vma + ee->code_size;
152 entries_address_ascending[start_idx]->vma;
159 " vma = %llx, offset = %llx,"
163 e->vma, e->vma - vma_start
    [all...]

Completed in 868 milliseconds

1 2 3 4 5 6