Home | History | Annotate | Download | only in src

Lines Matching refs:usize

689 arena_huge_malloc_stats_update(arena_t *arena, size_t usize)
691 index_t index = size2index(usize) - nlclasses - NBINS;
696 arena->stats.allocated_huge += usize;
702 arena_huge_malloc_stats_update_undo(arena_t *arena, size_t usize)
704 index_t index = size2index(usize) - nlclasses - NBINS;
709 arena->stats.allocated_huge -= usize;
715 arena_huge_dalloc_stats_update(arena_t *arena, size_t usize)
717 index_t index = size2index(usize) - nlclasses - NBINS;
722 arena->stats.allocated_huge -= usize;
728 arena_huge_dalloc_stats_update_undo(arena_t *arena, size_t usize)
730 index_t index = size2index(usize) - nlclasses - NBINS;
735 arena->stats.allocated_huge += usize;
741 arena_huge_ralloc_stats_update(arena_t *arena, size_t oldsize, size_t usize)
745 arena_huge_malloc_stats_update(arena, usize);
750 size_t usize)
754 arena_huge_malloc_stats_update_undo(arena, usize);
785 size_t usize, size_t alignment, bool *zero, size_t csize)
795 arena_huge_malloc_stats_update_undo(arena, usize);
796 arena->stats.mapped -= usize;
798 arena->nactive -= (usize >> LG_PAGE);
806 arena_chunk_alloc_huge(arena_t *arena, size_t usize, size_t alignment,
811 size_t csize = CHUNK_CEILING(usize);
817 arena_huge_malloc_stats_update(arena, usize);
818 arena->stats.mapped += usize;
820 arena->nactive += (usize >> LG_PAGE);
830 ret = arena_chunk_alloc_huge_hard(arena, chunk_alloc, usize,
835 stats_cactive_add(usize);
840 arena_chunk_dalloc_huge(arena_t *arena, void *chunk, size_t usize)
845 csize = CHUNK_CEILING(usize);
849 arena_huge_dalloc_stats_update(arena, usize);
850 arena->stats.mapped -= usize;
851 stats_cactive_sub(usize);
853 arena->nactive -= (usize >> LG_PAGE);
866 size_t usize)
869 assert(CHUNK_CEILING(oldsize) == CHUNK_CEILING(usize));
870 assert(oldsize != usize);
874 arena_huge_ralloc_stats_update(arena, oldsize, usize);
875 if (oldsize < usize) {
876 size_t udiff = usize - oldsize;
881 size_t udiff = oldsize - usize;
891 size_t usize)
893 size_t udiff = oldsize - usize;
894 size_t cdiff = CHUNK_CEILING(oldsize) - CHUNK_CEILING(usize);
898 arena_huge_ralloc_stats_update(arena, oldsize, usize);
909 CHUNK_CEILING(usize));
925 size_t oldsize, size_t usize, bool *zero, void *nchunk, size_t udiff,
937 usize);
948 size_t usize, bool *zero)
953 size_t udiff = usize - oldsize;
954 size_t cdiff = CHUNK_CEILING(usize) - CHUNK_CEILING(oldsize);
960 arena_huge_ralloc_stats_update(arena, oldsize, usize);
974 oldsize, usize, zero, nchunk, udiff, cdiff);
1875 arena_redzone_corruption(void *ptr, size_t usize, bool after,
1881 after ? "after" : "before", ptr, usize, byte);
1944 arena_quarantine_junk_small(void *ptr, size_t usize)
1951 assert(usize <= SMALL_MAXCLASS);
1953 binind = size2index(usize);
2016 size_t usize;
2024 usize = s2u(size);
2037 run = arena_run_alloc_large(arena, usize + large_pad, zero);
2046 index_t index = size2index(usize) - NBINS;
2050 arena->stats.allocated_large += usize;
2056 idump = arena_prof_accum_locked(arena, usize);
2064 memset(ret, 0xa5, usize);
2066 memset(ret, 0, usize);
2149 arena_palloc(tsd_t *tsd, arena_t *arena, size_t usize, size_t alignment,
2154 if (usize <= SMALL_MAXCLASS && (alignment < PAGE || (alignment == PAGE
2155 && (usize & PAGE_MASK) == 0))) {
2157 ret = arena_malloc(tsd, arena, usize, zero, tcache);
2158 } else if (usize <= arena_maxclass && alignment <= PAGE) {
2165 ret = arena_malloc(tsd, arena, usize, zero, tcache);
2169 if (likely(usize <= arena_maxclass)) {
2170 ret = arena_palloc_large(tsd, arena, usize, alignment,
2173 ret = huge_malloc(tsd, arena, usize, zero, tcache);
2175 ret = huge_palloc(tsd, arena, usize, alignment, zero,
2348 arena_dalloc_junk_large(void *ptr, size_t usize)
2352 memset(ptr, 0x5a, usize);
2370 size_t usize = arena_mapbits_large_size_get(chunk, pageind) -
2374 arena_dalloc_junk_large(ptr, usize);
2376 index_t index = size2index(usize) - NBINS;
2379 arena->stats.allocated_large -= usize;
2466 size_t flag_dirty, splitsize, usize;
2468 usize = s2u(size + extra);
2469 while (oldsize + followsize < usize)
2470 usize = index2size(size2index(usize)-1);
2471 assert(usize >= usize_min);
2472 splitsize = usize - oldsize + large_pad;
2522 arena_ralloc_junk_large(void *ptr, size_t old_usize, size_t usize)
2526 memset((void *)((uintptr_t)ptr + usize), 0x5a,
2527 old_usize - usize);
2545 size_t usize;
2551 usize = s2u(size + extra);
2552 if (usize == oldsize) {
2562 if (usize < oldsize) {
2564 arena_ralloc_junk_large(ptr, oldsize, usize);
2566 usize);
2642 size_t usize = sa2u(size + extra, alignment);
2643 if (usize == 0)
2645 ret = ipalloct(tsd, usize, alignment, zero, tcache,
2657 size_t usize = sa2u(size, alignment);
2658 if (usize == 0)
2660 ret = ipalloct(tsd, usize, alignment, zero,