/external/jemalloc/src/ |
tcache.c | 11 static unsigned stack_nelms; /* Total stack elms per tcache. */ 37 tcache_event_hard(tsd_t *tsd, tcache_t *tcache) 39 szind_t binind = tcache->next_gc_bin; 40 tcache_bin_t *tbin = &tcache->tbins[binind]; 48 tcache_bin_flush_small(tsd, tcache, tbin, binind, 53 - tbin->low_water + (tbin->low_water >> 2), tcache); 71 tcache->next_gc_bin++; 72 if (tcache->next_gc_bin == nhbins) 73 tcache->next_gc_bin = 0; 77 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 323 tcache_t *tcache; local 367 tcache_arena_dissociate(tsd_tsdn(tsd), tcache, arena); local 398 idalloctm(tsd_tsdn(tsd), tcache, NULL, true, true); local 404 tcache_t *tcache; local 480 tcache_t *tcache; local [all...] |
jemalloc.c | 758 * individual threads do not lock when recording tcache stats 766 tcache_t *tcache; local 775 ql_foreach(tcache, &arena->tcache_ql, link) { 776 tcache_stats_merge(tsdn, tcache, arena); 1259 CONF_HANDLE_BOOL(opt_tcache, "tcache", 1261 if (CONF_MATCH("tcache")) { 1266 "tcache cannot be enabled " 2195 tcache_t *tcache; local 2228 tcache_t *tcache; local 2382 tcache_t *tcache; local 2611 tcache_t *tcache; local 2653 tcache_t *tcache; local [all...] |
huge.c | 380 size_t usize, size_t alignment, bool zero, tcache_t *tcache) 405 isqalloc(tsd, ptr, oldsize, tcache, true);
|
arena.c | [all...] |
stats.c | 52 CTL_GET("config.tcache", &config_tcache, bool); 623 CONFIG_WRITE_BOOL_JSON(tcache, ",") 742 OPT_WRITE_BOOL(tcache, ",") [all...] |
ctl.c | 245 {NAME("tcache"), CHILD(named, thread_tcache)}, 260 {NAME("tcache"), CTL(config_tcache)}, 282 {NAME("tcache"), CTL(opt_tcache)}, 500 {NAME("tcache"), CHILD(named, tcache)}, 1342 tcache_t *tcache = tsd_tcache_get(tsd); local 1344 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local [all...] |
/external/jemalloc/include/jemalloc/internal/ |
tcache.h | 10 * tcache pointers close to NULL are used to encode state information that is 58 /* Number of tcache allocation/deallocation events between incremental GCs. */ 103 * array. During tcache initialization, the avail pointer in each 109 /* Linkage for list of available (previously used) explicit tcache IDs. */ 112 tcache_t *tcache; member in union:tcaches_s::__anon22494 127 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more 136 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and 141 * explicit tcache is created without a disproportionate impact on memory usage. 146 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache); 147 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 240 tcache_t *tcache; local [all...] |
jemalloc_internal.h | 173 * jemalloc can conceptually be broken into components (arena, tcache, etc.), 200 * t: tcache 392 #include "jemalloc/internal/tcache.h" 429 #include "jemalloc/internal/tcache.h" 527 #include "jemalloc/internal/tcache.h" 977 * Include portions of arena.h interleaved with tcache.h in order to resolve 983 #include "jemalloc/internal/tcache.h" 994 tcache_t *tcache, bool is_metadata, arena_t *arena, bool slow_path); 998 tcache_t *tcache, bool is_metadata, arena_t *arena); 1000 tcache_t *tcache, arena_t *arena) 1181 idalloctm(tsd_tsdn(tsd), ptr, tcache, false, slow_path); local 1199 isdalloct(tsd_tsdn(tsd), ptr, size, tcache, slow_path); local [all...] |
huge.h | 18 size_t usize, size_t alignment, bool zero, tcache_t *tcache);
|
arena.h | 580 size_t alignment, bool zero, tcache_t *tcache); 605 size_t size, size_t alignment, bool zero, tcache_t *tcache); 701 bool zero, tcache_t *tcache, bool slow_path); 704 void arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache, bool slow_path); 705 void arena_sdalloc(tsdn_t *tsdn, void *ptr, size_t size, tcache_t *tcache, [all...] |
tsd.h | 593 O(tcache, tcache_t *) \
|
/external/strace/tests/ |
getcpu.c | 53 long *tcache = tail_alloc(128); local 62 res = syscall(__NR_getcpu, cpu, node, tcache); 66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
|
/external/strace/tests-m32/ |
getcpu.c | 53 long *tcache = tail_alloc(128); local 62 res = syscall(__NR_getcpu, cpu, node, tcache); 66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
|
/external/strace/tests-mx32/ |
getcpu.c | 53 long *tcache = tail_alloc(128); local 62 res = syscall(__NR_getcpu, cpu, node, tcache); 66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
|
/external/jemalloc/test/unit/ |
mallctl.c | 144 TEST_MALLCTL_CONFIG(tcache, bool); 183 TEST_MALLCTL_OPT(bool, tcache, tcache); 184 TEST_MALLCTL_OPT(size_t, lg_tcache_max, tcache); 236 /* Deallocate p and q, but bypass the tcache for q. */ 240 /* Make sure that tcache-based allocation returns p, not q. */ 243 assert_ptr_eq(p0, p1, "Expected tcache to allocate cached region"); 267 assert_d_eq(mallctl("tcache.create", (void *)&tis[i], &sz, NULL, 271 /* Exercise tcache ID recycling. */ 273 assert_d_eq(mallctl("tcache.destroy", NULL, NULL [all...] |
/external/valgrind/coregrind/m_syswrap/ |
syswrap-linux.c | [all...] |