HomeSort by relevance Sort by last modified time
    Searched refs:tcache (Results 1 - 16 of 16) sorted by null

  /external/jemalloc/src/
tcache.c 11 static unsigned stack_nelms; /* Total stack elms per tcache. */
37 tcache_event_hard(tsd_t *tsd, tcache_t *tcache)
39 szind_t binind = tcache->next_gc_bin;
40 tcache_bin_t *tbin = &tcache->tbins[binind];
48 tcache_bin_flush_small(tsd, tcache, tbin, binind,
53 - tbin->low_water + (tbin->low_water >> 2), tcache);
71 tcache->next_gc_bin++;
72 if (tcache->next_gc_bin == nhbins)
73 tcache->next_gc_bin = 0;
77 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
323 tcache_t *tcache; local
367 tcache_arena_dissociate(tsd_tsdn(tsd), tcache, arena); local
398 idalloctm(tsd_tsdn(tsd), tcache, NULL, true, true); local
404 tcache_t *tcache; local
480 tcache_t *tcache; local
    [all...]
jemalloc.c 758 * individual threads do not lock when recording tcache stats
766 tcache_t *tcache; local
775 ql_foreach(tcache, &arena->tcache_ql, link) {
776 tcache_stats_merge(tsdn, tcache, arena);
1259 CONF_HANDLE_BOOL(opt_tcache, "tcache",
1261 if (CONF_MATCH("tcache")) {
1266 "tcache cannot be enabled "
2195 tcache_t *tcache; local
2228 tcache_t *tcache; local
2382 tcache_t *tcache; local
2611 tcache_t *tcache; local
2653 tcache_t *tcache; local
    [all...]
huge.c 380 size_t usize, size_t alignment, bool zero, tcache_t *tcache)
405 isqalloc(tsd, ptr, oldsize, tcache, true);
arena.c     [all...]
stats.c 52 CTL_GET("config.tcache", &config_tcache, bool);
623 CONFIG_WRITE_BOOL_JSON(tcache, ",")
742 OPT_WRITE_BOOL(tcache, ",")
    [all...]
ctl.c 245 {NAME("tcache"), CHILD(named, thread_tcache)},
260 {NAME("tcache"), CTL(config_tcache)},
282 {NAME("tcache"), CTL(opt_tcache)},
500 {NAME("tcache"), CHILD(named, tcache)},
1342 tcache_t *tcache = tsd_tcache_get(tsd); local
1344 tcache_arena_reassociate(tsd_tsdn(tsd), tcache, local
    [all...]
  /external/jemalloc/include/jemalloc/internal/
tcache.h 10 * tcache pointers close to NULL are used to encode state information that is
58 /* Number of tcache allocation/deallocation events between incremental GCs. */
103 * array. During tcache initialization, the avail pointer in each
109 /* Linkage for list of available (previously used) explicit tcache IDs. */
112 tcache_t *tcache; member in union:tcaches_s::__anon22494
127 * Number of tcache bins. There are NBINS small-object bins, plus 0 or more
136 * Explicit tcaches, managed via the tcache.{create,flush,destroy} mallctls and
141 * explicit tcache is created without a disproportionate impact on memory usage.
146 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache);
147 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
240 tcache_t *tcache; local
    [all...]
jemalloc_internal.h 173 * jemalloc can conceptually be broken into components (arena, tcache, etc.),
200 * t: tcache
392 #include "jemalloc/internal/tcache.h"
429 #include "jemalloc/internal/tcache.h"
527 #include "jemalloc/internal/tcache.h"
977 * Include portions of arena.h interleaved with tcache.h in order to resolve
983 #include "jemalloc/internal/tcache.h"
994 tcache_t *tcache, bool is_metadata, arena_t *arena, bool slow_path);
998 tcache_t *tcache, bool is_metadata, arena_t *arena);
1000 tcache_t *tcache, arena_t *arena)
1181 idalloctm(tsd_tsdn(tsd), ptr, tcache, false, slow_path); local
1199 isdalloct(tsd_tsdn(tsd), ptr, size, tcache, slow_path); local
    [all...]
huge.h 18 size_t usize, size_t alignment, bool zero, tcache_t *tcache);
arena.h 580 size_t alignment, bool zero, tcache_t *tcache);
605 size_t size, size_t alignment, bool zero, tcache_t *tcache);
701 bool zero, tcache_t *tcache, bool slow_path);
704 void arena_dalloc(tsdn_t *tsdn, void *ptr, tcache_t *tcache, bool slow_path);
705 void arena_sdalloc(tsdn_t *tsdn, void *ptr, size_t size, tcache_t *tcache,
    [all...]
tsd.h 593 O(tcache, tcache_t *) \
  /external/strace/tests/
getcpu.c 53 long *tcache = tail_alloc(128); local
62 res = syscall(__NR_getcpu, cpu, node, tcache);
66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
  /external/strace/tests-m32/
getcpu.c 53 long *tcache = tail_alloc(128); local
62 res = syscall(__NR_getcpu, cpu, node, tcache);
66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
  /external/strace/tests-mx32/
getcpu.c 53 long *tcache = tail_alloc(128); local
62 res = syscall(__NR_getcpu, cpu, node, tcache);
66 printf("getcpu([%u], [%u], %p) = 0\n", *cpu, *node, tcache);
  /external/jemalloc/test/unit/
mallctl.c 144 TEST_MALLCTL_CONFIG(tcache, bool);
183 TEST_MALLCTL_OPT(bool, tcache, tcache);
184 TEST_MALLCTL_OPT(size_t, lg_tcache_max, tcache);
236 /* Deallocate p and q, but bypass the tcache for q. */
240 /* Make sure that tcache-based allocation returns p, not q. */
243 assert_ptr_eq(p0, p1, "Expected tcache to allocate cached region");
267 assert_d_eq(mallctl("tcache.create", (void *)&tis[i], &sz, NULL,
271 /* Exercise tcache ID recycling. */
273 assert_d_eq(mallctl("tcache.destroy", NULL, NULL
    [all...]
  /external/valgrind/coregrind/m_syswrap/
syswrap-linux.c     [all...]

Completed in 357 milliseconds