Home
last modified time | relevance | path

Searched refs:arena_t (Results 1 – 25 of 25) sorted by relevance

/src/contrib/jemalloc/include/jemalloc/internal/
H A Darena_externs.h37 void arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena,
40 void arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads,
45 void arena_handle_deferred_work(tsdn_t *tsdn, arena_t *arena);
46 edata_t *arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena,
48 void arena_extent_dalloc_large_prep(tsdn_t *tsdn, arena_t *arena,
50 void arena_extent_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena,
52 void arena_extent_ralloc_large_expand(tsdn_t *tsdn, arena_t *arena,
54 bool arena_decay_ms_set(tsdn_t *tsdn, arena_t *arena, extent_state_t state,
56 ssize_t arena_decay_ms_get(arena_t *arena, extent_state_t state);
57 void arena_decay(tsdn_t *tsdn, arena_t *arena, bool is_background_thread,
[all …]
H A Djemalloc_internal_inlines_b.h9 arena_t *oldarena = tsd_arena_get(tsd); in percpu_arena_update()
15 arena_t *newarena = arena_get(tsd_tsdn(tsd), newind, true); in percpu_arena_update()
31 static inline arena_t *
32 arena_choose_impl(tsd_t *tsd, arena_t *arena, bool internal) { in arena_choose_impl()
33 arena_t *ret; in arena_choose_impl()
86 static inline arena_t *
87 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose()
91 static inline arena_t *
92 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose()
97 arena_is_auto(arena_t *arena) { in arena_is_auto()
H A Djemalloc_internal_externs.h61 void arena_set(unsigned ind, arena_t *arena);
63 arena_t *arena_init(tsdn_t *tsdn, unsigned ind, const arena_config_t *config);
64 arena_t *arena_choose_hard(tsd_t *tsd, bool internal);
65 void arena_migrate(tsd_t *tsd, arena_t *oldarena, arena_t *newarena);
H A Darena_inlines_a.h5 arena_ind_get(const arena_t *arena) { in arena_ind_get()
10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add()
15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub()
20 arena_internal_get(arena_t *arena) { in arena_internal_get()
H A Darena_inlines_b.h14 static inline arena_t *
16 return (arena_t *)atomic_load_p(&arenas[edata_arena_ind_get(edata)], in arena_get_from_edata()
20 JEMALLOC_ALWAYS_INLINE arena_t *
21 arena_choose_maybe_huge(tsd_t *tsd, arena_t *arena, size_t size) { in arena_choose_maybe_huge()
32 arena_t *tsd_arena = tsd_arena_get(tsd); in arena_choose_maybe_huge()
118 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) { in arena_decay_ticks()
140 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) { in arena_decay_tick()
145 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, in arena_malloc()
165 JEMALLOC_ALWAYS_INLINE arena_t *
169 return (arena_t *)atomic_load_p(&arenas[arena_ind], ATOMIC_RELAXED); in arena_aalloc()
[all …]
H A Djemalloc_internal_inlines_a.h59 static inline arena_t *
61 arena_t *ret; in arena_get()
65 ret = (arena_t *)atomic_load_p(&arenas[ind], ATOMIC_ACQUIRE); in arena_get()
111 pre_reentrancy(tsd_t *tsd, arena_t *arena) { in pre_reentrancy()
H A Dtcache_externs.h38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
48 tcache_t *tcache, arena_t *arena);
51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
57 tcache_t *tcache, arena_t *arena);
H A Dlarge_externs.h6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
H A Djemalloc_internal_inlines_c.h29 JEMALLOC_ALWAYS_INLINE arena_t *
45 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm()
70 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm()
90 tcache_t *tcache, arena_t *arena) { in ipalloct()
138 size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, in iralloct_realign()
177 bool zero, tcache_t *tcache, arena_t *arena, hook_ralloc_args_t *hook_args) in iralloct()
H A Dtsd.h78 O(iarena, arena_t *, arena_t *) \
79 O(arena, arena_t *, arena_t *) \
H A Darena_types.h13 typedef struct arena_s arena_t; typedef
H A Dtcache_structs.h32 arena_t *arena;
H A Dextent_dss.h20 void *extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr,
H A Dtcache_inlines.h45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small()
88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large()
H A Dbackground_thread_inlines.h16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get()
/src/contrib/jemalloc/src/
H A Darena.c63 static bool arena_decay_dirty(tsdn_t *tsdn, arena_t *arena,
65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab,
68 arena_maybe_do_deferred_work(tsdn_t *tsdn, arena_t *arena, decay_t *decay,
74 arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_basic_stats_merge()
85 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_stats_merge()
200 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check()
216 void arena_handle_deferred_work(tsdn_t *tsdn, arena_t *arena) { in arena_handle_deferred_work()
291 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update()
307 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_dalloc_stats_update()
323 arena_large_ralloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t oldusize, in arena_large_ralloc_stats_update()
[all …]
H A Dlarge.c14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc()
21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc()
56 arena_t *arena = arena_get_from_edata(edata); in large_ralloc_no_move_shrink()
85 arena_t *arena = arena_get_from_edata(edata); in large_ralloc_no_move_expand()
175 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, in large_ralloc_move_helper()
184 large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize, in large_ralloc()
231 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata, in large_dalloc_prep_impl()
251 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata) { in large_dalloc_finish_impl()
271 arena_t *arena = arena_get_from_edata(edata); in large_dalloc()
H A Dtcache.c229 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, in tcache_alloc_small_hard()
323 arena_t *tcache_arena = tcache_slow->arena; in tcache_bin_flush_impl()
351 arena_t *cur_arena = arena_get(tsdn, cur_arena_ind, false); in tcache_bin_flush_impl()
582 tcache_t *tcache, arena_t *arena) { in tcache_arena_associate()
604 arena_t *arena = tcache_slow->arena; in tcache_arena_dissociate()
631 tcache_t *tcache, arena_t *arena) { in tcache_arena_reassociate()
733 arena_t *arena; in tsd_tcache_data_init()
810 arena_t *arena = tcache_slow->arena; in tcache_destroy()
863 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_stats_merge()
H A Dinspect.c55 arena_t *arena = (arena_t *)atomic_load_p( in inspect_extent_util_stats_verbose_get()
H A Djemalloc.c183 static arena_t *a0; /* arenas[0]. */
376 arena_set(unsigned ind, arena_t *arena) { in arena_set()
396 static arena_t *
398 arena_t *arena; in arena_init_locked()
442 arena_t *
444 arena_t *arena; in arena_init()
457 arena_t *arena = arena_get(tsd_tsdn(tsd), ind, false); in arena_bind()
476 arena_migrate(tsd_t *tsd, arena_t *oldarena, arena_t *newarena) { in arena_migrate()
493 arena_t *arena; in arena_unbind()
506 arena_t *
[all …]
H A Dehooks.c22 extent_alloc_core(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_core()
54 arena_t *arena = arena_get(tsdn, arena_ind, false); in ehooks_default_alloc_impl()
H A Dctl.c1055 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_t *ctl_arena, arena_t *arena) { in ctl_arena_stats_amerge()
1251 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, ctl_arena_t *ctl_sdarena, in ctl_arena_refresh()
1308 VARIABLE_ARRAY(arena_t *, tarenas, ctl_arenas->narenas); in ctl_refresh()
2228 arena_t *oldarena; in CTL_RO_CONFIG_GEN()
2240 arena_t *newarena; in CTL_RO_CONFIG_GEN()
2442 arena_t *arena = arena_choose(tsd, NULL); in thread_idle_ctl()
2545 VARIABLE_ARRAY(arena_t *, tarenas, narenas); in arena_i_decay()
2564 arena_t *tarena; in arena_i_decay()
2613 arena_t **arena) { in arena_i_reset_destroy_helper()
2666 arena_t *arena; in arena_i_reset_ctl()
[all …]
H A Dbackground_thread.c195 arena_t *arena = arena_get(tsdn, i, false);
564 arena_t *arena = arena_get(tsd_tsdn(tsd), i, false);
586 arena_t *arena = arena_get(tsd_tsdn(tsd), i, false);
H A Dextent_dss.c109 extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_dss()
/src/contrib/jemalloc/
H A DChangeLog676 - Fix over-sized allocation of arena_t (plus associated stats) data