| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | arena_externs.h | 37 void arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, 40 void arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, 45 void arena_handle_deferred_work(tsdn_t *tsdn, arena_t *arena); 46 edata_t *arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena, 48 void arena_extent_dalloc_large_prep(tsdn_t *tsdn, arena_t *arena, 50 void arena_extent_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena, 52 void arena_extent_ralloc_large_expand(tsdn_t *tsdn, arena_t *arena, 54 bool arena_decay_ms_set(tsdn_t *tsdn, arena_t *arena, extent_state_t state, 56 ssize_t arena_decay_ms_get(arena_t *arena, extent_state_t state); 57 void arena_decay(tsdn_t *tsdn, arena_t *arena, bool is_background_thread, [all …]
|
| H A D | jemalloc_internal_inlines_b.h | 9 arena_t *oldarena = tsd_arena_get(tsd); in percpu_arena_update() 15 arena_t *newarena = arena_get(tsd_tsdn(tsd), newind, true); in percpu_arena_update() 31 static inline arena_t * 32 arena_choose_impl(tsd_t *tsd, arena_t *arena, bool internal) { in arena_choose_impl() 33 arena_t *ret; in arena_choose_impl() 86 static inline arena_t * 87 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose() 91 static inline arena_t * 92 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose() 97 arena_is_auto(arena_t *arena) { in arena_is_auto()
|
| H A D | jemalloc_internal_externs.h | 61 void arena_set(unsigned ind, arena_t *arena); 63 arena_t *arena_init(tsdn_t *tsdn, unsigned ind, const arena_config_t *config); 64 arena_t *arena_choose_hard(tsd_t *tsd, bool internal); 65 void arena_migrate(tsd_t *tsd, arena_t *oldarena, arena_t *newarena);
|
| H A D | arena_inlines_a.h | 5 arena_ind_get(const arena_t *arena) { in arena_ind_get() 10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add() 15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub() 20 arena_internal_get(arena_t *arena) { in arena_internal_get()
|
| H A D | arena_inlines_b.h | 14 static inline arena_t * 16 return (arena_t *)atomic_load_p(&arenas[edata_arena_ind_get(edata)], in arena_get_from_edata() 20 JEMALLOC_ALWAYS_INLINE arena_t * 21 arena_choose_maybe_huge(tsd_t *tsd, arena_t *arena, size_t size) { in arena_choose_maybe_huge() 32 arena_t *tsd_arena = tsd_arena_get(tsd); in arena_choose_maybe_huge() 118 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) { in arena_decay_ticks() 140 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) { in arena_decay_tick() 145 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, in arena_malloc() 165 JEMALLOC_ALWAYS_INLINE arena_t * 169 return (arena_t *)atomic_load_p(&arenas[arena_ind], ATOMIC_RELAXED); in arena_aalloc() [all …]
|
| H A D | jemalloc_internal_inlines_a.h | 59 static inline arena_t * 61 arena_t *ret; in arena_get() 65 ret = (arena_t *)atomic_load_p(&arenas[ind], ATOMIC_ACQUIRE); in arena_get() 111 pre_reentrancy(tsd_t *tsd, arena_t *arena) { in pre_reentrancy()
|
| H A D | tcache_externs.h | 38 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, 48 tcache_t *tcache, arena_t *arena); 51 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena); 57 tcache_t *tcache, arena_t *arena);
|
| H A D | large_externs.h | 6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
|
| H A D | jemalloc_internal_inlines_c.h | 29 JEMALLOC_ALWAYS_INLINE arena_t * 45 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm() 70 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm() 90 tcache_t *tcache, arena_t *arena) { in ipalloct() 138 size_t alignment, bool zero, tcache_t *tcache, arena_t *arena, in iralloct_realign() 177 bool zero, tcache_t *tcache, arena_t *arena, hook_ralloc_args_t *hook_args) in iralloct()
|
| H A D | tsd.h | 78 O(iarena, arena_t *, arena_t *) \ 79 O(arena, arena_t *, arena_t *) \
|
| H A D | arena_types.h | 13 typedef struct arena_s arena_t; typedef
|
| H A D | tcache_structs.h | 32 arena_t *arena;
|
| H A D | extent_dss.h | 20 void *extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr,
|
| H A D | tcache_inlines.h | 45 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small() 88 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large()
|
| H A D | background_thread_inlines.h | 16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get()
|
| /src/contrib/jemalloc/src/ |
| H A D | arena.c | 63 static bool arena_decay_dirty(tsdn_t *tsdn, arena_t *arena, 65 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, edata_t *slab, 68 arena_maybe_do_deferred_work(tsdn_t *tsdn, arena_t *arena, decay_t *decay, 74 arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_basic_stats_merge() 85 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_stats_merge() 200 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check() 216 void arena_handle_deferred_work(tsdn_t *tsdn, arena_t *arena) { in arena_handle_deferred_work() 291 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update() 307 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_dalloc_stats_update() 323 arena_large_ralloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t oldusize, in arena_large_ralloc_stats_update() [all …]
|
| H A D | large.c | 14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc() 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc() 56 arena_t *arena = arena_get_from_edata(edata); in large_ralloc_no_move_shrink() 85 arena_t *arena = arena_get_from_edata(edata); in large_ralloc_no_move_expand() 175 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, in large_ralloc_move_helper() 184 large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize, in large_ralloc() 231 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata, in large_dalloc_prep_impl() 251 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata) { in large_dalloc_finish_impl() 271 arena_t *arena = arena_get_from_edata(edata); in large_dalloc()
|
| H A D | tcache.c | 229 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, in tcache_alloc_small_hard() 323 arena_t *tcache_arena = tcache_slow->arena; in tcache_bin_flush_impl() 351 arena_t *cur_arena = arena_get(tsdn, cur_arena_ind, false); in tcache_bin_flush_impl() 582 tcache_t *tcache, arena_t *arena) { in tcache_arena_associate() 604 arena_t *arena = tcache_slow->arena; in tcache_arena_dissociate() 631 tcache_t *tcache, arena_t *arena) { in tcache_arena_reassociate() 733 arena_t *arena; in tsd_tcache_data_init() 810 arena_t *arena = tcache_slow->arena; in tcache_destroy() 863 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_stats_merge()
|
| H A D | inspect.c | 55 arena_t *arena = (arena_t *)atomic_load_p( in inspect_extent_util_stats_verbose_get()
|
| H A D | jemalloc.c | 183 static arena_t *a0; /* arenas[0]. */ 376 arena_set(unsigned ind, arena_t *arena) { in arena_set() 396 static arena_t * 398 arena_t *arena; in arena_init_locked() 442 arena_t * 444 arena_t *arena; in arena_init() 457 arena_t *arena = arena_get(tsd_tsdn(tsd), ind, false); in arena_bind() 476 arena_migrate(tsd_t *tsd, arena_t *oldarena, arena_t *newarena) { in arena_migrate() 493 arena_t *arena; in arena_unbind() 506 arena_t * [all …]
|
| H A D | ehooks.c | 22 extent_alloc_core(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_core() 54 arena_t *arena = arena_get(tsdn, arena_ind, false); in ehooks_default_alloc_impl()
|
| H A D | ctl.c | 1055 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_t *ctl_arena, arena_t *arena) { in ctl_arena_stats_amerge() 1251 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, ctl_arena_t *ctl_sdarena, in ctl_arena_refresh() 1308 VARIABLE_ARRAY(arena_t *, tarenas, ctl_arenas->narenas); in ctl_refresh() 2228 arena_t *oldarena; in CTL_RO_CONFIG_GEN() 2240 arena_t *newarena; in CTL_RO_CONFIG_GEN() 2442 arena_t *arena = arena_choose(tsd, NULL); in thread_idle_ctl() 2545 VARIABLE_ARRAY(arena_t *, tarenas, narenas); in arena_i_decay() 2564 arena_t *tarena; in arena_i_decay() 2613 arena_t **arena) { in arena_i_reset_destroy_helper() 2666 arena_t *arena; in arena_i_reset_ctl() [all …]
|
| H A D | background_thread.c | 195 arena_t *arena = arena_get(tsdn, i, false); 564 arena_t *arena = arena_get(tsd_tsdn(tsd), i, false); 586 arena_t *arena = arena_get(tsd_tsdn(tsd), i, false);
|
| H A D | extent_dss.c | 109 extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr, size_t size, in extent_alloc_dss()
|
| /src/contrib/jemalloc/ |
| H A D | ChangeLog | 676 - Fix over-sized allocation of arena_t (plus associated stats) data
|