| /src/contrib/jemalloc/src/ |
| H A D | bin_info.c | 6 bin_info_t bin_infos[SC_NBINS]; 9 bin_infos_init(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS], in bin_infos_init() argument 10 bin_info_t infos[SC_NBINS]) { in bin_infos_init() argument 11 for (unsigned i = 0; i < SC_NBINS; i++) { in bin_infos_init() 27 bin_info_boot(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]) { in bin_info_boot() argument
|
| H A D | bin.c | 10 bin_update_shard_size(unsigned bin_shard_sizes[SC_NBINS], size_t start_size, in bin_update_shard_size() argument 34 bin_shard_sizes_boot(unsigned bin_shard_sizes[SC_NBINS]) { in bin_shard_sizes_boot() argument 36 for (unsigned i = 0; i < SC_NBINS; i++) { in bin_shard_sizes_boot()
|
| H A D | tcache.c | 111 assert(szind < SC_NBINS); in tcache_gc_item_delay_compute() 126 assert(szind < SC_NBINS); in tcache_gc_small() 165 assert(szind >= SC_NBINS); in tcache_gc_large() 184 bool is_small = (szind < SC_NBINS); in tcache_event() 319 assert(binind < SC_NBINS); in tcache_bin_flush_impl() 667 unsigned n_reserved_bins = nhbins < SC_NBINS ? SC_NBINS : nhbins; in tcache_init() 674 if (i < SC_NBINS) { in tcache_init() 690 for (unsigned i = nhbins; i < SC_NBINS; i++) { in tcache_init() 789 if (i < SC_NBINS) { in tcache_flush_cache() 869 if (i < SC_NBINS) { in tcache_stats_merge() [all …]
|
| H A D | arena.c | 42 div_info_t arena_binind_div_info[SC_NBINS]; 47 uint32_t arena_bin_offsets[SC_NBINS]; 108 for (szind_t i = 0; i < SC_NSIZES - SC_NBINS; i++) { in arena_stats_merge() 143 curlextents * sz_index2size(SC_NBINS + i); in arena_stats_merge() 191 for (szind_t i = 0; i < SC_NBINS; i++) { in arena_stats_merge() 300 hindex = (index >= SC_NBINS) ? index - SC_NBINS : 0; in arena_large_malloc_stats_update() 316 hindex = (index >= SC_NBINS) ? index - SC_NBINS : 0; in arena_large_dalloc_stats_update() 703 for (unsigned i = 0; i < SC_NBINS; i++) { in arena_reset() 1058 assert(binind < SC_NBINS); in arena_fill_small_fresh() 1143 assert(binind < SC_NBINS); in arena_malloc_small() [all …]
|
| H A D | ctl.c | 1041 memset(ctl_arena->astats->bstats, 0, SC_NBINS * in ctl_arena_clear() 1043 memset(ctl_arena->astats->lstats, 0, (SC_NSIZES - SC_NBINS) * in ctl_arena_clear() 1067 for (i = 0; i < SC_NBINS; i++) { in ctl_arena_stats_amerge() 1188 for (i = 0; i < SC_NBINS; i++) { in ctl_arena_stats_sdmerge() 1215 for (i = 0; i < SC_NSIZES - SC_NBINS; i++) { in ctl_arena_stats_sdmerge() 3066 CTL_RO_NL_GEN(arenas_nbins, SC_NBINS, unsigned) in CTL_RO_NL_GEN() 3075 if (i > SC_NBINS) { in CTL_RO_NL_GEN() 3081 CTL_RO_NL_GEN(arenas_nlextents, SC_NSIZES - SC_NBINS, unsigned) 3082 CTL_RO_NL_GEN(arenas_lextent_i_size, sz_index2size(SC_NBINS+(szind_t)mib[2]), 3087 if (i > SC_NSIZES - SC_NBINS) { in arenas_lextent_i_index() [all …]
|
| H A D | jemalloc.c | 467 for (unsigned i = 0; i < SC_NBINS; i++) { in arena_bind() 1070 malloc_conf_init_helper(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS], in malloc_conf_init_helper() argument 1748 malloc_conf_init(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]) { in malloc_conf_init() argument 1810 unsigned bin_shard_sizes[SC_NBINS]; in malloc_init_hard_a0_locked() 2942 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree() 2950 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree() 2968 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree() 3108 assert(alloc_ctx.szind < SC_NBINS); in free_fastpath() 4168 if (likely(ind < SC_NBINS)) { in batch_alloc() 4192 if (likely(ind < SC_NBINS) && batch >= nregs) { in batch_alloc() [all …]
|
| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | tcache_structs.h | 36 uint8_t lg_fill_div[SC_NBINS]; 38 bool bin_refilled[SC_NBINS]; 43 uint8_t bin_flush_delay_items[SC_NBINS];
|
| H A D | bin_info.h | 46 extern bin_info_t bin_infos[SC_NBINS]; 48 void bin_info_boot(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]);
|
| H A D | arena_stats.h | 81 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS]; 106 arena_stats_large_t *lstats = &arena_stats->lstats[szind - SC_NBINS]; in arena_stats_large_flush_nrequests_add()
|
| H A D | bin.h | 51 void bin_shard_sizes_boot(unsigned bin_shards[SC_NBINS]); 52 bool bin_update_shard_size(unsigned bin_shards[SC_NBINS], size_t start_size,
|
| H A D | sc.h | 233 #define SC_NBINS ( \ macro 245 #if (SC_NBINS > 256)
|
| H A D | tcache_inlines.h | 32 assert(ind < SC_NBINS); in tcache_small_bin_disabled() 50 assert(binind < SC_NBINS); in tcache_alloc_small() 93 assert(binind >= SC_NBINS && binind < nhbins); in tcache_alloc_large()
|
| H A D | arena_inlines_b.h | 244 if (config_prof && unlikely(szind < SC_NBINS)) { in arena_dalloc_large_no_tcache() 284 if (config_prof && unlikely(szind < SC_NBINS)) { in arena_dalloc_large() 351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in arena_sdalloc_no_tcache() 360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)); in arena_sdalloc_no_tcache() 406 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in arena_sdalloc()
|
| H A D | ctl.h | 45 bin_stats_data_t bstats[SC_NBINS]; 46 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS];
|
| H A D | bin_types.h | 14 uint8_t binshard[SC_NBINS];
|
| H A D | tcache_types.h | 32 #define TCACHE_NBINS_MAX (SC_NBINS + SC_NGROUP * \
|
| H A D | arena_externs.h | 23 extern div_info_t arena_binind_div_info[SC_NBINS]; 35 extern uint32_t arena_bin_offsets[SC_NBINS];
|
| H A D | jemalloc_internal_inlines_c.h | 286 assert(ind < SC_NBINS); in imalloc_fastpath()
|