Home
last modified time | relevance | path

Searched refs:SC_NBINS (Results 1 – 18 of 18) sorted by relevance

/src/contrib/jemalloc/src/
H A Dbin_info.c6 bin_info_t bin_infos[SC_NBINS];
9 bin_infos_init(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS], in bin_infos_init() argument
10 bin_info_t infos[SC_NBINS]) { in bin_infos_init() argument
11 for (unsigned i = 0; i < SC_NBINS; i++) { in bin_infos_init()
27 bin_info_boot(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]) { in bin_info_boot() argument
H A Dbin.c10 bin_update_shard_size(unsigned bin_shard_sizes[SC_NBINS], size_t start_size, in bin_update_shard_size() argument
34 bin_shard_sizes_boot(unsigned bin_shard_sizes[SC_NBINS]) { in bin_shard_sizes_boot() argument
36 for (unsigned i = 0; i < SC_NBINS; i++) { in bin_shard_sizes_boot()
H A Dtcache.c111 assert(szind < SC_NBINS); in tcache_gc_item_delay_compute()
126 assert(szind < SC_NBINS); in tcache_gc_small()
165 assert(szind >= SC_NBINS); in tcache_gc_large()
184 bool is_small = (szind < SC_NBINS); in tcache_event()
319 assert(binind < SC_NBINS); in tcache_bin_flush_impl()
667 unsigned n_reserved_bins = nhbins < SC_NBINS ? SC_NBINS : nhbins; in tcache_init()
674 if (i < SC_NBINS) { in tcache_init()
690 for (unsigned i = nhbins; i < SC_NBINS; i++) { in tcache_init()
789 if (i < SC_NBINS) { in tcache_flush_cache()
869 if (i < SC_NBINS) { in tcache_stats_merge()
[all …]
H A Darena.c42 div_info_t arena_binind_div_info[SC_NBINS];
47 uint32_t arena_bin_offsets[SC_NBINS];
108 for (szind_t i = 0; i < SC_NSIZES - SC_NBINS; i++) { in arena_stats_merge()
143 curlextents * sz_index2size(SC_NBINS + i); in arena_stats_merge()
191 for (szind_t i = 0; i < SC_NBINS; i++) { in arena_stats_merge()
300 hindex = (index >= SC_NBINS) ? index - SC_NBINS : 0; in arena_large_malloc_stats_update()
316 hindex = (index >= SC_NBINS) ? index - SC_NBINS : 0; in arena_large_dalloc_stats_update()
703 for (unsigned i = 0; i < SC_NBINS; i++) { in arena_reset()
1058 assert(binind < SC_NBINS); in arena_fill_small_fresh()
1143 assert(binind < SC_NBINS); in arena_malloc_small()
[all …]
H A Dctl.c1041 memset(ctl_arena->astats->bstats, 0, SC_NBINS * in ctl_arena_clear()
1043 memset(ctl_arena->astats->lstats, 0, (SC_NSIZES - SC_NBINS) * in ctl_arena_clear()
1067 for (i = 0; i < SC_NBINS; i++) { in ctl_arena_stats_amerge()
1188 for (i = 0; i < SC_NBINS; i++) { in ctl_arena_stats_sdmerge()
1215 for (i = 0; i < SC_NSIZES - SC_NBINS; i++) { in ctl_arena_stats_sdmerge()
3066 CTL_RO_NL_GEN(arenas_nbins, SC_NBINS, unsigned) in CTL_RO_NL_GEN()
3075 if (i > SC_NBINS) { in CTL_RO_NL_GEN()
3081 CTL_RO_NL_GEN(arenas_nlextents, SC_NSIZES - SC_NBINS, unsigned)
3082 CTL_RO_NL_GEN(arenas_lextent_i_size, sz_index2size(SC_NBINS+(szind_t)mib[2]),
3087 if (i > SC_NSIZES - SC_NBINS) { in arenas_lextent_i_index()
[all …]
H A Djemalloc.c467 for (unsigned i = 0; i < SC_NBINS; i++) { in arena_bind()
1070 malloc_conf_init_helper(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS], in malloc_conf_init_helper() argument
1748 malloc_conf_init(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]) { in malloc_conf_init() argument
1810 unsigned bin_shard_sizes[SC_NBINS]; in malloc_init_hard_a0_locked()
2942 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree()
2950 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree()
2968 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in isfree()
3108 assert(alloc_ctx.szind < SC_NBINS); in free_fastpath()
4168 if (likely(ind < SC_NBINS)) { in batch_alloc()
4192 if (likely(ind < SC_NBINS) && batch >= nregs) { in batch_alloc()
[all …]
/src/contrib/jemalloc/include/jemalloc/internal/
H A Dtcache_structs.h36 uint8_t lg_fill_div[SC_NBINS];
38 bool bin_refilled[SC_NBINS];
43 uint8_t bin_flush_delay_items[SC_NBINS];
H A Dbin_info.h46 extern bin_info_t bin_infos[SC_NBINS];
48 void bin_info_boot(sc_data_t *sc_data, unsigned bin_shard_sizes[SC_NBINS]);
H A Darena_stats.h81 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS];
106 arena_stats_large_t *lstats = &arena_stats->lstats[szind - SC_NBINS]; in arena_stats_large_flush_nrequests_add()
H A Dbin.h51 void bin_shard_sizes_boot(unsigned bin_shards[SC_NBINS]);
52 bool bin_update_shard_size(unsigned bin_shards[SC_NBINS], size_t start_size,
H A Dsc.h233 #define SC_NBINS ( \ macro
245 #if (SC_NBINS > 256)
H A Dtcache_inlines.h32 assert(ind < SC_NBINS); in tcache_small_bin_disabled()
50 assert(binind < SC_NBINS); in tcache_alloc_small()
93 assert(binind >= SC_NBINS && binind < nhbins); in tcache_alloc_large()
H A Darena_inlines_b.h244 if (config_prof && unlikely(szind < SC_NBINS)) { in arena_dalloc_large_no_tcache()
284 if (config_prof && unlikely(szind < SC_NBINS)) { in arena_dalloc_large()
351 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in arena_sdalloc_no_tcache()
360 || alloc_ctx.slab == (alloc_ctx.szind < SC_NBINS)); in arena_sdalloc_no_tcache()
406 alloc_ctx.slab = (alloc_ctx.szind < SC_NBINS); in arena_sdalloc()
H A Dctl.h45 bin_stats_data_t bstats[SC_NBINS];
46 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS];
H A Dbin_types.h14 uint8_t binshard[SC_NBINS];
H A Dtcache_types.h32 #define TCACHE_NBINS_MAX (SC_NBINS + SC_NGROUP * \
H A Darena_externs.h23 extern div_info_t arena_binind_div_info[SC_NBINS];
35 extern uint32_t arena_bin_offsets[SC_NBINS];
H A Djemalloc_internal_inlines_c.h286 assert(ind < SC_NBINS); in imalloc_fastpath()