| /src/contrib/jemalloc/src/ |
| H A D | sz.c | 59 for (unsigned i = 0; i < SC_NSIZES; i++) { in sz_boot_pind2sz_tab() 73 size_t sz_index2size_tab[SC_NSIZES]; in JEMALLOC_ALIGNED() 77 for (unsigned i = 0; i < SC_NSIZES; i++) { 95 for (unsigned sc_ind = 0; sc_ind < SC_NSIZES && dst_ind < dst_max;
|
| H A D | emap.c | 231 emap_rtree_write_acquired(tsdn, emap, elm_a, elm_b, NULL, SC_NSIZES, in emap_deregister_boundary() 252 if (szind != SC_NSIZES) { in emap_remap() 321 prepare->lead_elm_b, lead, SC_NSIZES, /* slab */ false); in emap_split_commit() 323 prepare->trail_elm_b, trail, SC_NSIZES, /* slab */ false); in emap_split_commit() 341 clear_contents.metadata.szind = SC_NSIZES; in emap_merge_commit() 361 lead, SC_NSIZES, false); in emap_merge_commit()
|
| H A D | extent_dss.c | 159 SC_NSIZES, extent_sn_next( in extent_alloc_dss() 209 size, false, SC_NSIZES, in extent_alloc_dss()
|
| H A D | san.c | 89 emap_register_boundary(tsdn, emap, edata, SC_NSIZES, in san_guard_pages() 127 emap_register_boundary(tsdn, emap, edata, SC_NSIZES, in san_unguard_pages_impl()
|
| H A D | prof_stats.c | 14 assert(ind < SC_NSIZES); in prof_stats_enter()
|
| H A D | pa.c | 208 emap_remap(tsdn, shard->emap, edata, SC_NSIZES, /* slab */ false); in pa_dalloc() 217 edata_szind_set(edata, SC_NSIZES); in pa_dalloc()
|
| H A D | hpa.c | 587 SC_NSIZES, /* sn */ hpdata_age_get(ps), extent_state_active, in hpa_try_alloc_one_no_grow() 601 SC_NSIZES, /* slab */ false); in hpa_try_alloc_one_no_grow() 749 SC_NSIZES); in hpa_alloc_batch() 805 assert(edata_szind_get_maybe_invalid(edata) == SC_NSIZES); in hpa_dalloc_prepare_unlocked()
|
| H A D | base.c | 384 for (szind_t i = 0; i < SC_NSIZES; i++) { in base_new() 443 for (szind_t i = sz_size2index(asize); i < SC_NSIZES; i++) { in base_alloc_impl()
|
| H A D | extent.c | 328 if (emap_register_boundary(tsdn, pac->emap, edata, SC_NSIZES, in extent_register_impl() 681 alloc_size, false, SC_NSIZES, extent_sn_next(pac), in extent_grow_retained() 1010 size, /* slab */ false, SC_NSIZES, extent_sn_next(pac), in extent_alloc_wrapper() 1197 /* slab */ false, SC_NSIZES, edata_sn_get(edata), in extent_split_impl()
|
| H A D | ctl.c | 1043 memset(ctl_arena->astats->lstats, 0, (SC_NSIZES - SC_NBINS) * in ctl_arena_clear() 1215 for (i = 0; i < SC_NSIZES - SC_NBINS; i++) { in ctl_arena_stats_sdmerge() 3081 CTL_RO_NL_GEN(arenas_nlextents, SC_NSIZES - SC_NBINS, unsigned) 3087 if (i > SC_NSIZES - SC_NBINS) { in arenas_lextent_i_index() 3719 if (j > SC_NSIZES - SC_NBINS) { in stats_arenas_i_lextents_j_index() 4366 if (lextent_ind >= SC_NSIZES - SC_NBINS) { in prof_stats_lextents_i_live_ctl() 4392 if (lextent_ind >= SC_NSIZES - SC_NBINS) { in prof_stats_lextents_i_accum_ctl() 4410 if (i >= SC_NSIZES - SC_NBINS) { in prof_stats_lextents_i_index()
|
| H A D | jemalloc.c | 2305 if (unlikely(*ind >= SC_NSIZES)) { in aligned_usize_get() 2883 assert(alloc_ctx.szind != SC_NSIZES); in ifree() 3094 assert(alloc_ctx.szind != SC_NSIZES); in free_fastpath() 3525 assert(alloc_ctx.szind != SC_NSIZES); in do_rallocx() 3787 assert(alloc_ctx.szind != SC_NSIZES); in je_xallocx()
|
| H A D | arena.c | 108 for (szind_t i = 0; i < SC_NSIZES - SC_NBINS; i++) { in arena_stats_merge() 687 assert(alloc_ctx.szind != SC_NSIZES); in arena_reset()
|
| H A D | prof_data.c | 544 for (szind_t i = 0; i < SC_NSIZES; i++) { in prof_unbias_map_init()
|
| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | sc.h | 198 #define SC_NSIZES (SC_NTINY + SC_NPSEUDO + SC_NREGULAR) macro 344 sc_t sc[SC_NSIZES];
|
| H A D | edata.h | 173 #define EDATA_BITS_SZIND_WIDTH LG_CEIL(SC_NSIZES) 269 assert(szind <= SC_NSIZES); in edata_szind_get_maybe_invalid() 276 assert(szind < SC_NSIZES); /* Never call when "invalid". */ in edata_szind_get() 471 assert(szind <= SC_NSIZES); /* SC_NSIZES means "invalid". */ in edata_szind_set() 627 edata_szind_set(edata, SC_NSIZES); in edata_binit()
|
| H A D | sz.h | 40 extern size_t sz_index2size_tab[SC_NSIZES]; 157 return SC_NSIZES; in sz_size2index_compute() 250 assert(index < SC_NSIZES); in sz_index2size()
|
| H A D | prof_types.h | 44 # define PROF_SC_NSIZES SC_NSIZES
|
| H A D | arena_inlines_b.h | 177 assert(alloc_ctx.szind != SC_NSIZES); in arena_salloc() 208 assert(full_alloc_ctx.szind != SC_NSIZES); in arena_vsalloc() 268 assert(alloc_ctx.szind < SC_NSIZES); in arena_dalloc_no_tcache() 325 assert(alloc_ctx.szind < SC_NSIZES); in arena_dalloc()
|
| H A D | arena_stats.h | 81 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS];
|
| H A D | base.h | 73 edata_heap_t avail[SC_NSIZES];
|
| H A D | bitmap.h | 11 #if SC_LG_SLAB_MAXREGS > LG_CEIL(SC_NSIZES) 16 # define LG_BITMAP_MAXBITS LG_CEIL(SC_NSIZES)
|
| H A D | rtree.h | 34 #if RTREE_NHIB >= LG_CEIL(SC_NSIZES) 534 contents.metadata.szind = SC_NSIZES; in rtree_clear() 546 contents.metadata.szind = SC_NSIZES; in rtree_clear_range()
|
| H A D | ctl.h | 46 arena_stats_large_t lstats[SC_NSIZES - SC_NBINS];
|