Searched refs:SC_LARGE_MINCLASS (Results 1 – 9 of 9) sorted by relevance
137 assert(oldusize >= SC_LARGE_MINCLASS in large_ralloc_no_move()138 && usize_max >= SC_LARGE_MINCLASS); in large_ralloc_no_move()193 assert(oldusize >= SC_LARGE_MINCLASS in large_ralloc()194 && usize >= SC_LARGE_MINCLASS); in large_ralloc()
296 if (usize < SC_LARGE_MINCLASS) { in arena_large_malloc_stats_update()297 usize = SC_LARGE_MINCLASS; in arena_large_malloc_stats_update()312 if (usize < SC_LARGE_MINCLASS) { in arena_large_dalloc_stats_update()313 usize = SC_LARGE_MINCLASS; in arena_large_dalloc_stats_update()1240 assert(isalloc(tsdn, ptr) == SC_LARGE_MINCLASS); in arena_prof_promote()1244 safety_check_set_redzone(ptr, usize, SC_LARGE_MINCLASS); in arena_prof_promote()1264 assert(isalloc(tsdn, ptr) == SC_LARGE_MINCLASS); in arena_prof_demote()1266 return SC_LARGE_MINCLASS; in arena_prof_demote()1278 if (config_opt_safety_checks && usize < SC_LARGE_MINCLASS) { in arena_dalloc_promoted()1283 assert(bumped_usize == SC_LARGE_MINCLASS); in arena_dalloc_promoted()[all …]
241 assert(sc_data->large_minclass == SC_LARGE_MINCLASS); in size_classes()
2421 sz_s2u(SC_LARGE_MINCLASS) : in imalloc_sample()2422 sz_sa2u(SC_LARGE_MINCLASS, dopts->alignment)) in imalloc_sample()2423 == SC_LARGE_MINCLASS); in imalloc_sample()2424 ind_large = sz_size2index(SC_LARGE_MINCLASS); in imalloc_sample()2425 bumped_usize = sz_s2u(SC_LARGE_MINCLASS); in imalloc_sample()3452 SC_LARGE_MINCLASS, alignment, zero, tcache, in irallocx_prof_sample()
270 assert(edata_size_get(edata) >= SC_LARGE_MINCLASS || in pac_dalloc_impl()
407 bool ranged_mprotect = head_and_tail && range <= SC_LARGE_MINCLASS; in pages_unmark_guards()
931 } else if (edata_size_get(edata) >= SC_LARGE_MINCLASS) { in extent_record()
335 if (usize < SC_LARGE_MINCLASS) { in sz_sa2u()347 if (size <= SC_LARGE_MINCLASS) { in sz_sa2u()348 usize = SC_LARGE_MINCLASS; in sz_sa2u()
267 #define SC_LARGE_MINCLASS ((size_t)1ULL << (LG_PAGE + SC_LG_NGROUP)) macro