| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | pai.h | 11 bool *deferred_work_generated); 20 bool *deferred_work_generated); 23 bool *deferred_work_generated); 25 size_t old_size, size_t new_size, bool *deferred_work_generated); 27 bool *deferred_work_generated); 30 edata_list_active_t *list, bool *deferred_work_generated); 42 bool *deferred_work_generated) { in pai_alloc() argument 44 frequent_reuse, deferred_work_generated); in pai_alloc() 49 edata_list_active_t *results, bool *deferred_work_generated) { in pai_alloc_batch() argument 51 deferred_work_generated); in pai_alloc_batch() [all …]
|
| H A D | pa.h | 171 bool *deferred_work_generated); 174 size_t new_size, szind_t szind, bool zero, bool *deferred_work_generated); 180 size_t new_size, szind_t szind, bool *deferred_work_generated); 189 bool *deferred_work_generated);
|
| /src/contrib/jemalloc/src/ |
| H A D | sec.c | 8 bool *deferred_work_generated); 10 size_t old_size, size_t new_size, bool zero, bool *deferred_work_generated); 12 size_t old_size, size_t new_size, bool *deferred_work_generated); 14 bool *deferred_work_generated); 152 bool deferred_work_generated = false; in sec_flush_some_and_unlock() local 154 &deferred_work_generated); in sec_flush_some_and_unlock() 182 bool deferred_work_generated = false; in sec_batch_fill_and_alloc() local 184 1 + sec->opts.batch_fill_extra, &result, &deferred_work_generated); in sec_batch_fill_and_alloc() 222 bool guarded, bool frequent_reuse, bool *deferred_work_generated) { in sec_alloc() argument 232 deferred_work_generated); in sec_alloc() [all …]
|
| H A D | pai.c | 6 edata_list_active_t *results, bool *deferred_work_generated) { in pai_alloc_batch_default() argument 12 *deferred_work_generated |= deferred_by_alloc; in pai_alloc_batch_default() 23 edata_list_active_t *list, bool *deferred_work_generated) { in pai_dalloc_batch_default() argument 29 *deferred_work_generated |= deferred_by_dalloc; in pai_dalloc_batch_default()
|
| H A D | large.c | 67 bool deferred_work_generated = false; in large_ralloc_no_move_shrink() local 70 &deferred_work_generated); in large_ralloc_no_move_shrink() 74 if (deferred_work_generated) { in large_ralloc_no_move_shrink() 93 bool deferred_work_generated = false; in large_ralloc_no_move_expand() local 95 szind, zero, &deferred_work_generated); in large_ralloc_no_move_expand() 97 if (deferred_work_generated) { in large_ralloc_no_move_expand() 252 bool deferred_work_generated = false; in large_dalloc_finish_impl() local 253 pa_dalloc(tsdn, &arena->pa_shard, edata, &deferred_work_generated); in large_dalloc_finish_impl() 254 if (deferred_work_generated) { in large_dalloc_finish_impl()
|
| H A D | pa.c | 124 bool *deferred_work_generated) { in pa_alloc() argument 132 zero, /* guarded */ false, slab, deferred_work_generated); in pa_alloc() 140 guarded, slab, deferred_work_generated); in pa_alloc() 158 size_t new_size, szind_t szind, bool zero, bool *deferred_work_generated) { in pa_expand() argument 170 deferred_work_generated); in pa_expand() 183 size_t new_size, szind_t szind, bool *deferred_work_generated) { in pa_shrink() argument 194 deferred_work_generated); in pa_shrink() 207 bool *deferred_work_generated) { in pa_dalloc() argument 220 pai_dalloc(tsdn, pai, edata, deferred_work_generated); in pa_dalloc()
|
| H A D | hpa.c | 13 bool *deferred_work_generated); 15 size_t nallocs, edata_list_active_t *results, bool *deferred_work_generated); 17 size_t old_size, size_t new_size, bool zero, bool *deferred_work_generated); 19 size_t old_size, size_t new_size, bool *deferred_work_generated); 21 bool *deferred_work_generated); 23 edata_list_active_t *list, bool *deferred_work_generated); 628 bool *deferred_work_generated) { in hpa_try_alloc_batch_no_grow() argument 641 *deferred_work_generated = hpa_shard_has_deferred_work(tsdn, shard); in hpa_try_alloc_batch_no_grow() 649 bool *deferred_work_generated) { in hpa_alloc_batch_psset() argument 654 nallocs, results, deferred_work_generated); in hpa_alloc_batch_psset() [all …]
|
| H A D | pac.c | 9 bool *deferred_work_generated); 11 size_t old_size, size_t new_size, bool zero, bool *deferred_work_generated); 13 size_t old_size, size_t new_size, bool *deferred_work_generated); 15 bool *deferred_work_generated); 166 bool *deferred_work_generated) { in pac_alloc_impl() argument 192 size_t new_size, bool zero, bool *deferred_work_generated) { in pac_expand_impl() argument 230 size_t new_size, bool *deferred_work_generated) { in pac_shrink_impl() argument 246 *deferred_work_generated = true; in pac_shrink_impl() 252 bool *deferred_work_generated) { in pac_dalloc_impl() argument 279 *deferred_work_generated = true; in pac_dalloc_impl()
|
| H A D | arena.c | 332 bool deferred_work_generated = false; in arena_extent_alloc_large() local 339 /* slab */ false, szind, zero, guarded, &deferred_work_generated); in arena_extent_alloc_large() 340 assert(deferred_work_generated == false); in arena_extent_alloc_large() 569 bool deferred_work_generated = false; in arena_slab_dalloc() local 570 pa_dalloc(tsdn, &arena->pa_shard, slab, &deferred_work_generated); in arena_slab_dalloc() 571 if (deferred_work_generated) { in arena_slab_dalloc() 833 bool deferred_work_generated = false; in arena_slab_alloc() local 841 /* zero */ false, guarded, &deferred_work_generated); in arena_slab_alloc() 843 if (deferred_work_generated) { in arena_slab_alloc()
|