1d4d6bda0SWarner Losh #include "jemalloc/internal/jemalloc_preamble.h"
2d4d6bda0SWarner Losh #include "jemalloc/internal/jemalloc_internal_includes.h"
3d4d6bda0SWarner Losh
4d4d6bda0SWarner Losh #include "jemalloc/internal/assert.h"
548ec896eSWarner Losh #include "jemalloc/internal/emap.h"
6d4d6bda0SWarner Losh #include "jemalloc/internal/extent_mmap.h"
7d4d6bda0SWarner Losh #include "jemalloc/internal/mutex.h"
848ec896eSWarner Losh #include "jemalloc/internal/prof_recent.h"
9d4d6bda0SWarner Losh #include "jemalloc/internal/util.h"
10d4d6bda0SWarner Losh
11d4d6bda0SWarner Losh /******************************************************************************/
12d4d6bda0SWarner Losh
13d4d6bda0SWarner Losh void *
large_malloc(tsdn_t * tsdn,arena_t * arena,size_t usize,bool zero)14d4d6bda0SWarner Losh large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) {
15d4d6bda0SWarner Losh assert(usize == sz_s2u(usize));
16d4d6bda0SWarner Losh
17d4d6bda0SWarner Losh return large_palloc(tsdn, arena, usize, CACHELINE, zero);
18d4d6bda0SWarner Losh }
19d4d6bda0SWarner Losh
20d4d6bda0SWarner Losh void *
large_palloc(tsdn_t * tsdn,arena_t * arena,size_t usize,size_t alignment,bool zero)21d4d6bda0SWarner Losh large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
22d4d6bda0SWarner Losh bool zero) {
23d4d6bda0SWarner Losh size_t ausize;
2448ec896eSWarner Losh edata_t *edata;
25d4d6bda0SWarner Losh UNUSED bool idump JEMALLOC_CC_SILENCE_INIT(false);
26d4d6bda0SWarner Losh
27d4d6bda0SWarner Losh assert(!tsdn_null(tsdn) || arena != NULL);
28d4d6bda0SWarner Losh
29d4d6bda0SWarner Losh ausize = sz_sa2u(usize, alignment);
30d4d6bda0SWarner Losh if (unlikely(ausize == 0 || ausize > SC_LARGE_MAXCLASS)) {
31d4d6bda0SWarner Losh return NULL;
32d4d6bda0SWarner Losh }
33d4d6bda0SWarner Losh
34d4d6bda0SWarner Losh if (likely(!tsdn_null(tsdn))) {
35d4d6bda0SWarner Losh arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize);
36d4d6bda0SWarner Losh }
3748ec896eSWarner Losh if (unlikely(arena == NULL) || (edata = arena_extent_alloc_large(tsdn,
3848ec896eSWarner Losh arena, usize, alignment, zero)) == NULL) {
39d4d6bda0SWarner Losh return NULL;
40d4d6bda0SWarner Losh }
41d4d6bda0SWarner Losh
42d4d6bda0SWarner Losh /* See comments in arena_bin_slabs_full_insert(). */
43d4d6bda0SWarner Losh if (!arena_is_auto(arena)) {
4448ec896eSWarner Losh /* Insert edata into large. */
45d4d6bda0SWarner Losh malloc_mutex_lock(tsdn, &arena->large_mtx);
4648ec896eSWarner Losh edata_list_active_append(&arena->large, edata);
47d4d6bda0SWarner Losh malloc_mutex_unlock(tsdn, &arena->large_mtx);
48d4d6bda0SWarner Losh }
49d4d6bda0SWarner Losh
50d4d6bda0SWarner Losh arena_decay_tick(tsdn, arena);
5148ec896eSWarner Losh return edata_addr_get(edata);
52d4d6bda0SWarner Losh }
53d4d6bda0SWarner Losh
54d4d6bda0SWarner Losh static bool
large_ralloc_no_move_shrink(tsdn_t * tsdn,edata_t * edata,size_t usize)5548ec896eSWarner Losh large_ralloc_no_move_shrink(tsdn_t *tsdn, edata_t *edata, size_t usize) {
5648ec896eSWarner Losh arena_t *arena = arena_get_from_edata(edata);
5748ec896eSWarner Losh ehooks_t *ehooks = arena_get_ehooks(arena);
5848ec896eSWarner Losh size_t old_size = edata_size_get(edata);
5948ec896eSWarner Losh size_t old_usize = edata_usize_get(edata);
60d4d6bda0SWarner Losh
6148ec896eSWarner Losh assert(old_usize > usize);
62d4d6bda0SWarner Losh
6348ec896eSWarner Losh if (ehooks_split_will_fail(ehooks)) {
64d4d6bda0SWarner Losh return true;
65d4d6bda0SWarner Losh }
66d4d6bda0SWarner Losh
6748ec896eSWarner Losh bool deferred_work_generated = false;
6848ec896eSWarner Losh bool err = pa_shrink(tsdn, &arena->pa_shard, edata, old_size,
6948ec896eSWarner Losh usize + sz_large_pad, sz_size2index(usize),
7048ec896eSWarner Losh &deferred_work_generated);
7148ec896eSWarner Losh if (err) {
72d4d6bda0SWarner Losh return true;
73d4d6bda0SWarner Losh }
7448ec896eSWarner Losh if (deferred_work_generated) {
7548ec896eSWarner Losh arena_handle_deferred_work(tsdn, arena);
76d4d6bda0SWarner Losh }
7748ec896eSWarner Losh arena_extent_ralloc_large_shrink(tsdn, arena, edata, old_usize);
78d4d6bda0SWarner Losh
79d4d6bda0SWarner Losh return false;
80d4d6bda0SWarner Losh }
81d4d6bda0SWarner Losh
82d4d6bda0SWarner Losh static bool
large_ralloc_no_move_expand(tsdn_t * tsdn,edata_t * edata,size_t usize,bool zero)8348ec896eSWarner Losh large_ralloc_no_move_expand(tsdn_t *tsdn, edata_t *edata, size_t usize,
84d4d6bda0SWarner Losh bool zero) {
8548ec896eSWarner Losh arena_t *arena = arena_get_from_edata(edata);
86d4d6bda0SWarner Losh
8748ec896eSWarner Losh size_t old_size = edata_size_get(edata);
8848ec896eSWarner Losh size_t old_usize = edata_usize_get(edata);
8948ec896eSWarner Losh size_t new_size = usize + sz_large_pad;
90d4d6bda0SWarner Losh
91d4d6bda0SWarner Losh szind_t szind = sz_size2index(usize);
92d4d6bda0SWarner Losh
9348ec896eSWarner Losh bool deferred_work_generated = false;
9448ec896eSWarner Losh bool err = pa_expand(tsdn, &arena->pa_shard, edata, old_size, new_size,
9548ec896eSWarner Losh szind, zero, &deferred_work_generated);
9648ec896eSWarner Losh
9748ec896eSWarner Losh if (deferred_work_generated) {
9848ec896eSWarner Losh arena_handle_deferred_work(tsdn, arena);
9948ec896eSWarner Losh }
10048ec896eSWarner Losh
10148ec896eSWarner Losh if (err) {
10248ec896eSWarner Losh return true;
103d4d6bda0SWarner Losh }
104d4d6bda0SWarner Losh
105d4d6bda0SWarner Losh if (zero) {
10648ec896eSWarner Losh if (opt_cache_oblivious) {
10748ec896eSWarner Losh assert(sz_large_pad == PAGE);
108d4d6bda0SWarner Losh /*
109d4d6bda0SWarner Losh * Zero the trailing bytes of the original allocation's
110d4d6bda0SWarner Losh * last page, since they are in an indeterminate state.
111d4d6bda0SWarner Losh * There will always be trailing bytes, because ptr's
112d4d6bda0SWarner Losh * offset from the beginning of the extent is a multiple
113d4d6bda0SWarner Losh * of CACHELINE in [0 .. PAGE).
114d4d6bda0SWarner Losh */
115d4d6bda0SWarner Losh void *zbase = (void *)
11648ec896eSWarner Losh ((uintptr_t)edata_addr_get(edata) + old_usize);
117d4d6bda0SWarner Losh void *zpast = PAGE_ADDR2BASE((void *)((uintptr_t)zbase +
118d4d6bda0SWarner Losh PAGE));
119d4d6bda0SWarner Losh size_t nzero = (uintptr_t)zpast - (uintptr_t)zbase;
120d4d6bda0SWarner Losh assert(nzero > 0);
121d4d6bda0SWarner Losh memset(zbase, 0, nzero);
122d4d6bda0SWarner Losh }
123d4d6bda0SWarner Losh }
12448ec896eSWarner Losh arena_extent_ralloc_large_expand(tsdn, arena, edata, old_usize);
125d4d6bda0SWarner Losh
126d4d6bda0SWarner Losh return false;
127d4d6bda0SWarner Losh }
128d4d6bda0SWarner Losh
129d4d6bda0SWarner Losh bool
large_ralloc_no_move(tsdn_t * tsdn,edata_t * edata,size_t usize_min,size_t usize_max,bool zero)13048ec896eSWarner Losh large_ralloc_no_move(tsdn_t *tsdn, edata_t *edata, size_t usize_min,
131d4d6bda0SWarner Losh size_t usize_max, bool zero) {
13248ec896eSWarner Losh size_t oldusize = edata_usize_get(edata);
133d4d6bda0SWarner Losh
134d4d6bda0SWarner Losh /* The following should have been caught by callers. */
135d4d6bda0SWarner Losh assert(usize_min > 0 && usize_max <= SC_LARGE_MAXCLASS);
136d4d6bda0SWarner Losh /* Both allocation sizes must be large to avoid a move. */
137d4d6bda0SWarner Losh assert(oldusize >= SC_LARGE_MINCLASS
138d4d6bda0SWarner Losh && usize_max >= SC_LARGE_MINCLASS);
139d4d6bda0SWarner Losh
140d4d6bda0SWarner Losh if (usize_max > oldusize) {
141d4d6bda0SWarner Losh /* Attempt to expand the allocation in-place. */
14248ec896eSWarner Losh if (!large_ralloc_no_move_expand(tsdn, edata, usize_max,
143d4d6bda0SWarner Losh zero)) {
14448ec896eSWarner Losh arena_decay_tick(tsdn, arena_get_from_edata(edata));
145d4d6bda0SWarner Losh return false;
146d4d6bda0SWarner Losh }
147d4d6bda0SWarner Losh /* Try again, this time with usize_min. */
148d4d6bda0SWarner Losh if (usize_min < usize_max && usize_min > oldusize &&
14948ec896eSWarner Losh large_ralloc_no_move_expand(tsdn, edata, usize_min, zero)) {
15048ec896eSWarner Losh arena_decay_tick(tsdn, arena_get_from_edata(edata));
151d4d6bda0SWarner Losh return false;
152d4d6bda0SWarner Losh }
153d4d6bda0SWarner Losh }
154d4d6bda0SWarner Losh
155d4d6bda0SWarner Losh /*
156d4d6bda0SWarner Losh * Avoid moving the allocation if the existing extent size accommodates
157d4d6bda0SWarner Losh * the new size.
158d4d6bda0SWarner Losh */
159d4d6bda0SWarner Losh if (oldusize >= usize_min && oldusize <= usize_max) {
16048ec896eSWarner Losh arena_decay_tick(tsdn, arena_get_from_edata(edata));
161d4d6bda0SWarner Losh return false;
162d4d6bda0SWarner Losh }
163d4d6bda0SWarner Losh
164d4d6bda0SWarner Losh /* Attempt to shrink the allocation in-place. */
165d4d6bda0SWarner Losh if (oldusize > usize_max) {
16648ec896eSWarner Losh if (!large_ralloc_no_move_shrink(tsdn, edata, usize_max)) {
16748ec896eSWarner Losh arena_decay_tick(tsdn, arena_get_from_edata(edata));
168d4d6bda0SWarner Losh return false;
169d4d6bda0SWarner Losh }
170d4d6bda0SWarner Losh }
171d4d6bda0SWarner Losh return true;
172d4d6bda0SWarner Losh }
173d4d6bda0SWarner Losh
174d4d6bda0SWarner Losh static void *
large_ralloc_move_helper(tsdn_t * tsdn,arena_t * arena,size_t usize,size_t alignment,bool zero)175d4d6bda0SWarner Losh large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize,
176d4d6bda0SWarner Losh size_t alignment, bool zero) {
177d4d6bda0SWarner Losh if (alignment <= CACHELINE) {
178d4d6bda0SWarner Losh return large_malloc(tsdn, arena, usize, zero);
179d4d6bda0SWarner Losh }
180d4d6bda0SWarner Losh return large_palloc(tsdn, arena, usize, alignment, zero);
181d4d6bda0SWarner Losh }
182d4d6bda0SWarner Losh
183d4d6bda0SWarner Losh void *
large_ralloc(tsdn_t * tsdn,arena_t * arena,void * ptr,size_t usize,size_t alignment,bool zero,tcache_t * tcache,hook_ralloc_args_t * hook_args)184d4d6bda0SWarner Losh large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
185d4d6bda0SWarner Losh size_t alignment, bool zero, tcache_t *tcache,
186d4d6bda0SWarner Losh hook_ralloc_args_t *hook_args) {
18748ec896eSWarner Losh edata_t *edata = emap_edata_lookup(tsdn, &arena_emap_global, ptr);
188d4d6bda0SWarner Losh
18948ec896eSWarner Losh size_t oldusize = edata_usize_get(edata);
190d4d6bda0SWarner Losh /* The following should have been caught by callers. */
191d4d6bda0SWarner Losh assert(usize > 0 && usize <= SC_LARGE_MAXCLASS);
192d4d6bda0SWarner Losh /* Both allocation sizes must be large to avoid a move. */
193d4d6bda0SWarner Losh assert(oldusize >= SC_LARGE_MINCLASS
194d4d6bda0SWarner Losh && usize >= SC_LARGE_MINCLASS);
195d4d6bda0SWarner Losh
196d4d6bda0SWarner Losh /* Try to avoid moving the allocation. */
19748ec896eSWarner Losh if (!large_ralloc_no_move(tsdn, edata, usize, usize, zero)) {
198d4d6bda0SWarner Losh hook_invoke_expand(hook_args->is_realloc
199d4d6bda0SWarner Losh ? hook_expand_realloc : hook_expand_rallocx, ptr, oldusize,
200d4d6bda0SWarner Losh usize, (uintptr_t)ptr, hook_args->args);
20148ec896eSWarner Losh return edata_addr_get(edata);
202d4d6bda0SWarner Losh }
203d4d6bda0SWarner Losh
204d4d6bda0SWarner Losh /*
205d4d6bda0SWarner Losh * usize and old size are different enough that we need to use a
206d4d6bda0SWarner Losh * different size class. In that case, fall back to allocating new
207d4d6bda0SWarner Losh * space and copying.
208d4d6bda0SWarner Losh */
209d4d6bda0SWarner Losh void *ret = large_ralloc_move_helper(tsdn, arena, usize, alignment,
210d4d6bda0SWarner Losh zero);
211d4d6bda0SWarner Losh if (ret == NULL) {
212d4d6bda0SWarner Losh return NULL;
213d4d6bda0SWarner Losh }
214d4d6bda0SWarner Losh
215d4d6bda0SWarner Losh hook_invoke_alloc(hook_args->is_realloc
216d4d6bda0SWarner Losh ? hook_alloc_realloc : hook_alloc_rallocx, ret, (uintptr_t)ret,
217d4d6bda0SWarner Losh hook_args->args);
218d4d6bda0SWarner Losh hook_invoke_dalloc(hook_args->is_realloc
219d4d6bda0SWarner Losh ? hook_dalloc_realloc : hook_dalloc_rallocx, ptr, hook_args->args);
220d4d6bda0SWarner Losh
221d4d6bda0SWarner Losh size_t copysize = (usize < oldusize) ? usize : oldusize;
22248ec896eSWarner Losh memcpy(ret, edata_addr_get(edata), copysize);
22348ec896eSWarner Losh isdalloct(tsdn, edata_addr_get(edata), oldusize, tcache, NULL, true);
224d4d6bda0SWarner Losh return ret;
225d4d6bda0SWarner Losh }
226d4d6bda0SWarner Losh
227d4d6bda0SWarner Losh /*
22848ec896eSWarner Losh * locked indicates whether the arena's large_mtx is currently held.
229d4d6bda0SWarner Losh */
230d4d6bda0SWarner Losh static void
large_dalloc_prep_impl(tsdn_t * tsdn,arena_t * arena,edata_t * edata,bool locked)23148ec896eSWarner Losh large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata,
23248ec896eSWarner Losh bool locked) {
23348ec896eSWarner Losh if (!locked) {
234d4d6bda0SWarner Losh /* See comments in arena_bin_slabs_full_insert(). */
235d4d6bda0SWarner Losh if (!arena_is_auto(arena)) {
236d4d6bda0SWarner Losh malloc_mutex_lock(tsdn, &arena->large_mtx);
23748ec896eSWarner Losh edata_list_active_remove(&arena->large, edata);
238d4d6bda0SWarner Losh malloc_mutex_unlock(tsdn, &arena->large_mtx);
239d4d6bda0SWarner Losh }
240d4d6bda0SWarner Losh } else {
241d4d6bda0SWarner Losh /* Only hold the large_mtx if necessary. */
242d4d6bda0SWarner Losh if (!arena_is_auto(arena)) {
243d4d6bda0SWarner Losh malloc_mutex_assert_owner(tsdn, &arena->large_mtx);
24448ec896eSWarner Losh edata_list_active_remove(&arena->large, edata);
245d4d6bda0SWarner Losh }
246d4d6bda0SWarner Losh }
24748ec896eSWarner Losh arena_extent_dalloc_large_prep(tsdn, arena, edata);
248d4d6bda0SWarner Losh }
249d4d6bda0SWarner Losh
250d4d6bda0SWarner Losh static void
large_dalloc_finish_impl(tsdn_t * tsdn,arena_t * arena,edata_t * edata)25148ec896eSWarner Losh large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, edata_t *edata) {
25248ec896eSWarner Losh bool deferred_work_generated = false;
25348ec896eSWarner Losh pa_dalloc(tsdn, &arena->pa_shard, edata, &deferred_work_generated);
25448ec896eSWarner Losh if (deferred_work_generated) {
25548ec896eSWarner Losh arena_handle_deferred_work(tsdn, arena);
25648ec896eSWarner Losh }
257d4d6bda0SWarner Losh }
258d4d6bda0SWarner Losh
259d4d6bda0SWarner Losh void
large_dalloc_prep_locked(tsdn_t * tsdn,edata_t * edata)26048ec896eSWarner Losh large_dalloc_prep_locked(tsdn_t *tsdn, edata_t *edata) {
26148ec896eSWarner Losh large_dalloc_prep_impl(tsdn, arena_get_from_edata(edata), edata, true);
262d4d6bda0SWarner Losh }
263d4d6bda0SWarner Losh
264d4d6bda0SWarner Losh void
large_dalloc_finish(tsdn_t * tsdn,edata_t * edata)26548ec896eSWarner Losh large_dalloc_finish(tsdn_t *tsdn, edata_t *edata) {
26648ec896eSWarner Losh large_dalloc_finish_impl(tsdn, arena_get_from_edata(edata), edata);
267d4d6bda0SWarner Losh }
268d4d6bda0SWarner Losh
269d4d6bda0SWarner Losh void
large_dalloc(tsdn_t * tsdn,edata_t * edata)27048ec896eSWarner Losh large_dalloc(tsdn_t *tsdn, edata_t *edata) {
27148ec896eSWarner Losh arena_t *arena = arena_get_from_edata(edata);
27248ec896eSWarner Losh large_dalloc_prep_impl(tsdn, arena, edata, false);
27348ec896eSWarner Losh large_dalloc_finish_impl(tsdn, arena, edata);
274d4d6bda0SWarner Losh arena_decay_tick(tsdn, arena);
275d4d6bda0SWarner Losh }
276d4d6bda0SWarner Losh
277d4d6bda0SWarner Losh size_t
large_salloc(tsdn_t * tsdn,const edata_t * edata)27848ec896eSWarner Losh large_salloc(tsdn_t *tsdn, const edata_t *edata) {
27948ec896eSWarner Losh return edata_usize_get(edata);
280d4d6bda0SWarner Losh }
281d4d6bda0SWarner Losh
282d4d6bda0SWarner Losh void
large_prof_info_get(tsd_t * tsd,edata_t * edata,prof_info_t * prof_info,bool reset_recent)28348ec896eSWarner Losh large_prof_info_get(tsd_t *tsd, edata_t *edata, prof_info_t *prof_info,
28448ec896eSWarner Losh bool reset_recent) {
28548ec896eSWarner Losh assert(prof_info != NULL);
28648ec896eSWarner Losh
28748ec896eSWarner Losh prof_tctx_t *alloc_tctx = edata_prof_tctx_get(edata);
28848ec896eSWarner Losh prof_info->alloc_tctx = alloc_tctx;
28948ec896eSWarner Losh
29048ec896eSWarner Losh if ((uintptr_t)alloc_tctx > (uintptr_t)1U) {
29148ec896eSWarner Losh nstime_copy(&prof_info->alloc_time,
29248ec896eSWarner Losh edata_prof_alloc_time_get(edata));
29348ec896eSWarner Losh prof_info->alloc_size = edata_prof_alloc_size_get(edata);
29448ec896eSWarner Losh if (reset_recent) {
29548ec896eSWarner Losh /*
29648ec896eSWarner Losh * Reset the pointer on the recent allocation record,
29748ec896eSWarner Losh * so that this allocation is recorded as released.
29848ec896eSWarner Losh */
29948ec896eSWarner Losh prof_recent_alloc_reset(tsd, edata);
30048ec896eSWarner Losh }
30148ec896eSWarner Losh }
30248ec896eSWarner Losh }
30348ec896eSWarner Losh
30448ec896eSWarner Losh static void
large_prof_tctx_set(edata_t * edata,prof_tctx_t * tctx)30548ec896eSWarner Losh large_prof_tctx_set(edata_t *edata, prof_tctx_t *tctx) {
30648ec896eSWarner Losh edata_prof_tctx_set(edata, tctx);
307d4d6bda0SWarner Losh }
308d4d6bda0SWarner Losh
309d4d6bda0SWarner Losh void
large_prof_tctx_reset(edata_t * edata)31048ec896eSWarner Losh large_prof_tctx_reset(edata_t *edata) {
31148ec896eSWarner Losh large_prof_tctx_set(edata, (prof_tctx_t *)(uintptr_t)1U);
312d4d6bda0SWarner Losh }
313d4d6bda0SWarner Losh
314d4d6bda0SWarner Losh void
large_prof_info_set(edata_t * edata,prof_tctx_t * tctx,size_t size)31548ec896eSWarner Losh large_prof_info_set(edata_t *edata, prof_tctx_t *tctx, size_t size) {
31648ec896eSWarner Losh nstime_t t;
31748ec896eSWarner Losh nstime_prof_init_update(&t);
31848ec896eSWarner Losh edata_prof_alloc_time_set(edata, &t);
31948ec896eSWarner Losh edata_prof_alloc_size_set(edata, size);
32048ec896eSWarner Losh edata_prof_recent_alloc_init(edata);
32148ec896eSWarner Losh large_prof_tctx_set(edata, tctx);
322d4d6bda0SWarner Losh }
323