Lines Matching refs:allocation

24   - Support C++17 over-aligned allocation.  (@marksantaniello)
146 usage is improved with certain allocation patterns. As usual, the release and
262 - Allow arena index lookup based on allocation addresses via mallctl.
267 allocation. (@interwq, @davidtgoldblatt)
326 - Improve the fit for aligned allocation. (@interwq, @edwinsmith)
373 only allocation activity is to call free() after TLS destructors have been
473 - Improve reentrant allocation support, such that deadlock is less likely if
485 - Unify the allocation paths, and merge most fast-path branching decisions.
578 a single allocation's size exceeds the interval. (@jasone)
598 - Fix DSS (sbrk(2)-based) allocation. This regression was first released in
607 - Fix huge-aligned allocation. This regression was first released in 4.4.0.
627 is higher priority than address, so that the allocation policy prefers older
667 - Fix large allocation to search starting in the optimal size class heap,
674 - Fix over-sized allocation of radix tree leaf nodes. (@mjp41, @ogaun,
676 - Fix over-sized allocation of arena_t (plus associated stats) data
686 allocation. (@kspinka, @Whissi, @jasone)
692 - Fix TSD fetches to avoid (recursive) allocation. This is relevant to
700 - Fix bootstrapping issues for configurations that require allocation during
781 numerical overflow, and all allocation functions are guaranteed to indicate
810 - Move retained memory allocation out of the default chunk allocation
812 a custom chunk allocation function. This resolves a virtual memory leak.
886 allocation/deallocation within the application's thread-specific data
899 + Make one call to prof_active_get_unlocked() per allocation event, and use
900 the result throughout the relevant functions that handle an allocation
902 allocation events against concurrent prof_active changes.
954 - Refactor huge allocation to be managed by arenas, so that arenas now
960 mallctls provide high level per arena huge allocation statistics.
972 allocation versus deallocation.
1003 reduces the cost of repeated huge allocation/deallocation, because it
1011 - Randomly distribute large allocation base pointer alignment relative to page
1018 - Implement in-place huge allocation growing and shrinking.
1022 allocation, because a global data structure is critical for determining
1085 small/large allocation if chunk allocation failed. In the absence of this
1086 bug, chunk allocation failure would result in allocation failure, e.g. NULL
1092 - Use dss allocation precedence for huge allocations as well as small/large
1150 allocation, not just deallocation.
1151 - Fix a data race for large allocation stats counters.
1184 - Preserve errno during the first allocation. A readlink(2) call during
1186 set during the first allocation prior to this fix.
1250 allocation and dirty page purging algorithms in order to better control
1257 - Fix dss/mmap allocation precedence code to use recyclable mmap memory only
1258 after primary dss allocation fails.
1440 - Fix deadlocks on OS X that were due to memory allocation in
1448 + Fix error detection bugs for aligned memory allocation.
1526 - Add per thread allocation counters that can be accessed via the
1543 - Fix an allocation bug for small allocations that could be triggered if
1596 - Add support for allocation backed by one or more swap files, and allow the
1598 - Implement allocation profiling and leak checking.
1605 - Modify chunk allocation to work when address space layout randomization
1608 - Handle 0-size allocation requests in posix_memalign().