xref: /src/contrib/jemalloc/include/jemalloc/internal/thread_event.h (revision c43cad87172039ccf38172129c79755ea79e6102)
1 #ifndef JEMALLOC_INTERNAL_THREAD_EVENT_H
2 #define JEMALLOC_INTERNAL_THREAD_EVENT_H
3 
4 #include "jemalloc/internal/tsd.h"
5 
6 /* "te" is short for "thread_event" */
7 
8 /*
9  * TE_MIN_START_WAIT should not exceed the minimal allocation usize.
10  */
11 #define TE_MIN_START_WAIT ((uint64_t)1U)
12 #define TE_MAX_START_WAIT UINT64_MAX
13 
14 /*
15  * Maximum threshold on thread_(de)allocated_next_event_fast, so that there is
16  * no need to check overflow in malloc fast path. (The allocation size in malloc
17  * fast path never exceeds SC_LOOKUP_MAXCLASS.)
18  */
19 #define TE_NEXT_EVENT_FAST_MAX (UINT64_MAX - SC_LOOKUP_MAXCLASS + 1U)
20 
21 /*
22  * The max interval helps make sure that malloc stays on the fast path in the
23  * common case, i.e. thread_allocated < thread_allocated_next_event_fast.  When
24  * thread_allocated is within an event's distance to TE_NEXT_EVENT_FAST_MAX
25  * above, thread_allocated_next_event_fast is wrapped around and we fall back to
26  * the medium-fast path. The max interval makes sure that we're not staying on
27  * the fallback case for too long, even if there's no active event or if all
28  * active events have long wait times.
29  */
30 #define TE_MAX_INTERVAL ((uint64_t)(4U << 20))
31 
32 /*
33  * Invalid elapsed time, for situations where elapsed time is not needed.  See
34  * comments in thread_event.c for more info.
35  */
36 #define TE_INVALID_ELAPSED UINT64_MAX
37 
38 typedef struct te_ctx_s {
39 	bool is_alloc;
40 	uint64_t *current;
41 	uint64_t *last_event;
42 	uint64_t *next_event;
43 	uint64_t *next_event_fast;
44 } te_ctx_t;
45 
46 void te_assert_invariants_debug(tsd_t *tsd);
47 void te_event_trigger(tsd_t *tsd, te_ctx_t *ctx);
48 void te_recompute_fast_threshold(tsd_t *tsd);
49 void tsd_te_init(tsd_t *tsd);
50 
51 /*
52  * List of all events, in the following format:
53  *  E(event,		(condition), is_alloc_event)
54  */
55 #define ITERATE_OVER_ALL_EVENTS						\
56     E(tcache_gc,		(opt_tcache_gc_incr_bytes > 0), true)	\
57     E(prof_sample,		(config_prof && opt_prof), true)  	\
58     E(stats_interval,		(opt_stats_interval >= 0), true)   	\
59     E(tcache_gc_dalloc,		(opt_tcache_gc_incr_bytes > 0), false)	\
60     E(peak_alloc,		config_stats, true)			\
61     E(peak_dalloc,		config_stats, false)
62 
63 #define E(event, condition_unused, is_alloc_event_unused)		\
64     C(event##_event_wait)
65 
66 /* List of all thread event counters. */
67 #define ITERATE_OVER_ALL_COUNTERS					\
68     C(thread_allocated)							\
69     C(thread_allocated_last_event)					\
70     ITERATE_OVER_ALL_EVENTS						\
71     C(prof_sample_last_event)						\
72     C(stats_interval_last_event)
73 
74 /* Getters directly wrap TSD getters. */
75 #define C(counter)							\
76 JEMALLOC_ALWAYS_INLINE uint64_t						\
77 counter##_get(tsd_t *tsd) {						\
78 	return tsd_##counter##_get(tsd);				\
79 }
80 
81 ITERATE_OVER_ALL_COUNTERS
82 #undef C
83 
84 /*
85  * Setters call the TSD pointer getters rather than the TSD setters, so that
86  * the counters can be modified even when TSD state is reincarnated or
87  * minimal_initialized: if an event is triggered in such cases, we will
88  * temporarily delay the event and let it be immediately triggered at the next
89  * allocation call.
90  */
91 #define C(counter)							\
92 JEMALLOC_ALWAYS_INLINE void						\
93 counter##_set(tsd_t *tsd, uint64_t v) {					\
94 	*tsd_##counter##p_get(tsd) = v;					\
95 }
96 
97 ITERATE_OVER_ALL_COUNTERS
98 #undef C
99 
100 /*
101  * For generating _event_wait getter / setter functions for each individual
102  * event.
103  */
104 #undef E
105 
106 /*
107  * The malloc and free fastpath getters -- use the unsafe getters since tsd may
108  * be non-nominal, in which case the fast_threshold will be set to 0.  This
109  * allows checking for events and tsd non-nominal in a single branch.
110  *
111  * Note that these can only be used on the fastpath.
112  */
113 JEMALLOC_ALWAYS_INLINE void
te_malloc_fastpath_ctx(tsd_t * tsd,uint64_t * allocated,uint64_t * threshold)114 te_malloc_fastpath_ctx(tsd_t *tsd, uint64_t *allocated, uint64_t *threshold) {
115 	*allocated = *tsd_thread_allocatedp_get_unsafe(tsd);
116 	*threshold = *tsd_thread_allocated_next_event_fastp_get_unsafe(tsd);
117 	assert(*threshold <= TE_NEXT_EVENT_FAST_MAX);
118 }
119 
120 JEMALLOC_ALWAYS_INLINE void
te_free_fastpath_ctx(tsd_t * tsd,uint64_t * deallocated,uint64_t * threshold)121 te_free_fastpath_ctx(tsd_t *tsd, uint64_t *deallocated, uint64_t *threshold) {
122 	/* Unsafe getters since this may happen before tsd_init. */
123 	*deallocated = *tsd_thread_deallocatedp_get_unsafe(tsd);
124 	*threshold = *tsd_thread_deallocated_next_event_fastp_get_unsafe(tsd);
125 	assert(*threshold <= TE_NEXT_EVENT_FAST_MAX);
126 }
127 
128 JEMALLOC_ALWAYS_INLINE bool
te_ctx_is_alloc(te_ctx_t * ctx)129 te_ctx_is_alloc(te_ctx_t *ctx) {
130 	return ctx->is_alloc;
131 }
132 
133 JEMALLOC_ALWAYS_INLINE uint64_t
te_ctx_current_bytes_get(te_ctx_t * ctx)134 te_ctx_current_bytes_get(te_ctx_t *ctx) {
135 	return *ctx->current;
136 }
137 
138 JEMALLOC_ALWAYS_INLINE void
te_ctx_current_bytes_set(te_ctx_t * ctx,uint64_t v)139 te_ctx_current_bytes_set(te_ctx_t *ctx, uint64_t v) {
140 	*ctx->current = v;
141 }
142 
143 JEMALLOC_ALWAYS_INLINE uint64_t
te_ctx_last_event_get(te_ctx_t * ctx)144 te_ctx_last_event_get(te_ctx_t *ctx) {
145 	return *ctx->last_event;
146 }
147 
148 JEMALLOC_ALWAYS_INLINE void
te_ctx_last_event_set(te_ctx_t * ctx,uint64_t v)149 te_ctx_last_event_set(te_ctx_t *ctx, uint64_t v) {
150 	*ctx->last_event = v;
151 }
152 
153 /* Below 3 for next_event_fast. */
154 JEMALLOC_ALWAYS_INLINE uint64_t
te_ctx_next_event_fast_get(te_ctx_t * ctx)155 te_ctx_next_event_fast_get(te_ctx_t *ctx) {
156 	uint64_t v = *ctx->next_event_fast;
157 	assert(v <= TE_NEXT_EVENT_FAST_MAX);
158 	return v;
159 }
160 
161 JEMALLOC_ALWAYS_INLINE void
te_ctx_next_event_fast_set(te_ctx_t * ctx,uint64_t v)162 te_ctx_next_event_fast_set(te_ctx_t *ctx, uint64_t v) {
163 	assert(v <= TE_NEXT_EVENT_FAST_MAX);
164 	*ctx->next_event_fast = v;
165 }
166 
167 JEMALLOC_ALWAYS_INLINE void
te_next_event_fast_set_non_nominal(tsd_t * tsd)168 te_next_event_fast_set_non_nominal(tsd_t *tsd) {
169 	/*
170 	 * Set the fast thresholds to zero when tsd is non-nominal.  Use the
171 	 * unsafe getter as this may get called during tsd init and clean up.
172 	 */
173 	*tsd_thread_allocated_next_event_fastp_get_unsafe(tsd) = 0;
174 	*tsd_thread_deallocated_next_event_fastp_get_unsafe(tsd) = 0;
175 }
176 
177 /* For next_event.  Setter also updates the fast threshold. */
178 JEMALLOC_ALWAYS_INLINE uint64_t
te_ctx_next_event_get(te_ctx_t * ctx)179 te_ctx_next_event_get(te_ctx_t *ctx) {
180 	return *ctx->next_event;
181 }
182 
183 JEMALLOC_ALWAYS_INLINE void
te_ctx_next_event_set(tsd_t * tsd,te_ctx_t * ctx,uint64_t v)184 te_ctx_next_event_set(tsd_t *tsd, te_ctx_t *ctx, uint64_t v) {
185 	*ctx->next_event = v;
186 	te_recompute_fast_threshold(tsd);
187 }
188 
189 /*
190  * The function checks in debug mode whether the thread event counters are in
191  * a consistent state, which forms the invariants before and after each round
192  * of thread event handling that we can rely on and need to promise.
193  * The invariants are only temporarily violated in the middle of
194  * te_event_advance() if an event is triggered (the te_event_trigger() call at
195  * the end will restore the invariants).
196  */
197 JEMALLOC_ALWAYS_INLINE void
te_assert_invariants(tsd_t * tsd)198 te_assert_invariants(tsd_t *tsd) {
199 	if (config_debug) {
200 		te_assert_invariants_debug(tsd);
201 	}
202 }
203 
204 JEMALLOC_ALWAYS_INLINE void
te_ctx_get(tsd_t * tsd,te_ctx_t * ctx,bool is_alloc)205 te_ctx_get(tsd_t *tsd, te_ctx_t *ctx, bool is_alloc) {
206 	ctx->is_alloc = is_alloc;
207 	if (is_alloc) {
208 		ctx->current = tsd_thread_allocatedp_get(tsd);
209 		ctx->last_event = tsd_thread_allocated_last_eventp_get(tsd);
210 		ctx->next_event = tsd_thread_allocated_next_eventp_get(tsd);
211 		ctx->next_event_fast =
212 		    tsd_thread_allocated_next_event_fastp_get(tsd);
213 	} else {
214 		ctx->current = tsd_thread_deallocatedp_get(tsd);
215 		ctx->last_event = tsd_thread_deallocated_last_eventp_get(tsd);
216 		ctx->next_event = tsd_thread_deallocated_next_eventp_get(tsd);
217 		ctx->next_event_fast =
218 		    tsd_thread_deallocated_next_event_fastp_get(tsd);
219 	}
220 }
221 
222 /*
223  * The lookahead functionality facilitates events to be able to lookahead, i.e.
224  * without touching the event counters, to determine whether an event would be
225  * triggered.  The event counters are not advanced until the end of the
226  * allocation / deallocation calls, so the lookahead can be useful if some
227  * preparation work for some event must be done early in the allocation /
228  * deallocation calls.
229  *
230  * Currently only the profiling sampling event needs the lookahead
231  * functionality, so we don't yet define general purpose lookahead functions.
232  *
233  * Surplus is a terminology referring to the amount of bytes beyond what's
234  * needed for triggering an event, which can be a useful quantity to have in
235  * general when lookahead is being called.
236  */
237 
238 JEMALLOC_ALWAYS_INLINE bool
te_prof_sample_event_lookahead_surplus(tsd_t * tsd,size_t usize,size_t * surplus)239 te_prof_sample_event_lookahead_surplus(tsd_t *tsd, size_t usize,
240     size_t *surplus) {
241 	if (surplus != NULL) {
242 		/*
243 		 * This is a dead store: the surplus will be overwritten before
244 		 * any read.  The initialization suppresses compiler warnings.
245 		 * Meanwhile, using SIZE_MAX to initialize is good for
246 		 * debugging purpose, because a valid surplus value is strictly
247 		 * less than usize, which is at most SIZE_MAX.
248 		 */
249 		*surplus = SIZE_MAX;
250 	}
251 	if (unlikely(!tsd_nominal(tsd) || tsd_reentrancy_level_get(tsd) > 0)) {
252 		return false;
253 	}
254 	/* The subtraction is intentionally susceptible to underflow. */
255 	uint64_t accumbytes = tsd_thread_allocated_get(tsd) + usize -
256 	    tsd_thread_allocated_last_event_get(tsd);
257 	uint64_t sample_wait = tsd_prof_sample_event_wait_get(tsd);
258 	if (accumbytes < sample_wait) {
259 		return false;
260 	}
261 	assert(accumbytes - sample_wait < (uint64_t)usize);
262 	if (surplus != NULL) {
263 		*surplus = (size_t)(accumbytes - sample_wait);
264 	}
265 	return true;
266 }
267 
268 JEMALLOC_ALWAYS_INLINE bool
te_prof_sample_event_lookahead(tsd_t * tsd,size_t usize)269 te_prof_sample_event_lookahead(tsd_t *tsd, size_t usize) {
270 	return te_prof_sample_event_lookahead_surplus(tsd, usize, NULL);
271 }
272 
273 JEMALLOC_ALWAYS_INLINE void
te_event_advance(tsd_t * tsd,size_t usize,bool is_alloc)274 te_event_advance(tsd_t *tsd, size_t usize, bool is_alloc) {
275 	te_assert_invariants(tsd);
276 
277 	te_ctx_t ctx;
278 	te_ctx_get(tsd, &ctx, is_alloc);
279 
280 	uint64_t bytes_before = te_ctx_current_bytes_get(&ctx);
281 	te_ctx_current_bytes_set(&ctx, bytes_before + usize);
282 
283 	/* The subtraction is intentionally susceptible to underflow. */
284 	if (likely(usize < te_ctx_next_event_get(&ctx) - bytes_before)) {
285 		te_assert_invariants(tsd);
286 	} else {
287 		te_event_trigger(tsd, &ctx);
288 	}
289 }
290 
291 JEMALLOC_ALWAYS_INLINE void
thread_dalloc_event(tsd_t * tsd,size_t usize)292 thread_dalloc_event(tsd_t *tsd, size_t usize) {
293 	te_event_advance(tsd, usize, false);
294 }
295 
296 JEMALLOC_ALWAYS_INLINE void
thread_alloc_event(tsd_t * tsd,size_t usize)297 thread_alloc_event(tsd_t *tsd, size_t usize) {
298 	te_event_advance(tsd, usize, true);
299 }
300 
301 #endif /* JEMALLOC_INTERNAL_THREAD_EVENT_H */
302