| /src/contrib/jemalloc/src/ |
| H A D | large.c | 14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc() argument 15 assert(usize == sz_s2u(usize)); in large_malloc() 17 return large_palloc(tsdn, arena, usize, CACHELINE, zero); in large_malloc() 21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc() argument 29 ausize = sz_sa2u(usize, alignment); in large_palloc() 35 arena = arena_choose_maybe_huge(tsdn_tsd(tsdn), arena, usize); in large_palloc() 38 arena, usize, alignment, zero)) == NULL) { in large_palloc() 55 large_ralloc_no_move_shrink(tsdn_t *tsdn, edata_t *edata, size_t usize) { in large_ralloc_no_move_shrink() argument 61 assert(old_usize > usize); in large_ralloc_no_move_shrink() 69 usize + sz_large_pad, sz_size2index(usize), in large_ralloc_no_move_shrink() [all …]
|
| H A D | jemalloc.c | 141 static void default_junk_alloc(void *ptr, size_t usize) { in default_junk_alloc() argument 142 memset(ptr, junk_alloc_byte, usize); in default_junk_alloc() 145 static void default_junk_free(void *ptr, size_t usize) { in default_junk_free() argument 146 memset(ptr, junk_free_byte, usize); in default_junk_free() 2244 bool usize; member 2258 static_opts->usize = false; in static_opts_init() 2273 size_t usize; member 2285 dynamic_opts->usize = 0; in dynamic_opts_init() 2299 aligned_usize_get(size_t size, size_t alignment, size_t *usize, szind_t *ind, in aligned_usize_get() argument 2301 assert(usize != NULL); in aligned_usize_get() [all …]
|
| H A D | arena.c | 291 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update() argument 296 if (usize < SC_LARGE_MINCLASS) { in arena_large_malloc_stats_update() 297 usize = SC_LARGE_MINCLASS; in arena_large_malloc_stats_update() 299 index = sz_size2index(usize); in arena_large_malloc_stats_update() 307 arena_large_dalloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_dalloc_stats_update() argument 312 if (usize < SC_LARGE_MINCLASS) { in arena_large_dalloc_stats_update() 313 usize = SC_LARGE_MINCLASS; in arena_large_dalloc_stats_update() 315 index = sz_size2index(usize); in arena_large_dalloc_stats_update() 324 size_t usize) { in arena_large_ralloc_stats_update() argument 325 arena_large_malloc_stats_update(tsdn, arena, usize); in arena_large_ralloc_stats_update() [all …]
|
| H A D | ckh.c | 273 size_t usize; in ckh_grow() local 276 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_grow() 277 if (unlikely(usize == 0 in ckh_grow() 278 || usize > SC_LARGE_MAXCLASS)) { in ckh_grow() 282 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, in ckh_grow() 313 size_t usize; in ckh_shrink() local 322 usize = sz_sa2u(sizeof(ckhc_t) << lg_curcells, CACHELINE); in ckh_shrink() 323 if (unlikely(usize == 0 || usize > SC_LARGE_MAXCLASS)) { in ckh_shrink() 326 tab = (ckhc_t *)ipallocztm(tsd_tsdn(tsd), usize, CACHELINE, true, NULL, in ckh_shrink() 362 size_t mincells, usize; in ckh_new() local [all …]
|
| H A D | san.c | 72 size_t usize = (left && right) in san_guard_pages() local 77 san_find_guarded_addr(edata, &guard1, &guard2, &addr, usize, left, in san_guard_pages() 84 edata_size_set(edata, usize); in san_guard_pages() 176 san_check_stashed_ptrs(void **ptrs, size_t nstashed, size_t usize) { in san_check_stashed_ptrs() argument 185 if (unlikely(san_stashed_corrupted(stashed, usize))) { in san_check_stashed_ptrs() 188 stashed, usize); in san_check_stashed_ptrs()
|
| H A D | prof.c | 100 size_t usize, prof_tctx_t *tctx) { in prof_malloc_sample_object() argument 111 szind_t szind = sz_size2index(usize); in prof_malloc_sample_object() 128 tctx->cnts.curbytes += usize; in prof_malloc_sample_object() 133 tctx->cnts.accumbytes += usize; in prof_malloc_sample_object() 141 prof_recent_alloc(tsd, edata, size, usize); in prof_malloc_sample_object() 150 prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_info_t *prof_info) { in prof_free_sampled_object() argument 157 szind_t szind = sz_size2index(usize); in prof_free_sampled_object() 161 assert(tctx->cnts.curbytes >= usize); in prof_free_sampled_object() 170 tctx->cnts.curbytes -= usize; in prof_free_sampled_object() 173 prof_try_log(tsd, usize, prof_info); in prof_free_sampled_object()
|
| H A D | base.c | 247 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_block_alloc() local 259 + usize)); in base_block_alloc() 438 size_t usize = ALIGNMENT_CEILING(size, alignment); in base_alloc_impl() local 439 size_t asize = usize + alignment - QUANTUM; in base_alloc_impl() 452 edata = base_extent_alloc(tsdn, base, usize, alignment); in base_alloc_impl() 460 ret = base_extent_bump_alloc(base, edata, usize, alignment); in base_alloc_impl()
|
| H A D | prof_log.c | 86 size_t usize; member 205 prof_try_log(tsd_t *tsd, size_t usize, prof_info_t *prof_info) { in prof_try_log() argument 268 new_node->usize = usize; in prof_try_log() 572 &alloc_node->usize); in prof_log_emit_allocs()
|
| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | sz.h | 150 size_t usize = (psz + delta_mask) & ~delta_mask; in sz_psz2u() local 151 return usize; in sz_psz2u() 231 size_t usize = grp_size + mod_size; in sz_index2size_compute() local 232 return usize; in sz_index2size_compute() 255 sz_size2index_usize_fastpath(size_t size, szind_t *ind, size_t *usize) { in sz_size2index_usize_fastpath() argument 257 *usize = sz_index2size_lookup_impl(*ind); in sz_size2index_usize_fastpath() 283 size_t usize = (size + delta_mask) & ~delta_mask; in sz_s2u_compute() local 284 return usize; in sz_s2u_compute() 314 size_t usize; in sz_sa2u() local 334 usize = sz_s2u(ALIGNMENT_CEILING(size, alignment)); in sz_sa2u() [all …]
|
| H A D | safety_check.h | 14 safety_check_set_redzone(void *ptr, size_t usize, size_t bumped_usize) { in safety_check_set_redzone() argument 15 assert(usize < bumped_usize); in safety_check_set_redzone() 16 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_set_redzone() 22 safety_check_verify_redzone(const void *ptr, size_t usize, size_t bumped_usize) in safety_check_verify_redzone() argument 24 for (size_t i = usize; i < bumped_usize && i < usize + 32; ++i) { in safety_check_verify_redzone()
|
| H A D | jemalloc_internal_inlines_c.h | 69 ipallocztm(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipallocztm() argument 73 assert(usize != 0); in ipallocztm() 74 assert(usize == sz_sa2u(usize, alignment)); in ipallocztm() 80 ret = arena_palloc(tsdn, arena, usize, alignment, zero, tcache); in ipallocztm() 89 ipalloct(tsdn_t *tsdn, size_t usize, size_t alignment, bool zero, in ipalloct() argument 91 return ipallocztm(tsdn, usize, alignment, zero, tcache, false, arena); in ipalloct() 95 ipalloc(tsd_t *tsd, size_t usize, size_t alignment, bool zero) { in ipalloc() argument 96 return ipallocztm(tsd_tsdn(tsd), usize, alignment, zero, in ipalloc() 143 size_t usize, copysize; in iralloct_realign() local 145 usize = sz_sa2u(size, alignment); in iralloct_realign() [all …]
|
| H A D | thread_event.h | 239 te_prof_sample_event_lookahead_surplus(tsd_t *tsd, size_t usize, in te_prof_sample_event_lookahead_surplus() argument 255 uint64_t accumbytes = tsd_thread_allocated_get(tsd) + usize - in te_prof_sample_event_lookahead_surplus() 261 assert(accumbytes - sample_wait < (uint64_t)usize); in te_prof_sample_event_lookahead_surplus() 269 te_prof_sample_event_lookahead(tsd_t *tsd, size_t usize) { in te_prof_sample_event_lookahead() argument 270 return te_prof_sample_event_lookahead_surplus(tsd, usize, NULL); in te_prof_sample_event_lookahead() 274 te_event_advance(tsd_t *tsd, size_t usize, bool is_alloc) { in te_event_advance() argument 281 te_ctx_current_bytes_set(&ctx, bytes_before + usize); in te_event_advance() 284 if (likely(usize < te_ctx_next_event_get(&ctx) - bytes_before)) { in te_event_advance() 292 thread_dalloc_event(tsd_t *tsd, size_t usize) { in thread_dalloc_event() argument 293 te_event_advance(tsd, usize, false); in thread_dalloc_event() [all …]
|
| H A D | san.h | 34 void san_check_stashed_ptrs(void **ptrs, size_t nstashed, size_t usize); 134 san_junk_ptr_locations(void *ptr, size_t usize, void **first, void **mid, in san_junk_ptr_locations() argument 140 *mid = (void *)((uintptr_t)ptr + ((usize >> 1) & ~(ptr_sz - 1))); in san_junk_ptr_locations() 141 assert(*first != *mid || usize == ptr_sz); in san_junk_ptr_locations() 151 *last = (void *)((uintptr_t)ptr + usize - sizeof(uaf_detect_junk)); in san_junk_ptr_locations() 152 assert(*first != *last || usize == ptr_sz); in san_junk_ptr_locations() 153 assert(*mid != *last || usize <= ptr_sz * 2); in san_junk_ptr_locations() 167 san_junk_ptr(void *ptr, size_t usize) { in san_junk_ptr() argument 169 memset(ptr, (char)uaf_detect_junk, usize); in san_junk_ptr() 174 san_junk_ptr_locations(ptr, usize, &first, &mid, &last); in san_junk_ptr()
|
| H A D | prof_inlines.h | 148 prof_malloc(tsd_t *tsd, const void *ptr, size_t size, size_t usize, in prof_malloc() argument 152 assert(usize == isalloc(tsd_tsdn(tsd), ptr)); in prof_malloc() 155 prof_malloc_sample_object(tsd, ptr, size, usize, tctx); in prof_malloc() 162 prof_realloc(tsd_t *tsd, const void *ptr, size_t size, size_t usize, in prof_realloc() argument 171 assert(usize == isalloc(tsd_tsdn(tsd), ptr)); in prof_realloc() 190 prof_malloc_sample_object(tsd, ptr, size, usize, tctx); in prof_realloc() 247 prof_free(tsd_t *tsd, const void *ptr, size_t usize, in prof_free() argument 253 assert(usize == isalloc(tsd_tsdn(tsd), ptr)); in prof_free() 257 prof_free_sampled_object(tsd, usize, &prof_info); in prof_free()
|
| H A D | tcache_inlines.h | 77 size_t usize = sz_index2size(binind); in tcache_alloc_small() local 78 assert(tcache_salloc(tsd_tsdn(tsd), ret) == usize); in tcache_alloc_small() 79 memset(ret, 0, usize); in tcache_alloc_small() 115 size_t usize = sz_index2size(binind); in tcache_alloc_large() local 116 assert(usize <= tcache_maxclass); in tcache_alloc_large() 117 memset(ret, 0, usize); in tcache_alloc_large()
|
| H A D | large_externs.h | 6 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero); 7 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, 11 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, void *ptr, size_t usize,
|
| H A D | prof_externs.h | 64 size_t usize, prof_tctx_t *tctx); 65 void prof_free_sampled_object(tsd_t *tsd, size_t usize, prof_info_t *prof_info);
|
| H A D | arena_externs.h | 47 size_t usize, size_t alignment, bool zero); 69 void *arena_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, 71 void arena_prof_promote(tsdn_t *tsdn, void *ptr, size_t usize);
|
| /src/contrib/libarchive/libarchive/ |
| H A D | archive_write_add_filter_lzop.c | 321 lzo_uint usize, csize; in drive_compressor() local 331 usize = (lzo_uint) in drive_compressor() 337 r = lzo1x_1_compress(data->uncompressed, usize, in drive_compressor() 342 r = lzo1x_1_15_compress(data->uncompressed, usize, in drive_compressor() 347 r = lzo1x_999_compress_level(data->uncompressed, usize, in drive_compressor() 359 archive_be32enc(p + header_bytes, (uint32_t)usize); in drive_compressor() 361 checksum = lzo_adler32(1, data->uncompressed, usize); in drive_compressor() 364 if (csize < usize) { in drive_compressor() 374 archive_be32enc(p + header_bytes + 4, (uint32_t)usize); in drive_compressor() 380 usize); in drive_compressor()
|
| /src/bin/ed/ |
| H A D | undo.c | 36 static long usize = 0; /* stack size variable */ variable 47 (ustack = (undo_t *) malloc((usize = USIZE) * sizeof(undo_t))) == NULL) { in push_undo_stack() 54 if (u_p < usize || in push_undo_stack() 55 (t = (undo_t *) realloc(ustack, (usize += USIZE) * sizeof(undo_t))) != NULL) { in push_undo_stack() 68 usize = 0; in push_undo_stack()
|
| /src/usr.bin/gzip/ |
| H A D | gzip.c | 1750 off_t usize, gsize; in handle_stdin() local 1793 usize = cat_fd(fourbytes, sizeof fourbytes, &gsize, STDIN_FILENO); in handle_stdin() 1796 usize = gz_uncompress(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1801 usize = unbzip2(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1812 usize = zuncompress(in, stdout, (char *)fourbytes, in handle_stdin() 1819 usize = unpack(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1825 usize = unxz(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1831 usize = unlz(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1837 usize = unzstd(STDIN_FILENO, STDOUT_FILENO, in handle_stdin() 1843 if (vflag && !tflag && usize != -1 && gsize != -1) in handle_stdin() [all …]
|
| /src/sys/contrib/openzfs/tests/zfs-tests/tests/functional/mount/ |
| H A D | umount_unlinked_drain.ksh | 55 usize=$(($nunlinks - $nunlinked)) 56 if [[ $iters == $MAX_ITERS && $usize == $1 ]]; then 59 if [[ $usize == $last_usize ]]; then 64 last_usize=$usize
|
| /src/contrib/bc/scripts/ |
| H A D | release.pkg.yao | 203 defbits: usize = if no64 { usize(32); } else { usize(64); }; 274 long_bit: usize 345 long_bit: usize 390 long_bit: usize 491 long_bit: usize, 555 long_bit: usize, 630 run_config_tests_make(cflags, cc, flags, true, usize(64)); 635 run_config_tests_make(cflags, cc, flags, false, usize(64)); 639 cc, flags, true, usize(64)); 646 cc, flags +~ " -e", true, usize(64)); [all …]
|
| /src/contrib/bc/ |
| H A D | build.pkg.rig | 334 MAX_WIDTH: usize = usize(72); 388 i: !usize = usize(0); 389 count: !usize = usize(0); 390 slashes: !usize = usize(0); 398 slashes! = slashes + usize(1); 399 i! = i + usize(2); 403 slashes! = slashes + usize(1); 404 i! = i + usize(1); 407 i! = i + usize(1); 417 i! = i + usize(1); [all …]
|
| /src/sys/compat/linux/ |
| H A D | linux_fork.c | 435 if (args->usize > PAGE_SIZE) in linux_clone3() 437 if (args->usize < LINUX_CLONE_ARGS_SIZE_VER0) in linux_clone3() 445 size = max(args->usize, sizeof(*uca)); in linux_clone3() 447 error = copyin(args->uargs, uca, args->usize); in linux_clone3()
|