| /src/contrib/jemalloc/src/ |
| H A D | psset.c | 27 dst->nactive += src->nactive; in psset_bin_stats_accum() 57 binstats[huge_idx].nactive += mul * hpdata_nactive_get(ps); in psset_bin_stats_insert_remove() 61 psset->merged_stats.nactive += mul * hpdata_nactive_get(ps); in psset_bin_stats_insert_remove() 78 assert(psset->merged_stats.nactive == check_stats.nactive); in psset_bin_stats_insert_remove()
|
| H A D | pa.c | 9 atomic_fetch_add_zu(&shard->nactive, add_pages, ATOMIC_RELAXED); in pa_nactive_add() 14 assert(atomic_load_zu(&shard->nactive, ATOMIC_RELAXED) >= sub_pages); in pa_nactive_sub() 15 atomic_fetch_sub_zu(&shard->nactive, sub_pages, ATOMIC_RELAXED); in pa_nactive_sub() 54 atomic_store_zu(&shard->nactive, 0, ATOMIC_RELAXED); in pa_shard_init() 95 atomic_store_zu(&shard->nactive, 0, ATOMIC_RELAXED); in pa_shard_reset()
|
| H A D | pa_extra.c | 78 pa_shard_basic_stats_merge(pa_shard_t *shard, size_t *nactive, size_t *ndirty, in pa_shard_basic_stats_merge() argument 80 *nactive += atomic_load_zu(&shard->nactive, ATOMIC_RELAXED); in pa_shard_basic_stats_merge() 98 resident_pgs += atomic_load_zu(&shard->nactive, ATOMIC_RELAXED); in pa_shard_stats_merge()
|
| H A D | arena.c | 76 size_t *nactive, size_t *ndirty, size_t *nmuzzy) { in arena_basic_stats_merge() argument 81 pa_shard_basic_stats_merge(&arena->pa_shard, nactive, ndirty, nmuzzy); in arena_basic_stats_merge() 87 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats, in arena_stats_merge() argument 93 muzzy_decay_ms, nactive, ndirty, nmuzzy); in arena_stats_merge()
|
| H A D | ctl.c | 3761 arenas_i(mib[2])->astats->hpastats.psset_stats.full_slabs[0].nactive, size_t); 3770 arenas_i(mib[2])->astats->hpastats.psset_stats.full_slabs[1].nactive, size_t); 3779 arenas_i(mib[2])->astats->hpastats.psset_stats.empty_slabs[0].nactive, size_t); 3788 arenas_i(mib[2])->astats->hpastats.psset_stats.empty_slabs[1].nactive, size_t); 3797 arenas_i(mib[2])->astats->hpastats.psset_stats.nonfull_slabs[mib[5]][0].nactive, 3808 arenas_i(mib[2])->astats->hpastats.psset_stats.nonfull_slabs[mib[5]][1].nactive, 4182 pactivep = (size_t *)&(arena->pa_shard.nactive.repr); in experimental_arenas_i_pactivep_ctl()
|
| H A D | hpa.c | 954 assert(bin_stats->nactive == 0); in hpa_shard_assert_stats_empty()
|
| /src/contrib/jemalloc/include/jemalloc/internal/ |
| H A D | psset.h | 39 size_t nactive; member 123 return psset->merged_stats.nactive; in psset_nactive()
|
| H A D | pa.h | 77 atomic_zu_t nactive; member 226 void pa_shard_basic_stats_merge(pa_shard_t *shard, size_t *nactive,
|
| H A D | arena_externs.h | 39 ssize_t *muzzy_decay_ms, size_t *nactive, size_t *ndirty, size_t *nmuzzy); 42 size_t *nactive, size_t *ndirty, size_t *nmuzzy, arena_stats_t *astats,
|
| /src/sys/dev/cxgbe/ |
| H A D | t4_netmap.c | 594 int nactive[2] = {0, 0}; in cxgbe_netmap_split_rss() local 608 nactive[j]++; in cxgbe_netmap_split_rss() 614 if (nactive[0] == 0 || nactive[1] == 0) in cxgbe_netmap_split_rss() 618 if (nactive[0] > nactive[1]) { in cxgbe_netmap_split_rss() 621 } else if (nactive[0] < nactive[1]) { in cxgbe_netmap_split_rss() 786 int rc, i, nm_state, nactive; in cxgbe_netmap_off() local 822 nactive = 0; in cxgbe_netmap_off() 827 nactive++; in cxgbe_netmap_off() 849 if (nactive == 0) in cxgbe_netmap_off()
|
| /src/sys/net/altq/ |
| H A D | altq_hfsc.h | 113 int nactive; /* number of active children */ member 156 int nactive; /* number of active children */ member
|
| H A D | altq_hfsc.c | 1602 sp->nactive = cl->cl_nactive; in get_class_stats_v0() 1680 sp->nactive = cl->cl_nactive; in get_class_stats_v1()
|
| /src/crypto/openssh/ |
| H A D | scp.c | 957 size_t i, nactive = 0, ndone = 0; in brace_expand() local 966 if (append(cp, &active, &nactive) != 0) { in brace_expand() 970 while (nactive > 0) { in brace_expand() 971 cp = active[nactive - 1]; in brace_expand() 972 nactive--; in brace_expand() 973 if (brace_expand_one(cp, &active, &nactive, in brace_expand() 1009 for (i = 0; i < nactive; i++) in brace_expand()
|