| /linux/kernel/sched/ |
| H A D | loadavg.c | 82 long nr_active, delta = 0; in calc_load_fold_active() local 84 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 85 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active() 87 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 88 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 89 this_rq->calc_load_active = nr_active; in calc_load_fold_active()
|
| /linux/mm/ |
| H A D | dmapool.c | 53 size_t nr_active; member 84 pool->name, pool->nr_active, in pools_show() 188 pool->nr_active++; in pool_block_pop() 380 if (pool->nr_active) { in dma_pool_destroy() 461 pool->nr_active--; in dma_pool_free()
|
| /linux/include/trace/events/ |
| H A D | vmscan.h | 448 unsigned long nr_active, unsigned long nr_deactivated, 451 TP_ARGS(nid, nr_taken, nr_active, nr_deactivated, nr_referenced, priority, file), 456 __field(unsigned long, nr_active) 466 __entry->nr_active = nr_active; 476 __entry->nr_active, __entry->nr_deactivated, __entry->nr_referenced,
|
| /linux/arch/powerpc/platforms/cell/spufs/ |
| H A D | sched.c | 680 cbe_spu_info[node].nr_active--; in find_victim() 709 cbe_spu_info[node].nr_active++; in __spu_schedule() 750 cbe_spu_info[node].nr_active--; in spu_unschedule() 951 int nr_active = 0, node; in count_active_contexts() local 954 nr_active += cbe_spu_info[node].nr_active; in count_active_contexts() 955 nr_active += spu_prio->nr_waiting; in count_active_contexts() 957 return nr_active; in count_active_contexts()
|
| /linux/tools/sched_ext/ |
| H A D | scx_flatcg.bpf.c | 196 if (!cgc->nr_active) { in cgrp_refresh_hweight() 233 is_active = cgc->nr_active; in cgrp_refresh_hweight() 477 if (!cgc->nr_active++) { in update_active_weight_sums() 485 if (!--cgc->nr_active) { in update_active_weight_sums() 601 if (pcgc && cgc->nr_active) in BPF_STRUCT_OPS()
|
| H A D | scx_flatcg.h | 40 u32 nr_active; member
|
| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | net_timestamping.c | 11 int nr_active; variable 225 nr_active += !bpf_test_sockopt(skops, sk, 0); in skops_sockopt()
|
| H A D | setget_sockopt.c | 18 int nr_active; variable 415 nr_active += !(bpf_test_sockopt(skops, sk) || in skops_sockopt()
|
| /linux/block/ |
| H A D | blk-mq.h | 300 atomic_add(val, &hctx->nr_active); in __blk_mq_add_active_requests() 314 atomic_sub(val, &hctx->nr_active); in __blk_mq_sub_active_requests() 352 return atomic_read(&hctx->nr_active); in __blk_mq_active_requests()
|
| H A D | blk-mq.c | 4013 atomic_set(&hctx->nr_active, 0); in blk_mq_alloc_hctx()
|
| /linux/tools/testing/selftests/bpf/prog_tests/ |
| H A D | setget_sockopt.c | 64 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp() 144 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_ktls()
|
| H A D | net_timestamping.c | 213 ASSERT_EQ(bss->nr_active, 1, "nr_active"); in test_tcp()
|
| /linux/fs/ceph/ |
| H A D | mdsmap.c | 438 int i, nr_active = 0; in ceph_mdsmap_is_cluster_available() local 447 nr_active++; in ceph_mdsmap_is_cluster_available() 449 return nr_active > 0; in ceph_mdsmap_is_cluster_available()
|
| /linux/arch/x86/events/amd/ |
| H A D | uncore.c | 43 int nr_active; member 107 if (!ctx->nr_active || ctx->cpu != smp_processor_id()) in amd_uncore_hrtimer() 169 if (!ctx->nr_active++) in amd_uncore_start() 195 if (!--ctx->nr_active) in amd_uncore_stop() 947 if (!ctx->nr_active++) in amd_uncore_umc_start()
|
| /linux/drivers/nvme/host/ |
| H A D | multipath.c | 187 atomic_inc(&ns->ctrl->nr_active); in nvme_mpath_start_request() 206 atomic_dec_if_positive(&ns->ctrl->nr_active); in nvme_mpath_end_request() 421 depth = atomic_read(&ns->ctrl->nr_active); in nvme_queue_depth_path() 1097 return sysfs_emit(buf, "%d\n", atomic_read(&ns->ctrl->nr_active)); in queue_depth_show() 1362 atomic_set(&ctrl->nr_active, 0); in nvme_mpath_init_identify()
|
| H A D | nvme.h | 374 atomic_t nr_active; member
|
| /linux/kernel/ |
| H A D | workqueue.c | 284 int nr_active; /* L: nr of active works */ member 1676 return !pwq->nr_active && list_empty(&pwq->inactive_works); in pwq_is_empty() 1724 obtained = pwq->nr_active < READ_ONCE(wq->max_active); in pwq_tryinc_nr_active() 1776 pwq->nr_active++; in pwq_tryinc_nr_active() 1905 pwq->nr_active++; in node_activate_pending_pwq() 1944 pwq->nr_active--; in pwq_dec_nr_active() 3534 if (pwq->nr_active && need_to_create_worker(pool)) { in rescuer_thread() 6283 pwq->nr_active, pwq->refcnt, in show_pwq() 6865 WARN_ON_ONCE(pwq->nr_active < 0); in freeze_workqueues_busy() 6866 if (pwq->nr_active) { in freeze_workqueues_busy()
|
| /linux/arch/powerpc/include/asm/ |
| H A D | spu.h | 183 int nr_active; member
|
| /linux/include/linux/ |
| H A D | blk-mq.h | 421 atomic_t nr_active; member
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 1423 int i, nr_active = 0; in mptcp_subflow_get_send() local 1443 nr_active += !backup; in mptcp_subflow_get_send() 1462 if (!nr_active) in mptcp_subflow_get_send()
|