| /linux/tools/verification/models/sched/ ! |
| H A D | sssw.dot | 5 {node [shape = doublecircle] "runnable"}; 6 {node [shape = circle] "runnable"}; 10 "__init_runnable" -> "runnable"; 11 "runnable" [label = "runnable", color = green3]; 12 …"runnable" -> "runnable" [ label = "sched_set_state_runnable\nsched_wakeup\nsched_switch_in\nsched… 13 "runnable" -> "sleepable" [ label = "sched_set_state_sleepable" ]; 14 "runnable" -> "sleeping" [ label = "sched_switch_blocking" ]; 16 "signal_wakeup" -> "runnable" [ label = "signal_deliver" ]; 20 "sleepable" -> "runnable" [ label = "sched_set_state_runnable\nsched_wakeup" ]; 25 "sleeping" -> "runnable" [ label = "sched_wakeup" ]; [all …]
|
| /linux/kernel/sched/ ! |
| H A D | pelt.c | 104 unsigned long load, unsigned long runnable, int running) in accumulate_sum() argument 144 if (runnable) in accumulate_sum() 145 sa->runnable_sum += runnable * contrib << SCHED_CAPACITY_SHIFT; in accumulate_sum() 182 unsigned long load, unsigned long runnable, int running) in ___update_load_sum() argument 218 runnable = running = 0; in ___update_load_sum() 227 if (!accumulate_sum(delta, sa, load, runnable, running)) in ___update_load_sum()
|
| H A D | ext_internal.h | 389 void (*runnable)(struct task_struct *p, u64 enq_flags); member 426 void (*stopping)(struct task_struct *p, bool runnable);
|
| H A D | fair.c | 2112 unsigned long runnable; member 2148 ((ns->compute_capacity * imbalance_pct) < (ns->runnable * 100)))) in numa_classify() 2153 ((ns->compute_capacity * imbalance_pct) > (ns->runnable * 100)))) in numa_classify() 2204 ns->runnable += cpu_runnable(rq); in update_numa_stats() 4382 __update_sa(&cfs_rq->avg, runnable, delta_avg, delta_sum); in update_tg_cfs_runnable() 4651 __update_sa(sa, runnable, -r, -r*divider); in update_cfs_rq_load_avg() 4737 __update_sa(&cfs_rq->avg, runnable, -se->avg.runnable_avg, -se->avg.runnable_sum); in detach_entity_load_avg() 7266 unsigned int runnable; in cpu_runnable_without() local 7273 runnable = READ_ONCE(cfs_rq->avg.runnable_avg); in cpu_runnable_without() 7276 lsub_positive(&runnable, p->se.avg.runnable_avg); in cpu_runnable_without() [all …]
|
| /linux/tools/sched_ext/ ! |
| H A D | scx_flatcg.bpf.c | 416 static void update_active_weight_sums(struct cgroup *cgrp, bool runnable) in update_active_weight_sums() argument 433 if (runnable) { in update_active_weight_sums() 450 if (!runnable) in update_active_weight_sums() 476 if (runnable) { in update_active_weight_sums() 503 if (runnable) in update_active_weight_sums() 539 void BPF_STRUCT_OPS(fcg_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument 952 .runnable = (void *)fcg_runnable,
|
| H A D | scx_simple.bpf.c | 110 void BPF_STRUCT_OPS(simple_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument
|
| H A D | scx_central.bpf.c | 246 void BPF_STRUCT_OPS(central_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument
|
| /linux/Documentation/scheduler/ ! |
| H A D | schedutil.rst | 35 Using this we track 2 key metrics: 'running' and 'runnable'. 'Running' 36 reflects the time an entity spends on the CPU, while 'runnable' reflects the 40 while 'runnable' will increase to reflect the amount of contention. 83 The result is that the above 'running' and 'runnable' metrics become invariant 104 A further runqueue wide sum (of runnable tasks) is maintained of:
|
| H A D | sched-ext.rst | 20 a runnable task stalls, or on invoking the SysRq key sequence 64 detection of any internal error including stalled runnable tasks aborts the 210 from the global DSQ. If that doesn't yield a runnable task either, 279 * If the previous task is an SCX task and still runnable, keep executing 312 ops.runnable(); /* Task becomes ready to run */ 314 while (task is runnable) { 323 while (task->scx.slice > 0 && task is runnable)
|
| H A D | sched-design-CFS.rst | 48 up CPU time between runnable tasks as close to "ideal multitasking hardware" as 75 CFS maintains a time-ordered rbtree, where all runnable tasks are sorted by the 174 Called when a task enters a runnable state. 180 When a task is no longer runnable, this function is called to keep the 192 This function checks if a task that entered the runnable state should
|
| H A D | sched-eevdf.rst | 14 runnable tasks with the same priority. To do so, it assigns a virtual run
|
| H A D | sched-bwc.rst | 15 slices as threads in the cgroup become runnable. Once all quota has been 202 decide which application is chosen to run, as they will both be runnable and
|
| /linux/Documentation/timers/ ! |
| H A D | no_hz.rst | 24 have only one runnable task (CONFIG_NO_HZ_FULL=y). Unless you 44 will frequently be multiple runnable tasks per CPU. In these cases, 107 If a CPU has only one runnable task, there is little point in sending it 109 Note that omitting scheduling-clock ticks for CPUs with only one runnable 113 sending scheduling-clock interrupts to CPUs with a single runnable task, 257 runnable task for a given CPU, even though there are a number 260 runnable high-priority SCHED_FIFO task and an arbitrary number 267 single runnable SCHED_FIFO task and multiple runnable SCHED_OTHER 270 And even when there are multiple runnable tasks on a given CPU,
|
| /linux/tools/testing/selftests/sched_ext/ ! |
| H A D | maximal.bpf.c | 42 void BPF_STRUCT_OPS(maximal_stopping, struct task_struct *p, bool runnable) in BPF_STRUCT_OPS() argument 144 .runnable = (void *) maximal_runnable,
|
| H A D | select_cpu_vtime.bpf.c | 67 bool runnable) in BPF_STRUCT_OPS() argument
|
| /linux/Documentation/core-api/real-time/ ! |
| H A D | theory.rst | 22 becoming runnable and its actual execution on the CPU. 32 consumed a fair share of CPU time relative to other runnable tasks. However, 34 becomes runnable. The currently running task may continue executing. 37 SCHED_FIFO. When a task with a real-time policy becomes runnable, the
|
| /linux/Documentation/trace/rv/ ! |
| H A D | monitor_sched.rst | 67 The set non runnable on its own context (snroc) monitor ensures changes in a 270 is woken up or set to ``runnable``. 287 back to runnable, the resulting switch (if there) looks like a yield to the 292 This monitor doesn't include a running state, ``sleepable`` and ``runnable`` 312 | | _blocking H runnable H | |
|
| /linux/Documentation/tools/rv/ ! |
| H A D | rv-mon-sched.rst | 52 * snroc: set non runnable on its own context
|
| /linux/drivers/gpu/drm/panthor/ ! |
| H A D | panthor_sched.c | 231 struct list_head runnable[PANTHOR_CSG_PRIORITY_COUNT]; member 2241 &sched->groups.runnable[group->priority]); in tick_ctx_cleanup() 2261 &sched->groups.runnable[group->priority]); in tick_ctx_cleanup() 2415 list_move_tail(&group->run_node, &sched->groups.runnable[prio]); in tick_ctx_apply() 2494 &sched->groups.runnable[prio], in tick_work() 2516 tick_ctx_pick_groups_from_list(sched, &ctx, &sched->groups.runnable[prio], in tick_work() 2636 &sched->groups.runnable[group->priority]); in sync_upd_work() 2676 struct list_head *queue = &sched->groups.runnable[group->priority]; in group_schedule_locked() 2770 &sched->groups.runnable[group->priority]); in panthor_group_start() 2962 for (i = 0; i < ARRAY_SIZE(sched->groups.runnable); i++) { in panthor_sched_pre_reset() [all …]
|
| /linux/Documentation/virt/ ! |
| H A D | guest-halt-polling.rst | 18 even with other runnable tasks in the host.
|
| /linux/Documentation/virt/kvm/ ! |
| H A D | halt-polling.rst | 18 interval or some other task on the runqueue is runnable the scheduler is 150 - Halt polling will only be conducted by the host when no other tasks are runnable on
|
| /linux/Documentation/accounting/ ! |
| H A D | delay-accounting.rst | 7 runnable task may wait for a free CPU to run on. 12 a) waiting for a CPU (while being runnable)
|
| H A D | taskstats-struct.rst | 112 /* Delay waiting for cpu, while runnable
|
| /linux/Documentation/admin-guide/pm/ ! |
| H A D | cpuidle.rst | 90 Tasks can be in various states. In particular, they are *runnable* if there are 93 events to occur or similar). When a task becomes runnable, the CPU scheduler 94 assigns it to one of the available CPUs to run and if there are no more runnable 97 another CPU). [If there are multiple runnable tasks assigned to one CPU 101 The special "idle" task becomes runnable if there are no other runnable tasks 193 multiple runnable tasks assigned to one CPU at the same time, the only way to 216 not be shared between multiple runnable tasks, the primary reason for using the
|
| /linux/Documentation/core-api/ ! |
| H A D | workqueue.rst | 126 number of the currently runnable workers. Generally, work items are 129 stalling should be optimal. As long as there are one or more runnable 230 concurrency level. In other words, runnable CPU intensive 238 regulated by the concurrency management and runnable
|