Home
last modified time | relevance | path

Searched refs:sched_domain (Results 1 – 17 of 17) sorted by relevance

/linux/block/
H A Dkyber-iosched.c205 unsigned int sched_domain, unsigned int type) in flush_latency_buckets() argument
207 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in flush_latency_buckets()
208 atomic_t *cpu_buckets = cpu_latency->buckets[sched_domain][type]; in flush_latency_buckets()
220 unsigned int sched_domain, unsigned int type, in calculate_percentile() argument
223 unsigned int *buckets = kqd->latency_buckets[sched_domain][type]; in calculate_percentile()
236 if (!kqd->latency_timeout[sched_domain]) in calculate_percentile()
237 kqd->latency_timeout[sched_domain] = max(jiffies + HZ, 1UL); in calculate_percentile()
239 time_is_after_jiffies(kqd->latency_timeout[sched_domain])) { in calculate_percentile()
242 kqd->latency_timeout[sched_domain] = 0; in calculate_percentile()
250 memset(buckets, 0, sizeof(kqd->latency_buckets[sched_domain][type])); in calculate_percentile()
[all …]
/linux/kernel/sched/
H A Dtopology.c43 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one()
139 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
170 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
188 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
624 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain()
640 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu()
643 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu()
649 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains()
664 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
669 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
[all …]
H A Dext_idle.c233 struct sched_domain *sd; in llc_weight()
248 struct sched_domain *sd; in llc_span()
263 struct sched_domain *sd; in numa_weight()
282 struct sched_domain *sd; in numa_span()
H A Dfair.c2577 struct sched_domain *sd; in task_numa_migrate()
7377 wake_affine_weight(struct sched_domain *sd, struct task_struct *p, in wake_affine_weight()
7419 static int wake_affine(struct sched_domain *sd, struct task_struct *p, in wake_affine()
7440 sched_balance_find_dst_group(struct sched_domain *sd, struct task_struct *p, int this_cpu);
7502 static inline int sched_balance_find_dst_cpu(struct sched_domain *sd, struct task_struct *p, in sched_balance_find_dst_cpu()
7519 struct sched_domain *tmp; in sched_balance_find_dst_cpu()
7653 static int select_idle_smt(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_smt()
7689 static inline int select_idle_smt(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_smt()
7701 static int select_idle_cpu(struct task_struct *p, struct sched_domain *sd, bool has_idle_core, int … in select_idle_cpu()
7771 select_idle_capacity(struct task_struct *p, struct sched_domain *sd, int target) in select_idle_capacity()
[all …]
H A Dsched.h1234 struct sched_domain __rcu *sd;
2117 static inline struct sched_domain *highest_flag_domain(int cpu, int flag) in highest_flag_domain()
2119 struct sched_domain *sd, *hsd = NULL; in highest_flag_domain()
2138 static inline struct sched_domain *lowest_flag_domain(int cpu, int flag) in lowest_flag_domain()
2140 struct sched_domain *sd; in lowest_flag_domain()
2150 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_llc);
2155 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_numa);
2156 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
2157 DECLARE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
2794 extern void update_group_capacity(struct sched_domain *sd, int cpu);
H A Dstats.c118 struct sched_domain *sd; in show_schedstat()
H A Ddebug.c667 static void register_sd(struct sched_domain *sd, struct dentry *parent) in register_sd()
720 struct sched_domain *sd; in update_sched_domain_debugfs()
H A Dcore.c1208 struct sched_domain *sd; in get_nohz_timer_target()
3602 struct sched_domain *sd; in ttwu_stat()
6301 static bool steal_cookie_task(int cpu, struct sched_domain *sd) in steal_cookie_task()
6322 struct sched_domain *sd; in sched_core_balance()
H A Drt.c1765 struct sched_domain *sd; in find_lowest_rq()
H A Ddeadline.c2717 struct sched_domain *sd; in find_later_rq()
/linux/include/linux/sched/
H A Dtopology.h73 struct sched_domain { struct
75 struct sched_domain __rcu *parent; /* top domain must be null terminated */ argument
76 struct sched_domain __rcu *child; /* bottom domain must be null terminated */ argument
153 static inline struct cpumask *sched_domain_span(struct sched_domain *sd) in sched_domain_span() argument
173 struct sched_domain *__percpu *sd;
/linux/tools/testing/selftests/bpf/progs/
H A Dtest_access_variable_array.c12 struct sched_domain *sd) in BPF_PROG()
/linux/Documentation/scheduler/
H A Dsched-domains.rst5 Each CPU has a "base" scheduling domain (struct sched_domain). The domain
45 the parent sched_domain (if it exists), and the parent of the parent and so
70 of a sched_domain.
H A Dsched-capacity.rst288 - The SD_ASYM_CPUCAPACITY_FULL flag will be set at the lowest sched_domain
290 - The SD_ASYM_CPUCAPACITY flag will be set for any sched_domain that spans
319 sched_asym_cpucapacity static key will be enabled. However, the sched_domain
328 the sched_domain hierarchy (if relevant, i.e. the codepath targets a specific
H A Dsched-energy.rst335 flag to be set in the sched_domain hierarchy.
/linux/Documentation/translations/zh_CN/scheduler/
H A Dsched-domains.rst18 每个CPU有一个“基”调度域(struct sched_domain)。调度域层次结构从基调度域构建而来,可
H A Dsched-energy.rst275 参阅Documentation/scheduler/sched-capacity.rst以了解在sched_domain层次结构