Lines Matching +full:domain +full:- +full:idle +full:- +full:state

1 // SPDX-License-Identifier: GPL-2.0-only
3 * RISC-V SBI CPU idle driver.
9 #define pr_fmt(fmt) "cpuidle-riscv-sbi: " fmt
38 u32 state; member
47 static inline void sbi_set_domain_state(u32 state) in sbi_set_domain_state() argument
51 data->available = true; in sbi_set_domain_state()
52 data->state = state; in sbi_set_domain_state()
59 return data->state; in sbi_get_domain_state()
66 data->available = false; in sbi_clear_domain_state()
73 return data->available; in sbi_is_domain_state_available()
88 static int sbi_suspend(u32 state) in sbi_suspend() argument
90 if (state & SBI_HSM_SUSP_NON_RET_BIT) in sbi_suspend()
91 return cpu_suspend(state, sbi_suspend_finisher); in sbi_suspend()
93 return sbi_suspend_finisher(state, 0, 0); in sbi_suspend()
100 u32 state = states[idx]; in sbi_cpuidle_enter_state() local
102 if (state & SBI_HSM_SUSP_NON_RET_BIT) in sbi_cpuidle_enter_state()
103 return CPU_PM_CPU_IDLE_ENTER_PARAM(sbi_suspend, idx, state); in sbi_cpuidle_enter_state()
106 idx, state); in sbi_cpuidle_enter_state()
114 u32 *states = data->states; in __sbi_enter_domain_idle_state()
115 struct device *pd_dev = data->dev; in __sbi_enter_domain_idle_state()
116 u32 state; in __sbi_enter_domain_idle_state() local
121 return -1; in __sbi_enter_domain_idle_state()
132 state = sbi_get_domain_state(); in __sbi_enter_domain_idle_state()
134 state = states[idx]; in __sbi_enter_domain_idle_state()
136 ret = sbi_suspend(state) ? -1 : idx; in __sbi_enter_domain_idle_state()
147 /* Clear the domain state to start fresh when back from idle. */ in __sbi_enter_domain_idle_state()
181 /* Clear domain state to start fresh at next online. */ in sbi_cpuidle_cpuhp_down()
200 pr_warn("Failed %d while setup cpuhp state\n", err); in sbi_idle_init_cpuhp()
204 { .compatible = "riscv,idle-state",
209 static bool sbi_suspend_state_is_valid(u32 state) in sbi_suspend_state_is_valid() argument
211 if (state > SBI_HSM_SUSPEND_RET_DEFAULT && in sbi_suspend_state_is_valid()
212 state < SBI_HSM_SUSPEND_RET_PLATFORM) in sbi_suspend_state_is_valid()
214 if (state > SBI_HSM_SUSPEND_NON_RET_DEFAULT && in sbi_suspend_state_is_valid()
215 state < SBI_HSM_SUSPEND_NON_RET_PLATFORM) in sbi_suspend_state_is_valid()
220 static int sbi_dt_parse_state_node(struct device_node *np, u32 *state) in sbi_dt_parse_state_node() argument
222 int err = of_property_read_u32(np, "riscv,sbi-suspend-param", state); in sbi_dt_parse_state_node()
225 pr_warn("%pOF missing riscv,sbi-suspend-param property\n", np); in sbi_dt_parse_state_node()
229 if (!sbi_suspend_state_is_valid(*state)) { in sbi_dt_parse_state_node()
230 pr_warn("Invalid SBI suspend state %#x\n", *state); in sbi_dt_parse_state_node()
231 return -EINVAL; in sbi_dt_parse_state_node()
245 data->dev = dt_idle_attach_cpu(cpu, "sbi"); in sbi_dt_cpu_init_topology()
246 if (IS_ERR_OR_NULL(data->dev)) in sbi_dt_cpu_init_topology()
247 return PTR_ERR_OR_ZERO(data->dev); in sbi_dt_cpu_init_topology()
250 * Using the deepest state for the CPU to trigger a potential selection in sbi_dt_cpu_init_topology()
251 * of a shared state for the domain, assumes the domain states are all in sbi_dt_cpu_init_topology()
254 drv->states[state_count - 1].flags |= CPUIDLE_FLAG_RCU_IDLE; in sbi_dt_cpu_init_topology()
255 drv->states[state_count - 1].enter = sbi_enter_domain_idle_state; in sbi_dt_cpu_init_topology()
256 drv->states[state_count - 1].enter_s2idle = in sbi_dt_cpu_init_topology()
276 return -ENODEV; in sbi_cpuidle_dt_init_states()
280 ret = -ENOMEM; in sbi_cpuidle_dt_init_states()
284 /* Parse SBI specific details from state DT nodes */ in sbi_cpuidle_dt_init_states()
286 state_node = of_get_cpu_state_node(cpu_node, i - 1); in sbi_cpuidle_dt_init_states()
296 pr_debug("sbi-state %#x index %d\n", states[i], i); in sbi_cpuidle_dt_init_states()
299 ret = -ENODEV; in sbi_cpuidle_dt_init_states()
308 /* Store states in the per-cpu struct. */ in sbi_cpuidle_dt_init_states()
309 data->states = states; in sbi_cpuidle_dt_init_states()
321 dt_idle_detach_cpu(data->dev); in sbi_cpuidle_deinit_cpu()
333 return -ENOMEM; in sbi_cpuidle_init_cpu()
335 drv->name = "sbi_cpuidle"; in sbi_cpuidle_init_cpu()
336 drv->owner = THIS_MODULE; in sbi_cpuidle_init_cpu()
337 drv->cpumask = (struct cpumask *)cpumask_of(cpu); in sbi_cpuidle_init_cpu()
339 /* RISC-V architectural WFI to be represented as state index 0. */ in sbi_cpuidle_init_cpu()
340 drv->states[0].enter = sbi_cpuidle_enter_state; in sbi_cpuidle_init_cpu()
341 drv->states[0].exit_latency = 1; in sbi_cpuidle_init_cpu()
342 drv->states[0].target_residency = 1; in sbi_cpuidle_init_cpu()
343 drv->states[0].power_usage = UINT_MAX; in sbi_cpuidle_init_cpu()
344 strcpy(drv->states[0].name, "WFI"); in sbi_cpuidle_init_cpu()
345 strcpy(drv->states[0].desc, "RISC-V WFI"); in sbi_cpuidle_init_cpu()
348 * If no DT idle states are detected (ret == 0) let the driver in sbi_cpuidle_init_cpu()
350 * initialize the idle driver if only wfi is supported, the in sbi_cpuidle_init_cpu()
351 * default archictectural back-end already executes wfi in sbi_cpuidle_init_cpu()
352 * on idle entry. in sbi_cpuidle_init_cpu()
356 pr_debug("HART%ld: failed to parse DT idle states\n", in sbi_cpuidle_init_cpu()
358 return ret ? : -ENODEV; in sbi_cpuidle_init_cpu()
360 state_count = ret + 1; /* Include WFI state as well */ in sbi_cpuidle_init_cpu()
362 /* Initialize idle states from DT. */ in sbi_cpuidle_init_cpu()
365 pr_err("HART%ld: failed to init idle states\n", in sbi_cpuidle_init_cpu()
385 * All devices have now been attached/probed to the PM domain in sbi_cpuidle_domain_sync_state()
386 * topology, hence it's fine to allow domain states to be picked. in sbi_cpuidle_domain_sync_state()
395 struct genpd_power_state *state = &pd->states[pd->state_idx]; in sbi_cpuidle_pd_power_off() local
398 if (!state->data) in sbi_cpuidle_pd_power_off()
402 return -EBUSY; in sbi_cpuidle_pd_power_off()
404 /* OSI mode is enabled, set the corresponding domain state. */ in sbi_cpuidle_pd_power_off()
405 pd_state = state->data; in sbi_cpuidle_pd_power_off()
423 int ret = -ENOMEM; in sbi_pd_init()
433 pd->flags |= GENPD_FLAG_IRQ_SAFE | GENPD_FLAG_CPU_DOMAIN; in sbi_pd_init()
437 pd->power_off = sbi_cpuidle_pd_power_off; in sbi_pd_init()
439 pd->flags |= GENPD_FLAG_ALWAYS_ON; in sbi_pd_init()
442 pd_gov = pd->states ? &pm_domain_cpu_gov : NULL; in sbi_pd_init()
452 pd_provider->node = of_node_get(np); in sbi_pd_init()
453 list_add(&pd_provider->link, &sbi_pd_providers); in sbi_pd_init()
455 pr_debug("init PM domain %s\n", pd->name); in sbi_pd_init()
465 pr_err("failed to init PM domain ret=%d %pOF\n", ret, np); in sbi_pd_init()
475 of_genpd_del_provider(pd_provider->node); in sbi_pd_remove()
477 genpd = of_genpd_remove_last(pd_provider->node); in sbi_pd_remove()
481 of_node_put(pd_provider->node); in sbi_pd_remove()
482 list_del(&pd_provider->link); in sbi_pd_remove()
493 return -ENODEV; in sbi_genpd_probe()
496 * Parse child nodes for the "#power-domain-cells" property and in sbi_genpd_probe()
497 * initialize a genpd/genpd-of-provider pair when it's found. in sbi_genpd_probe()
500 if (!of_property_present(node, "#power-domain-cells")) in sbi_genpd_probe()
551 of_property_present(np, "power-domains") && in sbi_cpuidle_probe()
552 of_property_present(np, "power-domain-names")) { in sbi_cpuidle_probe()
561 pds_node = of_find_node_by_path("/cpus/power-domains"); in sbi_cpuidle_probe()
569 /* Initialize CPU idle driver for each CPU */ in sbi_cpuidle_probe()
571 ret = sbi_cpuidle_init_cpu(&pdev->dev, cpu); in sbi_cpuidle_probe()
573 pr_debug("HART%ld: idle driver init failed\n", in sbi_cpuidle_probe()
582 pr_info("idle driver registered for all CPUs\n"); in sbi_cpuidle_probe()
587 while (--cpu >= 0) { in sbi_cpuidle_probe()
600 .name = "sbi-cpuidle",
625 pdev = platform_device_register_simple("sbi-cpuidle", in sbi_cpuidle_init()
626 -1, NULL, 0); in sbi_cpuidle_init()