Lines Matching refs:curgrp
262 unsigned int *curgrp, in assign_cpus_to_groups() argument
270 for (v = 0; v < nv->ngroups; v++, *curgrp += 1) { in assign_cpus_to_groups()
283 if (*curgrp >= last_grp) in assign_cpus_to_groups()
284 *curgrp = 0; in assign_cpus_to_groups()
285 grp_spread_init_one(&masks[*curgrp], nmsk, cpus_per_grp); in assign_cpus_to_groups()
363 unsigned int *curgrp, in __try_group_cluster_cpus() argument
392 assign_cpus_to_groups(nc, nmsk, nv, masks, curgrp, last_grp); in __try_group_cluster_cpus()
411 unsigned int curgrp = startgrp; in __group_cpus_evenly() local
428 cpumask_or(&masks[curgrp], &masks[curgrp], nmsk); in __group_cpus_evenly()
429 if (++curgrp == last_grp) in __group_cpus_evenly()
430 curgrp = 0; in __group_cpus_evenly()
458 masks, &curgrp, last_grp)) { in __group_cpus_evenly()
463 assign_cpus_to_groups(ncpus, nmsk, nv, masks, &curgrp, in __group_cpus_evenly()
490 unsigned int curgrp = 0, nr_present = 0, nr_others = 0; in group_cpus_evenly() local
530 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()
543 curgrp = 0; in group_cpus_evenly()
545 curgrp = nr_present; in group_cpus_evenly()
547 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()