Lines Matching refs:numgrps
130 static void alloc_nodes_groups(unsigned int numgrps, in alloc_nodes_groups() argument
156 numgrps = min_t(unsigned, remaining_ncpus, numgrps); in alloc_nodes_groups()
235 WARN_ON_ONCE(numgrps == 0); in alloc_nodes_groups()
239 numgrps * ncpus / remaining_ncpus); in alloc_nodes_groups()
245 numgrps -= ngroups; in alloc_nodes_groups()
249 static int __group_cpus_evenly(unsigned int startgrp, unsigned int numgrps, in __group_cpus_evenly() argument
255 unsigned int last_grp = numgrps; in __group_cpus_evenly()
269 if (numgrps <= nodes) { in __group_cpus_evenly()
277 return numgrps; in __group_cpus_evenly()
287 alloc_nodes_groups(numgrps, node_to_cpumask, cpu_mask, in __group_cpus_evenly()
347 struct cpumask *group_cpus_evenly(unsigned int numgrps) in group_cpus_evenly() argument
365 masks = kcalloc(numgrps, sizeof(*masks), GFP_KERNEL); in group_cpus_evenly()
374 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()
386 if (nr_present >= numgrps) in group_cpus_evenly()
391 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()
400 WARN_ON(nr_present + nr_others < numgrps); in group_cpus_evenly()
417 struct cpumask *group_cpus_evenly(unsigned int numgrps) in group_cpus_evenly() argument
419 struct cpumask *masks = kcalloc(numgrps, sizeof(*masks), GFP_KERNEL); in group_cpus_evenly()