Lines Matching refs:numgrps

117 static void alloc_groups_to_nodes(unsigned int numgrps,  in alloc_groups_to_nodes()  argument
201 WARN_ON_ONCE(numgrps == 0); in alloc_groups_to_nodes()
205 numgrps * ncpus / remaining_ncpus); in alloc_groups_to_nodes()
211 numgrps -= ngroups; in alloc_groups_to_nodes()
228 static void alloc_nodes_groups(unsigned int numgrps, in alloc_nodes_groups() argument
254 numgrps = min_t(unsigned int, numcpus, numgrps); in alloc_nodes_groups()
255 alloc_groups_to_nodes(numgrps, numcpus, node_groups, nr_node_ids); in alloc_nodes_groups()
404 static int __group_cpus_evenly(unsigned int startgrp, unsigned int numgrps, in __group_cpus_evenly() argument
410 unsigned int last_grp = numgrps; in __group_cpus_evenly()
424 if (numgrps <= nodes) { in __group_cpus_evenly()
432 return numgrps; in __group_cpus_evenly()
440 alloc_nodes_groups(numgrps, node_to_cpumask, cpu_mask, in __group_cpus_evenly()
488 struct cpumask *group_cpus_evenly(unsigned int numgrps, unsigned int *nummasks) in group_cpus_evenly() argument
496 if (numgrps == 0) in group_cpus_evenly()
509 masks = kzalloc_objs(*masks, numgrps); in group_cpus_evenly()
530 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()
542 if (nr_present >= numgrps) in group_cpus_evenly()
547 ret = __group_cpus_evenly(curgrp, numgrps, node_to_cpumask, in group_cpus_evenly()
564 *nummasks = min(nr_present + nr_others, numgrps); in group_cpus_evenly()
568 struct cpumask *group_cpus_evenly(unsigned int numgrps, unsigned int *nummasks) in group_cpus_evenly() argument
572 if (numgrps == 0) in group_cpus_evenly()
575 masks = kzalloc_objs(*masks, numgrps); in group_cpus_evenly()