Searched refs:sched_group_span (Results 1 – 3 of 3) sorted by relevance
| /linux/kernel/sched/ |
| H A D | fair.c | 7708 return cpumask_first(sched_group_span(group)); in sched_balance_find_dst_group_cpu() 7711 for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { in sched_balance_find_dst_group_cpu() 7980 for_each_cpu_wrap(cpu, sched_group_span(sg), target + 1) { in select_idle_cpu() 7996 cpumask_andnot(cpus, cpus, sched_group_span(sg)); in select_idle_cpu() 10388 for_each_cpu(cpu, sched_group_span(sdg)) { in update_group_capacity() 10713 for_each_cpu_and(i, sched_group_span(group), env->cpus) { in update_sg_lb_stats() 11032 for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { in update_sg_wakeup_stats() 11149 if (!cpumask_intersects(sched_group_span(group), in sched_balance_find_dst_group() 11158 sched_group_span(group)); in sched_balance_find_dst_group() 11256 idlest_cpu = cpumask_first(sched_group_span(idlest)); in sched_balance_find_dst_group() [all …]
|
| H A D | ext_idle.c | 292 return sched_group_span(sg); in numa_span()
|
| H A D | sched.h | 1440 static inline struct cpumask *sched_group_span(struct sched_group *sg); 1531 for_each_cpu_and(cpu, sched_group_span(group), p->cpus_ptr) { in sched_group_cookie_match() 2223 static inline struct cpumask *sched_group_span(struct sched_group *sg) in sched_group_span() function
|