Lines Matching full:covered
423 /* Skip already covered CPUs. */ in build_perf_domains()
1029 struct cpumask *covered = sched_domains_tmpmask; in build_overlap_sched_groups() local
1034 cpumask_clear(covered); in build_overlap_sched_groups()
1039 if (cpumask_test_cpu(i, covered)) in build_overlap_sched_groups()
1097 cpumask_or(covered, covered, sg_span); in build_overlap_sched_groups()
1230 * covered by the given span, will set each group's ->cpumask correctly,
1241 struct cpumask *covered; in build_sched_groups() local
1245 covered = sched_domains_tmpmask; in build_sched_groups()
1247 cpumask_clear(covered); in build_sched_groups()
1252 if (cpumask_test_cpu(i, covered)) in build_sched_groups()
1257 cpumask_or(covered, covered, sched_group_span(sg)); in build_sched_groups()
2426 struct cpumask *covered, *id_seen; in topology_span_sane() local
2430 covered = sched_domains_tmpmask; in topology_span_sane()
2443 cpumask_clear(covered); in topology_span_sane()
2465 if (cpumask_intersects(tl_cpu_mask, covered)) in topology_span_sane()
2468 cpumask_or(covered, covered, tl_cpu_mask); in topology_span_sane()