Lines Matching refs:cg

411     struct cpu_group *cg, int indent);
757 cpu_search_lowest(const struct cpu_group *cg, const struct cpu_search *s, in cpu_search_lowest() argument
769 if (cg->cg_children > 0) { in cpu_search_lowest()
770 for (c = cg->cg_children - 1; c >= 0; c--) { in cpu_search_lowest()
771 load = cpu_search_lowest(&cg->cg_child[c], s, &lr); in cpu_search_lowest()
780 (cg->cg_child[c].cg_flags & CG_FLAG_THREAD) && in cpu_search_lowest()
795 for (c = cg->cg_last; c >= cg->cg_first; c--) { in cpu_search_lowest()
796 if (!CPU_ISSET(c, &cg->cg_mask)) in cpu_search_lowest()
839 cpu_search_highest(const struct cpu_group *cg, const struct cpu_search *s, in cpu_search_highest() argument
851 if (cg->cg_children > 0) { in cpu_search_highest()
852 for (c = cg->cg_children - 1; c >= 0; c--) { in cpu_search_highest()
853 load = cpu_search_highest(&cg->cg_child[c], s, &lr); in cpu_search_highest()
866 for (c = cg->cg_last; c >= cg->cg_first; c--) { in cpu_search_highest()
867 if (!CPU_ISSET(c, &cg->cg_mask)) in cpu_search_highest()
897 sched_lowest(const struct cpu_group *cg, cpuset_t *mask, int pri, int maxload, in sched_lowest() argument
908 cpu_search_lowest(cg, &s, &r); in sched_lowest()
916 sched_highest(const struct cpu_group *cg, cpuset_t *mask, int minload, in sched_highest() argument
925 cpu_search_highest(cg, &s, &r); in sched_highest()
930 sched_balance_group(struct cpu_group *cg) in sched_balance_group() argument
939 high = sched_highest(cg, &hmask, 1, 0); in sched_balance_group()
971 low = sched_lowest(cg, &lmask, -1, TDQ_LOAD(tdq) - 1, high, 1); in sched_balance_group()
1109 struct cpu_group *cg, *parent; in tdq_idled() local
1120 for (cg = tdq->tdq_cg, goup = 0; ; ) { in tdq_idled()
1121 cpu = sched_highest(cg, &mask, steal_thresh, 1); in tdq_idled()
1137 cg = cg->cg_parent; in tdq_idled()
1140 parent = cg->cg_parent; in tdq_idled()
1144 if (cg == &parent->cg_child[0]) in tdq_idled()
1145 cg = &parent->cg_child[1]; in tdq_idled()
1147 cg = &parent->cg_child[0]; in tdq_idled()
1150 cg = parent; in tdq_idled()
1408 struct cpu_group *cg, *ccg; in sched_pickcpu() local
1438 cg = tdq->tdq_cg; in sched_pickcpu()
1443 cg = tdq->tdq_cg; in sched_pickcpu()
1452 if (cg->cg_flags & CG_FLAG_THREAD) { in sched_pickcpu()
1454 for (cpu = cg->cg_first; cpu <= cg->cg_last; cpu++) { in sched_pickcpu()
1457 if (CPU_ISSET(cpu, &cg->cg_mask) && in sched_pickcpu()
1461 if (cpu > cg->cg_last) { in sched_pickcpu()
1476 for (ccg = NULL; cg != NULL; cg = cg->cg_parent) { in sched_pickcpu()
1477 if (cg->cg_flags & CG_FLAG_THREAD) in sched_pickcpu()
1479 if (cg->cg_children == 1 || cg->cg_count == 1) in sched_pickcpu()
1481 if (cg->cg_level == CG_SHARE_NONE || in sched_pickcpu()
1482 (!intr && !SCHED_AFFINITY(ts, cg->cg_level))) in sched_pickcpu()
1484 ccg = cg; in sched_pickcpu()
2178 struct cpu_group *cg, *parent; in tdq_trysteal() local
2191 for (i = 1, cg = tdq->tdq_cg, goup = 0; ; ) { in tdq_trysteal()
2192 cpu = sched_highest(cg, &mask, steal_thresh, 1); in tdq_trysteal()
2210 cg = cg->cg_parent; in tdq_trysteal()
2217 parent = cg->cg_parent; in tdq_trysteal()
2223 if (cg == &parent->cg_child[0]) in tdq_trysteal()
2224 cg = &parent->cg_child[1]; in tdq_trysteal()
2226 cg = &parent->cg_child[0]; in tdq_trysteal()
2229 cg = parent; in tdq_trysteal()
3368 sysctl_kern_sched_topology_spec_internal(struct sbuf *sb, struct cpu_group *cg, in sysctl_kern_sched_topology_spec_internal() argument
3375 "", 1 + indent / 2, cg->cg_level); in sysctl_kern_sched_topology_spec_internal()
3377 cg->cg_count, cpusetobj_strprint(cpusetbuf, &cg->cg_mask)); in sysctl_kern_sched_topology_spec_internal()
3379 for (i = cg->cg_first; i <= cg->cg_last; i++) { in sysctl_kern_sched_topology_spec_internal()
3380 if (CPU_ISSET(i, &cg->cg_mask)) { in sysctl_kern_sched_topology_spec_internal()
3390 if (cg->cg_flags != 0) { in sysctl_kern_sched_topology_spec_internal()
3392 if ((cg->cg_flags & CG_FLAG_HTT) != 0) in sysctl_kern_sched_topology_spec_internal()
3394 if ((cg->cg_flags & CG_FLAG_THREAD) != 0) in sysctl_kern_sched_topology_spec_internal()
3396 if ((cg->cg_flags & CG_FLAG_SMT) != 0) in sysctl_kern_sched_topology_spec_internal()
3398 if ((cg->cg_flags & CG_FLAG_NODE) != 0) in sysctl_kern_sched_topology_spec_internal()
3403 if (cg->cg_children > 0) { in sysctl_kern_sched_topology_spec_internal()
3405 for (i = 0; i < cg->cg_children; i++) in sysctl_kern_sched_topology_spec_internal()
3407 &cg->cg_child[i], indent+2); in sysctl_kern_sched_topology_spec_internal()