Lines Matching refs:sched_domain

35 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level,  in sched_domain_debug_one()
131 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
171 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
189 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
635 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain()
651 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu()
654 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu()
660 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains()
675 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
680 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
681 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
682 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
690 struct sched_domain *sd; in update_top_cache_domain()
732 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
735 struct sched_domain *tmp; in cpu_attach_domain()
739 struct sched_domain *parent = tmp->parent; in cpu_attach_domain()
795 struct sched_domain * __percpu *sd;
927 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) in build_balance_mask()
931 struct sched_domain *sibling; in build_balance_mask()
964 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain()
987 static void init_overlap_sched_group(struct sched_domain *sd, in init_overlap_sched_group()
1015 static struct sched_domain *
1016 find_descended_sibling(struct sched_domain *sd, struct sched_domain *sibling) in find_descended_sibling()
1041 build_overlap_sched_groups(struct sched_domain *sd, int cpu) in build_overlap_sched_groups()
1047 struct sched_domain *sibling; in build_overlap_sched_groups()
1208 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1209 struct sched_domain *child = sd->child; in get_group()
1252 build_sched_groups(struct sched_domain *sd, int cpu) in build_sched_groups()
1297 static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) in init_sched_groups_capacity()
1473 static void set_domain_attribute(struct sched_domain *sd, in set_domain_attribute()
1520 d->sd = alloc_percpu(struct sched_domain *); in __visit_domain_allocation_hell()
1535 static void claim_allocations(int cpu, struct sched_domain *sd) in claim_allocations()
1587 static struct sched_domain *
1590 struct sched_domain *child, int cpu) in sd_init()
1593 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in sd_init()
1612 *sd = (struct sched_domain){ in sd_init()
2230 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
2247 struct sched_domain *sd; in __sdt_alloc()
2252 sd = kzalloc_node(sizeof(struct sched_domain) + cpumask_size(), in __sdt_alloc()
2300 struct sched_domain *sd; in __sdt_free()
2327 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain()
2329 struct sched_domain *child, int cpu) in build_sched_domain()
2331 struct sched_domain *sd = sd_init(tl, cpu_map, child, cpu); in build_sched_domain()
2399 struct sched_domain *sd; in build_sched_domains()
2459 struct sched_domain *child = sd->child; in build_sched_domains()
2463 struct sched_domain __rcu *top_p; in build_sched_domains()