Lines Matching refs:sched_domain

35 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level,  in sched_domain_debug_one()
131 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
171 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
189 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
626 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain()
642 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu()
645 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu()
651 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains()
666 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
670 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
671 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
672 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
678 struct sched_domain *sd; in update_top_cache_domain()
709 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
712 struct sched_domain *tmp; in cpu_attach_domain()
716 struct sched_domain *parent = tmp->parent; in cpu_attach_domain()
772 struct sched_domain * __percpu *sd;
904 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) in build_balance_mask()
908 struct sched_domain *sibling; in build_balance_mask()
941 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain()
964 static void init_overlap_sched_group(struct sched_domain *sd, in init_overlap_sched_group()
992 static struct sched_domain *
993 find_descended_sibling(struct sched_domain *sd, struct sched_domain *sibling) in find_descended_sibling()
1018 build_overlap_sched_groups(struct sched_domain *sd, int cpu) in build_overlap_sched_groups()
1024 struct sched_domain *sibling; in build_overlap_sched_groups()
1185 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1186 struct sched_domain *child = sd->child; in get_group()
1229 build_sched_groups(struct sched_domain *sd, int cpu) in build_sched_groups()
1274 static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) in init_sched_groups_capacity()
1443 static void set_domain_attribute(struct sched_domain *sd, in set_domain_attribute()
1490 d->sd = alloc_percpu(struct sched_domain *); in __visit_domain_allocation_hell()
1505 static void claim_allocations(int cpu, struct sched_domain *sd) in claim_allocations()
1555 static struct sched_domain *
1558 struct sched_domain *child, int cpu) in sd_init()
1561 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in sd_init()
1580 *sd = (struct sched_domain){ in sd_init()
2189 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
2206 struct sched_domain *sd; in __sdt_alloc()
2211 sd = kzalloc_node(sizeof(struct sched_domain) + cpumask_size(), in __sdt_alloc()
2259 struct sched_domain *sd; in __sdt_free()
2286 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain()
2288 struct sched_domain *child, int cpu) in build_sched_domain()
2290 struct sched_domain *sd = sd_init(tl, cpu_map, child, cpu); in build_sched_domain()
2360 struct sched_domain *sd; in build_sched_domains()
2419 struct sched_domain *child = sd->child; in build_sched_domains()
2423 struct sched_domain __rcu *top_p; in build_sched_domains()