Lines Matching refs:sched_domain

35 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level,  in sched_domain_debug_one()
131 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
171 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
189 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
626 static void destroy_sched_domain(struct sched_domain *sd) in destroy_sched_domain()
642 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in destroy_sched_domains_rcu()
645 struct sched_domain *parent = sd->parent; in destroy_sched_domains_rcu()
651 static void destroy_sched_domains(struct sched_domain *sd) in destroy_sched_domains()
666 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_llc);
670 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_numa);
671 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_packing);
672 DEFINE_PER_CPU(struct sched_domain __rcu *, sd_asym_cpucapacity);
678 struct sched_domain *sd; in update_top_cache_domain()
709 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
712 struct sched_domain *tmp; in cpu_attach_domain()
716 struct sched_domain *parent = tmp->parent; in cpu_attach_domain()
768 struct sched_domain * __percpu *sd;
900 build_balance_mask(struct sched_domain *sd, struct sched_group *sg, struct cpumask *mask) in build_balance_mask()
904 struct sched_domain *sibling; in build_balance_mask()
937 build_group_from_child_sched_domain(struct sched_domain *sd, int cpu) in build_group_from_child_sched_domain()
960 static void init_overlap_sched_group(struct sched_domain *sd, in init_overlap_sched_group()
988 static struct sched_domain *
989 find_descended_sibling(struct sched_domain *sd, struct sched_domain *sibling) in find_descended_sibling()
1014 build_overlap_sched_groups(struct sched_domain *sd, int cpu) in build_overlap_sched_groups()
1020 struct sched_domain *sibling; in build_overlap_sched_groups()
1181 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
1182 struct sched_domain *child = sd->child; in get_group()
1225 build_sched_groups(struct sched_domain *sd, int cpu) in build_sched_groups()
1270 static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) in init_sched_groups_capacity()
1429 static void set_domain_attribute(struct sched_domain *sd, in set_domain_attribute()
1476 d->sd = alloc_percpu(struct sched_domain *); in __visit_domain_allocation_hell()
1491 static void claim_allocations(int cpu, struct sched_domain *sd) in claim_allocations()
1541 static struct sched_domain *
1544 struct sched_domain *child, int cpu) in sd_init()
1547 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in sd_init()
1566 *sd = (struct sched_domain){ in sd_init()
2175 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
2192 struct sched_domain *sd; in __sdt_alloc()
2197 sd = kzalloc_node(sizeof(struct sched_domain) + cpumask_size(), in __sdt_alloc()
2245 struct sched_domain *sd; in __sdt_free()
2272 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain()
2274 struct sched_domain *child, int cpu) in build_sched_domain()
2276 struct sched_domain *sd = sd_init(tl, cpu_map, child, cpu); in build_sched_domain()
2346 struct sched_domain *sd; in build_sched_domains()
2405 struct sched_domain *child = sd->child; in build_sched_domains()
2409 struct sched_domain __rcu *top_p; in build_sched_domains()