Lines Matching refs:tl
1322 sd_init(struct sched_domain_topology_level *tl, in sd_init() argument
1326 struct sd_data *sdd = &tl->data; in sd_init()
1334 sched_domains_curr_level = tl->numa_level; in sd_init()
1337 sd_weight = cpumask_weight(tl->mask(cpu)); in sd_init()
1339 if (tl->sd_flags) in sd_init()
1340 sd_flags = (*tl->sd_flags)(); in sd_init()
1375 .name = tl->name, in sd_init()
1379 cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); in sd_init()
1403 if (sched_domains_numa_distance[tl->numa_level] > node_reclaim_distance) { in sd_init()
1446 #define for_each_sd_topology(tl) \ argument
1447 for (tl = sched_domain_topology; tl->mask; tl++)
1449 void set_sched_topology(struct sched_domain_topology_level *tl) in set_sched_topology() argument
1454 sched_domain_topology = tl; in set_sched_topology()
1555 struct sched_domain_topology_level *tl; in sched_init_numa() local
1661 tl = kzalloc((i + level + 1) * in sched_init_numa()
1663 if (!tl) in sched_init_numa()
1670 tl[i] = sched_domain_topology[i]; in sched_init_numa()
1675 tl[i++] = (struct sched_domain_topology_level){ in sched_init_numa()
1685 tl[i] = (struct sched_domain_topology_level){ in sched_init_numa()
1694 sched_domain_topology = tl; in sched_init_numa()
1749 struct sched_domain_topology_level *tl; in __sdt_alloc() local
1752 for_each_sd_topology(tl) { in __sdt_alloc()
1753 struct sd_data *sdd = &tl->data; in __sdt_alloc()
1818 struct sched_domain_topology_level *tl; in __sdt_free() local
1821 for_each_sd_topology(tl) { in __sdt_free()
1822 struct sd_data *sdd = &tl->data; in __sdt_free()
1852 static struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain() argument
1856 struct sched_domain *sd = sd_init(tl, cpu_map, child, dflags, cpu); in build_sched_domain()
1886 static bool topology_span_sane(struct sched_domain_topology_level *tl, in topology_span_sane() argument
1892 if (tl->flags & SDTL_OVERLAP) in topology_span_sane()
1910 if (!cpumask_equal(tl->mask(cpu), tl->mask(i)) && in topology_span_sane()
1911 cpumask_intersects(tl->mask(cpu), tl->mask(i))) in topology_span_sane()
1927 struct sched_domain_topology_level *tl, *asym_tl = NULL; in asym_cpu_capacity_level() local
1952 for_each_sd_topology(tl) { in asym_cpu_capacity_level()
1956 for_each_cpu_and(j, tl->mask(i), cpu_map) { in asym_cpu_capacity_level()
1966 asym_tl = tl; in asym_cpu_capacity_level()
2003 struct sched_domain_topology_level *tl; in build_sched_domains() local
2007 for_each_sd_topology(tl) { in build_sched_domains()
2008 if (tl == tl_asym) { in build_sched_domains()
2013 if (WARN_ON(!topology_span_sane(tl, cpu_map, i))) in build_sched_domains()
2016 sd = build_sched_domain(tl, cpu_map, attr, sd, dflags, i); in build_sched_domains()
2018 if (tl == sched_domain_topology) in build_sched_domains()
2020 if (tl->flags & SDTL_OVERLAP) in build_sched_domains()