Lines Matching refs:cpu_map

276 static void perf_domain_debug(const struct cpumask *cpu_map,  in perf_domain_debug()  argument
282 printk(KERN_DEBUG "root_domain %*pbl:", cpumask_pr_args(cpu_map)); in perf_domain_debug()
342 static bool build_perf_domains(const struct cpumask *cpu_map) in build_perf_domains() argument
344 int i, nr_pd = 0, nr_cs = 0, nr_cpus = cpumask_weight(cpu_map); in build_perf_domains()
346 int cpu = cpumask_first(cpu_map); in build_perf_domains()
358 cpumask_pr_args(cpu_map)); in build_perf_domains()
363 for_each_cpu(i, cpu_map) { in build_perf_domains()
377 cpumask_pr_args(cpu_map)); in build_perf_domains()
399 cpumask_pr_args(cpu_map)); in build_perf_domains()
403 perf_domain_debug(cpu_map, pd); in build_perf_domains()
1217 static void __sdt_free(const struct cpumask *cpu_map);
1218 static int __sdt_alloc(const struct cpumask *cpu_map);
1221 const struct cpumask *cpu_map) in __free_domain_allocs() argument
1232 __sdt_free(cpu_map); in __free_domain_allocs()
1240 __visit_domain_allocation_hell(struct s_data *d, const struct cpumask *cpu_map) in __visit_domain_allocation_hell() argument
1244 if (__sdt_alloc(cpu_map)) in __visit_domain_allocation_hell()
1316 const struct cpumask *cpu_map, in sd_init() argument
1373 cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); in sd_init()
1750 static int __sdt_alloc(const struct cpumask *cpu_map) in __sdt_alloc() argument
1774 for_each_cpu(j, cpu_map) { in __sdt_alloc()
1819 static void __sdt_free(const struct cpumask *cpu_map) in __sdt_free() argument
1827 for_each_cpu(j, cpu_map) { in __sdt_free()
1856 const struct cpumask *cpu_map, struct sched_domain_attr *attr, in build_sched_domain() argument
1859 struct sched_domain *sd = sd_init(tl, cpu_map, child, dflags, cpu); in build_sched_domain()
1890 *asym_cpu_capacity_level(const struct cpumask *cpu_map) in asym_cpu_capacity_level() argument
1898 cap = arch_scale_cpu_capacity(cpumask_first(cpu_map)); in asym_cpu_capacity_level()
1900 for_each_cpu(i, cpu_map) { in asym_cpu_capacity_level()
1915 for_each_cpu(i, cpu_map) { in asym_cpu_capacity_level()
1923 for_each_cpu_and(j, tl->mask(i), cpu_map) { in asym_cpu_capacity_level()
1949 build_sched_domains(const struct cpumask *cpu_map, struct sched_domain_attr *attr) in build_sched_domains() argument
1959 if (WARN_ON(cpumask_empty(cpu_map))) in build_sched_domains()
1962 alloc_state = __visit_domain_allocation_hell(&d, cpu_map); in build_sched_domains()
1966 tl_asym = asym_cpu_capacity_level(cpu_map); in build_sched_domains()
1969 for_each_cpu(i, cpu_map) { in build_sched_domains()
1981 sd = build_sched_domain(tl, cpu_map, attr, sd, dflags, i); in build_sched_domains()
1987 if (cpumask_equal(cpu_map, sched_domain_span(sd))) in build_sched_domains()
1993 for_each_cpu(i, cpu_map) { in build_sched_domains()
2008 if (!cpumask_test_cpu(i, cpu_map)) in build_sched_domains()
2019 for_each_cpu(i, cpu_map) { in build_sched_domains()
2036 cpumask_pr_args(cpu_map), rq->rd->max_cpu_capacity); in build_sched_domains()
2041 __free_domain_allocs(&d, alloc_state, cpu_map); in build_sched_domains()
2101 int sched_init_domains(const struct cpumask *cpu_map) in sched_init_domains() argument
2114 cpumask_and(doms_cur[0], cpu_map, housekeeping_cpumask(HK_FLAG_DOMAIN)); in sched_init_domains()
2125 static void detach_destroy_domains(const struct cpumask *cpu_map) in detach_destroy_domains() argument
2127 unsigned int cpu = cpumask_any(cpu_map); in detach_destroy_domains()
2134 for_each_cpu(i, cpu_map) in detach_destroy_domains()