Searched refs:sched_group_span (Results 1 – 3 of 3) sorted by relevance
/Linux-v4.19/kernel/sched/ |
D | topology.c | 50 if (group && !cpumask_test_cpu(cpu, sched_group_span(group))) { in sched_domain_debug_one() 62 if (!cpumask_weight(sched_group_span(group))) { in sched_domain_debug_one() 69 cpumask_intersects(groupmask, sched_group_span(group))) { in sched_domain_debug_one() 75 cpumask_or(groupmask, groupmask, sched_group_span(group)); in sched_domain_debug_one() 79 cpumask_pr_args(sched_group_span(group))); in sched_domain_debug_one() 82 !cpumask_equal(group_balance_mask(group), sched_group_span(group))) { in sched_domain_debug_one() 92 sched_group_span(group))) { in sched_domain_debug_one() 614 const struct cpumask *sg_span = sched_group_span(sg); in build_balance_mask() 660 sg_span = sched_group_span(sg); in build_group_from_child_sched_domain() 679 cpu = cpumask_first_and(sched_group_span(sg), mask); in init_overlap_sched_group() [all …]
|
D | fair.c | 5713 if (!cpumask_intersects(sched_group_span(group), in find_idlest_group() 5718 sched_group_span(group)); in find_idlest_group() 5728 for_each_cpu(i, sched_group_span(group)) { in find_idlest_group() 5843 return cpumask_first(sched_group_span(group)); in find_idlest_group_cpu() 5846 for_each_cpu_and(i, sched_group_span(group), &p->cpus_allowed) { in find_idlest_group_cpu() 7649 for_each_cpu(cpu, sched_group_span(sdg)) { in update_group_capacity() 7850 for_each_cpu_and(i, sched_group_span(group), env->cpus) { in update_sg_lb_stats() 8022 local_group = cpumask_test_cpu(env->dst_cpu, sched_group_span(sg)); in update_sd_lb_stats() 8390 for_each_cpu_and(i, sched_group_span(group), env->cpus) { in find_busiest_queue() 8550 .dst_grpmask = sched_group_span(sd->groups), in load_balance()
|
D | sched.h | 1228 static inline struct cpumask *sched_group_span(struct sched_group *sg) in sched_group_span() function 1247 return cpumask_first(sched_group_span(group)); in group_first_cpu()
|