Lines Matching refs:effective_cpus
111 cpumask_var_t effective_cpus; member
412 while (!cpumask_intersects(cs->effective_cpus, pmask)) { in guarantee_online_cpus()
425 cpumask_and(pmask, pmask, cs->effective_cpus); in guarantee_online_cpus()
498 pmask2 = &cs->effective_cpus; in alloc_cpumasks()
533 free_cpumask_var(cs->effective_cpus); in free_cpumasks()
561 cpumask_copy(trial->effective_cpus, cs->effective_cpus); in alloc_trial_cpuset()
674 return cpumask_intersects(a->effective_cpus, b->effective_cpus); in cpusets_overlap()
795 cpumask_and(doms[0], top_cpuset.effective_cpus, in generate_sched_domains()
830 cpumask_subset(cp->cpus_allowed, top_cpuset.effective_cpus)) in generate_sched_domains()
834 !cpumask_empty(cp->effective_cpus)) in generate_sched_domains()
914 cpumask_or(dp, dp, b->effective_cpus); in generate_sched_domains()
974 if (cpumask_empty(cs->effective_cpus)) { in rebuild_root_domains()
1033 !cpumask_equal(top_cpuset.effective_cpus, cpu_active_mask)) in rebuild_sched_domains_locked()
1048 if (!cpumask_subset(cs->effective_cpus, in rebuild_sched_domains_locked()
1093 set_cpus_allowed_ptr(task, cs->effective_cpus); in update_tasks_cpumask()
1112 cpumask_or(new_cpus, parent->effective_cpus, in compute_effective_cpumask()
1117 cpumask_and(new_cpus, cs->cpus_allowed, parent->effective_cpus); in compute_effective_cpumask()
1207 (!cpumask_subset(cpuset->cpus_allowed, parent->effective_cpus) || in update_parent_subparts_cpumask()
1208 cpumask_equal(cpuset->cpus_allowed, parent->effective_cpus))) in update_parent_subparts_cpumask()
1234 cpumask_and(tmp->addmask, newmask, parent->effective_cpus); in update_parent_subparts_cpumask()
1241 cpumask_equal(parent->effective_cpus, tmp->addmask)) { in update_parent_subparts_cpumask()
1252 cpumask_copy(tmp->addmask, parent->effective_cpus); in update_parent_subparts_cpumask()
1265 parent->effective_cpus); in update_parent_subparts_cpumask()
1267 parent->effective_cpus); in update_parent_subparts_cpumask()
1317 cpumask_andnot(parent->effective_cpus, in update_parent_subparts_cpumask()
1318 parent->effective_cpus, tmp->addmask); in update_parent_subparts_cpumask()
1327 cpumask_or(parent->effective_cpus, in update_parent_subparts_cpumask()
1328 parent->effective_cpus, tmp->delmask); in update_parent_subparts_cpumask()
1372 cpumask_copy(tmp->new_cpus, parent->effective_cpus); in update_cpumasks_hier()
1388 cpumask_equal(tmp->new_cpus, cp->effective_cpus)) { in update_cpumasks_hier()
1442 cpumask_copy(cp->effective_cpus, tmp->new_cpus); in update_cpumasks_hier()
1456 cpumask_andnot(cp->effective_cpus, cp->effective_cpus, in update_cpumasks_hier()
1458 if (cpumask_empty(cp->effective_cpus)) { in update_cpumasks_hier()
1459 cpumask_copy(cp->effective_cpus, tmp->new_cpus); in update_cpumasks_hier()
1478 !cpumask_equal(cp->cpus_allowed, cp->effective_cpus)); in update_cpumasks_hier()
1580 tmp.delmask = trialcs->effective_cpus; in update_cpumask()
2486 seq_printf(sf, "%*pbl\n", cpumask_pr_args(cs->effective_cpus)); in cpuset_common_seq_show()
2818 cpumask_copy(cs->effective_cpus, parent->effective_cpus); in cpuset_css_online()
2854 cpumask_copy(cs->effective_cpus, parent->cpus_allowed); in cpuset_css_online()
2918 top_cpuset.effective_cpus); in cpuset_bind()
2968 BUG_ON(!alloc_cpumask_var(&top_cpuset.effective_cpus, GFP_KERNEL)); in cpuset_init()
2973 cpumask_setall(top_cpuset.effective_cpus); in cpuset_init()
3021 cpumask_copy(cs->effective_cpus, new_cpus); in hotplug_update_tasks_legacy()
3057 cpumask_copy(new_cpus, parent_cs(cs)->effective_cpus); in hotplug_update_tasks()
3062 cpumask_copy(cs->effective_cpus, new_cpus); in hotplug_update_tasks()
3173 cpus_updated = !cpumask_equal(&new_cpus, cs->effective_cpus); in cpuset_hotplug_update_tasks()
3224 cpus_updated = !cpumask_equal(top_cpuset.effective_cpus, &new_cpus); in cpuset_hotplug_workfn()
3255 cpumask_copy(top_cpuset.effective_cpus, &new_cpus); in cpuset_hotplug_workfn()
3343 cpumask_copy(top_cpuset.effective_cpus, cpu_active_mask); in cpuset_init_smp()