Home
last modified time | relevance | path

Searched refs:cpus_ptr (Results 1 – 19 of 19) sorted by relevance

/Linux-v5.4/kernel/sched/
Dcpupri.c97 if (cpumask_any_and(p->cpus_ptr, vec->mask) >= nr_cpu_ids) in cpupri_find()
101 cpumask_and(lowest_mask, p->cpus_ptr, vec->mask); in cpupri_find()
Dcpudeadline.c123 cpumask_and(later_mask, cp->free_cpus, p->cpus_ptr)) { in cpudl_find()
130 if (cpumask_test_cpu(best_cpu, p->cpus_ptr) && in cpudl_find()
Dcore.c1460 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in is_cpu_allowed()
1656 if (cpumask_equal(p->cpus_ptr, new_mask)) in __set_cpus_allowed_ptr()
1816 if (!cpumask_test_cpu(arg->dst_cpu, arg->src_task->cpus_ptr)) in migrate_swap_stop()
1819 if (!cpumask_test_cpu(arg->src_cpu, arg->dst_task->cpus_ptr)) in migrate_swap_stop()
1861 if (!cpumask_test_cpu(arg.dst_cpu, arg.src_task->cpus_ptr)) in migrate_swap()
1864 if (!cpumask_test_cpu(arg.src_cpu, arg.dst_task->cpus_ptr)) in migrate_swap()
2049 if (cpumask_test_cpu(dest_cpu, p->cpus_ptr)) in select_fallback_rq()
2056 for_each_cpu(dest_cpu, p->cpus_ptr) { in select_fallback_rq()
2110 cpu = cpumask_any(p->cpus_ptr); in select_task_rq()
4924 if (!cpumask_subset(span, p->cpus_ptr) || in __sched_setscheduler()
[all …]
Dfair.c1631 if (!cpumask_test_cpu(env->src_cpu, cur->cpus_ptr)) in task_numa_compare()
1729 if (!cpumask_test_cpu(cpu, env->p->cpus_ptr)) in task_numa_find_cpu()
5576 p->cpus_ptr)) in find_idlest_group()
5703 for_each_cpu_and(i, sched_group_span(group), p->cpus_ptr) { in find_idlest_group_cpu()
5752 if (!cpumask_intersects(sched_domain_span(sd), p->cpus_ptr)) in find_idlest_cpu()
5869 cpumask_and(cpus, sched_domain_span(sd), p->cpus_ptr); in select_idle_core()
5903 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in select_idle_smt()
5969 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in select_idle_cpu()
6009 cpumask_test_cpu(p->recent_used_cpu, p->cpus_ptr)) { in select_idle_sibling()
6372 if (!cpumask_test_cpu(cpu, p->cpus_ptr)) in find_energy_efficient_cpu()
[all …]
Ddeadline.c542 cpu = cpumask_any_and(cpu_active_mask, p->cpus_ptr); in dl_task_offline_migration()
1840 cpumask_test_cpu(cpu, p->cpus_ptr)) in pick_dl_task()
1990 !cpumask_test_cpu(later_rq->cpu, task->cpus_ptr) || in find_lock_later_rq()
Drt.c1604 cpumask_test_cpu(cpu, p->cpus_ptr)) in pick_rt_task()
1741 !cpumask_test_cpu(lowest_rq->cpu, task->cpus_ptr) || in find_lock_lowest_rq()
/Linux-v5.4/lib/
Dsmp_processor_id.c26 if (cpumask_equal(current->cpus_ptr, cpumask_of(this_cpu))) in check_preemption_disabled()
/Linux-v5.4/samples/trace_events/
Dtrace-events-sample.c37 current->cpus_ptr); in simple_thread_func()
/Linux-v5.4/arch/mips/kernel/
Dmips-mt-fpaff.c180 cpumask_or(&allowed, &p->thread.user_cpus_allowed, p->cpus_ptr); in mipsmt_sys_sched_getaffinity()
/Linux-v5.4/init/
Dinit_task.c73 .cpus_ptr = &init_task.cpus_mask,
/Linux-v5.4/fs/proc/
Darray.c384 cpumask_pr_args(task->cpus_ptr)); in task_cpus_allowed()
386 cpumask_pr_args(task->cpus_ptr)); in task_cpus_allowed()
/Linux-v5.4/kernel/trace/
Dtrace_hwlat.c282 if (!cpumask_equal(current_mask, current->cpus_ptr)) in move_to_next_cpu()
/Linux-v5.4/arch/x86/kernel/cpu/resctrl/
Dpseudo_lock.c1502 if (!cpumask_subset(current->cpus_ptr, &plr->d->cpu_mask)) { in pseudo_lock_dev_mmap()
/Linux-v5.4/arch/powerpc/platforms/cell/spufs/
Dsched.c131 cpumask_copy(&ctx->cpus_allowed, current->cpus_ptr); in __spu_update_sched_info()
/Linux-v5.4/drivers/infiniband/hw/hfi1/
Daffinity.c1041 *proc_mask = current->cpus_ptr; in hfi1_get_proc_affinity()
/Linux-v5.4/include/linux/
Dsched.h701 const cpumask_t *cpus_ptr; member
/Linux-v5.4/kernel/
Dfork.c914 if (orig->cpus_ptr == &orig->cpus_mask) in dup_task_struct()
915 tsk->cpus_ptr = &tsk->cpus_mask; in dup_task_struct()
/Linux-v5.4/kernel/cgroup/
Dcpuset.c2865 set_cpus_allowed_ptr(task, current->cpus_ptr); in cpuset_fork()
/Linux-v5.4/drivers/infiniband/hw/qib/
Dqib_file_ops.c1626 const unsigned int cpu = cpumask_first(current->cpus_ptr); in qib_assign_ctxt()