Home
last modified time | relevance | path

Searched refs:task_cpu (Results 1 – 25 of 34) sorted by relevance

12

/Linux-v5.10/arch/x86/um/
Dptrace_32.c198 int err, n, cpu = task_cpu(child); in get_fpregs()
215 int n, cpu = task_cpu(child); in set_fpregs()
228 int err, n, cpu = task_cpu(child); in get_fpxregs()
244 int n, cpu = task_cpu(child); in set_fpxregs()
/Linux-v5.10/arch/ia64/include/asm/
Dswitch_to.h62 (task_cpu(current) != \
64 task_thread_info(current)->last_cpu = task_cpu(current); \
/Linux-v5.10/kernel/rcu/
Dtasks.h452 cpu = task_cpu(t); in check_holdout_task()
858 int cpu = task_cpu(t); in trc_inspect_reader()
932 cpu = task_cpu(t); in trc_wait_for_one_reader()
1014 cpu = task_cpu(t); in show_stalled_task_trace()
Dtree_stall.h455 gpk ? gpk->state : ~0, gpk ? task_cpu(gpk) : -1); in rcu_check_gp_kthread_starvation()
/Linux-v5.10/kernel/sched/
Dstop_task.c16 return task_cpu(p); /* stop tasks as never migrate */ in select_task_rq_stop()
Dcore.c1659 return cpu_curr(task_cpu(p)) == p; in task_curr()
1745 BUG_ON(task_cpu(p) != new_cpu); in move_queued_task()
1926 if (cpumask_test_cpu(task_cpu(p), new_mask)) in __set_cpus_allowed_ptr()
1995 if (task_cpu(p) != new_cpu) { in set_task_cpu()
2058 if (task_cpu(arg->dst_task) != arg->dst_cpu) in migrate_swap_stop()
2061 if (task_cpu(arg->src_task) != arg->src_cpu) in migrate_swap_stop()
2249 cpu = task_cpu(p); in kick_process()
2371 cpu = select_fallback_rq(task_cpu(p), p); in select_task_rq()
2583 if (WARN_ON_ONCE(task_cpu(p) != cpu_of(rq))) in sched_ttwu_pending()
2950 ttwu_queue_wakelist(p, task_cpu(p), wake_flags | WF_ON_CPU)) in try_to_wake_up()
[all …]
Ddeadline.c336 struct dl_bw *dl_b = dl_bw_of(task_cpu(p)); in task_non_contending()
341 __dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in task_non_contending()
1357 struct dl_bw *dl_b = dl_bw_of(task_cpu(p)); in inactive_task_timer()
1366 __dl_sub(dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in inactive_task_timer()
1959 int cpu = task_cpu(task); in find_later_rq()
2109 BUG_ON(rq->cpu != task_cpu(p)); in pick_next_pushable_dl_task()
2335 __dl_sub(src_dl_b, p->dl.dl_bw, dl_bw_cpus(task_cpu(p))); in set_cpus_allowed_dl()
2636 int cpus, err = -1, cpu = task_cpu(p); in sched_dl_overflow()
Didle.c406 return task_cpu(p); /* IDLE tasks as never migrated */ in select_task_rq_idle()
Dcpudeadline.c138 (cpu == task_cpu(p) && cap == max_cap)) { in cpudl_find()
Dpsi.c774 task->pid, task->comm, task_cpu(task), in psi_flags_change()
785 int cpu = task_cpu(task); in psi_task_change()
814 int cpu = task_cpu(prev); in psi_task_switch()
Dfair.c1972 .src_cpu = task_cpu(p), in task_numa_migrate()
2937 int src_nid = cpu_to_node(task_cpu(p)); in update_scan_period()
5688 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_load_without()
5711 if (cpu_of(rq) != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_runnable_without()
6380 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in cpu_util_without()
6465 if (task_cpu(p) == cpu && dst_cpu != cpu) in cpu_util_next()
6467 else if (task_cpu(p) != cpu && dst_cpu == cpu) in cpu_util_next()
8613 if (cpu != task_cpu(p) || !READ_ONCE(p->se.avg.last_update_time)) in task_running_on_cpu()
10949 set_task_rq(p, task_cpu(p)); in task_set_group_fair()
10956 set_task_rq(p, task_cpu(p)); in task_move_group_fair()
Dsched.h1096 #define task_rq(p) cpu_rq(task_cpu(p))
1799 int (*select_task_rq)(struct task_struct *p, int task_cpu, int sd_flag, int flags);
Drt.c1694 int cpu = task_cpu(task); in find_lowest_rq()
1847 BUG_ON(rq->cpu != task_cpu(p)); in pick_next_pushable_task()
/Linux-v5.10/include/linux/
Dkdb.h180 unsigned int cpu = task_cpu(p); in kdb_process_cpu()
Dsched.h1905 static inline unsigned int task_cpu(const struct task_struct *p) in task_cpu() function
1918 static inline unsigned int task_cpu(const struct task_struct *p) in task_cpu() function
/Linux-v5.10/kernel/trace/
Dtrace_sched_wakeup.c396 entry->next_cpu = task_cpu(next); in tracing_sched_switch_trace()
424 entry->next_cpu = task_cpu(wakee); in tracing_sched_wakeup_trace()
569 wakeup_cpu = task_cpu(p); in probe_wakeup()
/Linux-v5.10/include/linux/sched/
Dtopology.h263 return cpu_to_node(task_cpu(p)); in task_node()
/Linux-v5.10/Documentation/scheduler/
Dsched-capacity.rst339 task_util(p) < capacity(task_cpu(p))
402 then it might become CPU-bound, IOW ``task_util(p) > capacity(task_cpu(p))``;
421 task_uclamp_min(p) <= capacity(task_cpu(cpu))
435 task_bandwidth(p) < capacity(task_cpu(p))
/Linux-v5.10/include/trace/events/
Dsched.h76 __entry->target_cpu = task_cpu(p);
207 __entry->orig_cpu = task_cpu(p);
/Linux-v5.10/kernel/locking/
Dmutex.c569 vcpu_is_preempted(task_cpu(owner))) { in mutex_spin_on_owner()
605 retval = owner->on_cpu && !vcpu_is_preempted(task_cpu(owner)); in mutex_can_spin_on_owner()
Drwsem.c652 return owner->on_cpu && !vcpu_is_preempted(task_cpu(owner)); in owner_on_cpu()
/Linux-v5.10/arch/arm64/kernel/
Dsmp.c206 cpu = task_cpu(current); in secondary_start_kernel()
/Linux-v5.10/arch/mips/kernel/
Dprocess.c834 cpumask_set_cpu(task_cpu(t), &process_cpus); in mips_set_process_fp_mode()
/Linux-v5.10/arch/powerpc/kernel/
Dprocess.c2024 unsigned long cpu = task_cpu(p); in valid_irq_stack()
2042 unsigned long cpu = task_cpu(p); in valid_emergency_stack()
/Linux-v5.10/fs/proc/
Darray.c595 seq_put_decimal_ll(m, " ", task_cpu(task)); in do_task_stat()

12