Lines Matching refs:task_rq
86 rq = task_rq(p); in __task_rq_lock()
88 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) { in __task_rq_lock()
110 rq = task_rq(p); in task_rq_lock()
129 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) { in task_rq_lock()
1567 if (task_rq(p) == rq) { in migration_cpu_stop()
1592 struct rq *rq = task_rq(p); in do_set_cpus_allowed()
1737 lockdep_is_held(&task_rq(p)->lock))); in set_task_cpu()
1765 src_rq = task_rq(p); in __migrate_swap_task()
1905 rq = task_rq(p); in wait_task_inactive()
2624 atomic_dec(&task_rq(p)->nr_iowait); in try_to_wake_up()
2638 atomic_dec(&task_rq(p)->nr_iowait); in try_to_wake_up()
3535 struct sched_entity *curr = (&task_rq(p)->cfs)->curr; in prefetch_curr_exec_start()
5442 if (!cpumask_subset(task_rq(p)->rd->span, new_mask)) { in sched_setaffinity()
5704 p_rq = task_rq(p); in yield_to()
5715 if (task_rq(p) != p_rq) { in yield_to()
6113 if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span, in task_can_attach()
6284 if (WARN_ON(task_rq(next) != rq || !task_on_rq_queued(next))) { in migrate_tasks()