Lines Matching refs:cpu_rq

339 		raw_spin_lock_nested(&cpu_rq(t)->__lock, i++);  in sched_core_lock()
348 raw_spin_unlock(&cpu_rq(t)->__lock); in sched_core_unlock()
369 cpu_rq(t)->core_enabled = enabled; in __sched_core_flip()
371 cpu_rq(cpu)->core->core_forceidle_start = 0; in __sched_core_flip()
382 cpu_rq(cpu)->core_enabled = enabled; in __sched_core_flip()
392 WARN_ON_ONCE(!RB_EMPTY_ROOT(&cpu_rq(cpu)->core_tree)); in sched_core_assert_empty()
1067 struct rq *rq = cpu_rq(cpu); in resched_cpu()
1130 struct rq *rq = cpu_rq(cpu); in wake_up_idle_cpu()
2034 init_uclamp_rq(cpu_rq(cpu)); in init_uclamp()
2525 rq = cpu_rq(new_cpu); in move_queued_task()
3404 dst_rq = cpu_rq(cpu); in __migrate_swap_task()
3440 src_rq = cpu_rq(arg->src_cpu); in migrate_swap_stop()
3441 dst_rq = cpu_rq(arg->dst_cpu); in migrate_swap_stop()
3655 struct task_struct *old_stop = cpu_rq(cpu)->stop; in sched_set_stop_task()
3685 cpu_rq(cpu)->stop = stop; in sched_set_stop_task()
3915 if (set_nr_if_polling(cpu_rq(cpu)->idle)) { in call_function_single_prep_ipi()
3931 struct rq *rq = cpu_rq(cpu); in __ttwu_queue_wakelist()
3941 struct rq *rq = cpu_rq(cpu); in wake_up_if_idle()
3993 if (!cpu_rq(cpu)->nr_running) in ttwu_queue_cond()
4021 struct rq *rq = cpu_rq(cpu); in ttwu_queue()
5399 sum += cpu_rq(i)->nr_running; in nr_running()
5425 return cpu_rq(cpu)->nr_switches; in nr_context_switches_cpu()
5434 sum += cpu_rq(i)->nr_switches; in nr_context_switches()
5448 return atomic_read(&cpu_rq(cpu)->nr_iowait); in nr_iowait_cpu()
5641 struct rq *rq = cpu_rq(cpu); in scheduler_tick()
5722 struct rq *rq = cpu_rq(cpu); in sched_tick_remote()
6176 rq_i = cpu_rq(i); in pick_next_task()
6198 rq_i = cpu_rq(i); in pick_next_task()
6243 rq_i = cpu_rq(i); in pick_next_task()
6295 struct rq *dst = cpu_rq(this), *src = cpu_rq(that); in try_steal_cookie()
6410 struct rq *rq = cpu_rq(cpu), *core_rq = NULL; in sched_core_cpu_starting()
6425 rq = cpu_rq(t); in sched_core_cpu_starting()
6437 rq = cpu_rq(t); in sched_core_cpu_starting()
6449 struct rq *rq = cpu_rq(cpu), *core_rq = NULL; in sched_core_cpu_deactivate()
6468 core_rq = cpu_rq(t); in sched_core_cpu_deactivate()
6492 rq = cpu_rq(t); in sched_core_cpu_deactivate()
6499 struct rq *rq = cpu_rq(cpu); in sched_core_cpu_dying()
6586 rq = cpu_rq(cpu); in __schedule()
7327 struct rq *rq = cpu_rq(cpu); in idle_cpu()
7368 return cpu_rq(cpu)->idle; in idle_task()
7374 struct rq *rq = cpu_rq(cpu); in sched_core_idle_cpu()
7410 struct rq *rq = cpu_rq(cpu); in effective_cpu_util()
9258 struct rq *rq = cpu_rq(cpu); in init_idle()
9521 struct rq *rq = cpu_rq(cpu); in balance_push_set()
9648 struct rq *rq = cpu_rq(cpu); in sched_cpu_activate()
9693 struct rq *rq = cpu_rq(cpu); in sched_cpu_deactivate()
9759 struct rq *rq = cpu_rq(cpu); in sched_rq_cpu_starting()
9830 struct rq *rq = cpu_rq(cpu); in sched_cpu_dying()
9976 rq = cpu_rq(i); in sched_init()
11833 src_rq = cpu_rq(src_cpu); in sched_mm_cid_migrate_to()
11853 struct rq *rq = cpu_rq(cpu); in sched_mm_cid_remote_clear()
11911 struct rq *rq = cpu_rq(cpu); in sched_mm_cid_remote_clear_old()