Lines Matching defs:rq
233 void sched_core_enqueue(struct rq *rq, struct task_struct *p) in sched_core_enqueue()
243 void sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags) in sched_core_dequeue()
265 static struct task_struct *sched_core_find(struct rq *rq, unsigned long cookie) in sched_core_find()
432 static inline void sched_core_enqueue(struct rq *rq, struct task_struct *p) { } in sched_core_enqueue()
434 sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags) { } in sched_core_dequeue()
530 void raw_spin_rq_lock_nested(struct rq *rq, int subclass) in raw_spin_rq_lock_nested()
555 bool raw_spin_rq_trylock(struct rq *rq) in raw_spin_rq_trylock()
579 void raw_spin_rq_unlock(struct rq *rq) in raw_spin_rq_unlock()
609 struct rq *rq; in __task_rq_lock() local
634 struct rq *rq; in task_rq_lock() local
673 static void update_rq_clock_task(struct rq *rq, s64 delta) in update_rq_clock_task()
728 void update_rq_clock(struct rq *rq) in update_rq_clock()
755 static void hrtick_clear(struct rq *rq) in hrtick_clear()
767 struct rq *rq = container_of(timer, struct rq, hrtick_timer); in hrtick() local
782 static void __hrtick_restart(struct rq *rq) in __hrtick_restart()
795 struct rq *rq = arg; in __hrtick_start() local
808 void hrtick_start(struct rq *rq, u64 delay) in hrtick_start()
832 void hrtick_start(struct rq *rq, u64 delay) in hrtick_start()
845 static void hrtick_rq_init(struct rq *rq) in hrtick_rq_init()
854 static inline void hrtick_clear(struct rq *rq) in hrtick_clear()
858 static inline void hrtick_rq_init(struct rq *rq) in hrtick_rq_init()
1019 void resched_curr(struct rq *rq) in resched_curr()
1045 struct rq *rq = cpu_rq(cpu); in resched_cpu() local
1111 struct rq *rq = cpu_rq(cpu); in wake_up_idle_cpu() local
1155 struct rq *rq = info; in nohz_csd_func() local
1175 bool sched_can_stop_tick(struct rq *rq) in sched_can_stop_tick()
1372 uclamp_idle_value(struct rq *rq, enum uclamp_id clamp_id, in uclamp_idle_value()
1388 static inline void uclamp_idle_reset(struct rq *rq, enum uclamp_id clamp_id, in uclamp_idle_reset()
1399 unsigned int uclamp_rq_max_value(struct rq *rq, enum uclamp_id clamp_id, in uclamp_rq_max_value()
1439 struct rq *rq; in uclamp_update_util_min_rt_default() local
1521 static inline void uclamp_rq_inc_id(struct rq *rq, struct task_struct *p, in uclamp_rq_inc_id()
1559 static inline void uclamp_rq_dec_id(struct rq *rq, struct task_struct *p, in uclamp_rq_dec_id()
1625 static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p) in uclamp_rq_inc()
1649 static inline void uclamp_rq_dec(struct rq *rq, struct task_struct *p) in uclamp_rq_dec()
1669 static inline void uclamp_rq_reinc_id(struct rq *rq, struct task_struct *p, in uclamp_rq_reinc_id()
1691 struct rq *rq; in uclamp_update_active() local
1971 static void __init init_uclamp_rq(struct rq *rq) in init_uclamp_rq()
2011 static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p) { } in uclamp_rq_inc()
2012 static inline void uclamp_rq_dec(struct rq *rq, struct task_struct *p) { } in uclamp_rq_dec()
2049 static inline void enqueue_task(struct rq *rq, struct task_struct *p, int flags) in enqueue_task()
2066 static inline void dequeue_task(struct rq *rq, struct task_struct *p, int flags) in dequeue_task()
2083 void activate_task(struct rq *rq, struct task_struct *p, int flags) in activate_task()
2090 void deactivate_task(struct rq *rq, struct task_struct *p, int flags) in deactivate_task()
2161 static inline void check_class_changed(struct rq *rq, struct task_struct *p, in check_class_changed()
2174 void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags) in check_preempt_curr()
2198 static void migrate_disable_switch(struct rq *rq, struct task_struct *p) in migrate_disable_switch()
2259 static inline bool rq_has_pinned_tasks(struct rq *rq) in rq_has_pinned_tasks()
2313 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task()
2359 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task()
2382 struct rq *rq = this_rq(); in migration_cpu_stop() local
2481 struct rq *lowest_rq = NULL, *rq = this_rq(); in push_cpu_stop() local
2540 struct rq *rq = task_rq(p); in __do_set_cpus_allowed() local
2691 static int affine_move_task(struct rq *rq, struct task_struct *p, struct rq_flags *rf, in affine_move_task()
2837 struct rq *rq, in __set_cpus_allowed_ptr_locked()
2932 struct rq *rq; in __set_cpus_allowed_ptr() local
2957 struct rq *rq; in restrict_cpus_allowed_ptr() local
3269 struct rq *rq; in wait_task_inactive() local
3557 static inline void migrate_disable_switch(struct rq *rq, struct task_struct *p) { } in migrate_disable_switch()
3559 static inline bool rq_has_pinned_tasks(struct rq *rq) in rq_has_pinned_tasks()
3569 struct rq *rq; in ttwu_stat() local
3608 static void ttwu_do_wakeup(struct rq *rq, struct task_struct *p, int wake_flags, in ttwu_do_wakeup()
3644 ttwu_do_activate(struct rq *rq, struct task_struct *p, int wake_flags, in ttwu_do_activate()
3696 struct rq *rq; in ttwu_runnable() local
3715 struct rq *rq = this_rq(); in sched_ttwu_pending() local
3747 struct rq *rq = cpu_rq(cpu); in send_call_function_single_ipi() local
3763 struct rq *rq = cpu_rq(cpu); in __ttwu_queue_wakelist() local
3773 struct rq *rq = cpu_rq(cpu); in wake_up_if_idle() local
3861 struct rq *rq = cpu_rq(cpu); in ttwu_queue() local
4253 struct rq *rq = NULL; in task_call_func() local
4685 struct rq *rq; in wake_up_new_task() local
4844 static void do_balance_callbacks(struct rq *rq, struct balance_callback *head) in do_balance_callbacks()
4880 __splice_balance_callbacks(struct rq *rq, bool split) in __splice_balance_callbacks()
4904 static inline struct balance_callback *splice_balance_callbacks(struct rq *rq) in splice_balance_callbacks()
4909 static void __balance_callbacks(struct rq *rq) in __balance_callbacks()
4914 static inline void balance_callbacks(struct rq *rq, struct balance_callback *head) in balance_callbacks()
4927 static inline void __balance_callbacks(struct rq *rq) in __balance_callbacks()
4931 static inline struct balance_callback *splice_balance_callbacks(struct rq *rq) in splice_balance_callbacks()
4936 static inline void balance_callbacks(struct rq *rq, struct balance_callback *head) in balance_callbacks()
4943 prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) in prepare_lock_switch()
4959 static inline void finish_lock_switch(struct rq *rq) in finish_lock_switch()
5013 prepare_task_switch(struct rq *rq, struct task_struct *prev, in prepare_task_switch()
5048 struct rq *rq = this_rq(); in finish_task_switch() local
5157 context_switch(struct rq *rq, struct task_struct *prev, in context_switch()
5374 struct rq *rq; in task_sched_runtime() local
5411 static u64 cpu_resched_latency(struct rq *rq) in cpu_resched_latency()
5456 static inline u64 cpu_resched_latency(struct rq *rq) { return 0; } in cpu_resched_latency()
5466 struct rq *rq = cpu_rq(cpu); in scheduler_tick() local
5541 struct rq *rq = cpu_rq(cpu); in sched_tick_remote() local
5791 static void put_prev_task_balance(struct rq *rq, struct task_struct *prev, in put_prev_task_balance()
5817 __pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in __pick_next_task()
5875 static inline struct task_struct *pick_task(struct rq *rq) in pick_task()
5894 pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task()
6189 static void sched_core_balance(struct rq *rq) in sched_core_balance()
6211 static void queue_core_balance(struct rq *rq) in queue_core_balance()
6228 struct rq *rq = cpu_rq(cpu), *core_rq = NULL; in sched_core_cpu_starting() local
6271 struct rq *rq = cpu_rq(cpu), *core_rq = NULL; in sched_core_cpu_deactivate() local
6325 struct rq *rq = cpu_rq(cpu); in sched_core_cpu_dying() local
6338 pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task()
6408 struct rq *rq; in __schedule() local
6895 struct rq *rq; in rt_mutex_setprio() local
7018 struct rq *rq; in set_user_nice() local
7153 struct rq *rq = cpu_rq(cpu); in idle_cpu() local
7223 struct rq *rq = cpu_rq(cpu); in effective_cpu_util() local
7444 struct rq *rq; in __sched_setscheduler() local
8294 struct rq *rq; in do_sched_yield() local
8649 struct rq *rq, *p_rq; in yield_to() local
8804 struct rq *rq; in sched_rr_get_interval() local
8967 struct rq *rq = cpu_rq(cpu); in init_idle() local
9106 struct rq *rq; in sched_setnuma() local
9150 struct rq *rq = this_rq(); in __balance_push_cpu_stop() local
9180 static void balance_push(struct rq *rq) in balance_push()
9243 struct rq *rq = cpu_rq(cpu); in balance_push_set() local
9264 struct rq *rq = this_rq(); in balance_hotplug_wait() local
9273 static inline void balance_push(struct rq *rq) in balance_push()
9287 void set_rq_online(struct rq *rq) in set_rq_online()
9302 void set_rq_offline(struct rq *rq) in set_rq_offline()
9369 struct rq *rq = cpu_rq(cpu); in sched_cpu_activate() local
9414 struct rq *rq = cpu_rq(cpu); in sched_cpu_deactivate() local
9481 struct rq *rq = cpu_rq(cpu); in sched_rq_cpu_starting() local
9523 static void calc_load_migrate(struct rq *rq) in calc_load_migrate()
9531 static void dump_rq_tasks(struct rq *rq, const char *loglvl) in dump_rq_tasks()
9552 struct rq *rq = cpu_rq(cpu); in sched_cpu_dying() local
9696 struct rq *rq; in sched_init() local
10216 struct rq *rq; in sched_move_task() local
10602 struct rq *rq = cfs_rq->rq; in tg_set_cfs_bandwidth() local
11227 void call_trace_sched_update_nr_running(struct rq *rq, int count) in call_trace_sched_update_nr_running()