Lines Matching refs:rf
78 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock() argument
89 rq_pin_lock(rq, rf); in __task_rq_lock()
102 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock() argument
109 raw_spin_lock_irqsave(&p->pi_lock, rf->flags); in task_rq_lock()
130 rq_pin_lock(rq, rf); in task_rq_lock()
134 raw_spin_unlock_irqrestore(&p->pi_lock, rf->flags); in task_rq_lock()
240 struct rq_flags rf; in hrtick() local
244 rq_lock(rq, &rf); in hrtick()
247 rq_unlock(rq, &rf); in hrtick()
267 struct rq_flags rf; in __hrtick_start() local
269 rq_lock(rq, &rf); in __hrtick_start()
272 rq_unlock(rq, &rf); in __hrtick_start()
1049 struct rq_flags rf; in uclamp_update_active() local
1060 rq = task_rq_lock(p, &rf); in uclamp_update_active()
1073 task_rq_unlock(rq, p, &rf); in uclamp_update_active()
1488 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task() argument
1496 rq_unlock(rq, rf); in move_queued_task()
1500 rq_lock(rq, rf); in move_queued_task()
1523 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task() argument
1531 rq = move_queued_task(rq, rf, p, dest_cpu); in __migrate_task()
1546 struct rq_flags rf; in migration_cpu_stop() local
1561 rq_lock(rq, &rf); in migration_cpu_stop()
1569 rq = __migrate_task(rq, &rf, p, arg->dest_cpu); in migration_cpu_stop()
1573 rq_unlock(rq, &rf); in migration_cpu_stop()
1633 struct rq_flags rf; in __set_cpus_allowed_ptr() local
1637 rq = task_rq_lock(p, &rf); in __set_cpus_allowed_ptr()
1684 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
1692 rq = move_queued_task(rq, &rf, p, dest_cpu); in __set_cpus_allowed_ptr()
1695 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
1894 struct rq_flags rf; in wait_task_inactive() local
1929 rq = task_rq_lock(p, &rf); in wait_task_inactive()
1936 task_rq_unlock(rq, p, &rf); in wait_task_inactive()
2217 struct rq_flags *rf) in ttwu_do_wakeup() argument
2229 rq_unpin_lock(rq, rf); in ttwu_do_wakeup()
2231 rq_repin_lock(rq, rf); in ttwu_do_wakeup()
2250 struct rq_flags *rf) in ttwu_do_activate() argument
2265 ttwu_do_wakeup(rq, p, wake_flags, rf); in ttwu_do_activate()
2276 struct rq_flags rf; in ttwu_remote() local
2280 rq = __task_rq_lock(p, &rf); in ttwu_remote()
2284 ttwu_do_wakeup(rq, p, wake_flags, &rf); in ttwu_remote()
2287 __task_rq_unlock(rq, &rf); in ttwu_remote()
2298 struct rq_flags rf; in sched_ttwu_pending() local
2303 rq_lock_irqsave(rq, &rf); in sched_ttwu_pending()
2307 ttwu_do_activate(rq, p, p->sched_remote_wakeup ? WF_MIGRATED : 0, &rf); in sched_ttwu_pending()
2309 rq_unlock_irqrestore(rq, &rf); in sched_ttwu_pending()
2367 struct rq_flags rf; in wake_up_if_idle() local
2377 rq_lock_irqsave(rq, &rf); in wake_up_if_idle()
2381 rq_unlock_irqrestore(rq, &rf); in wake_up_if_idle()
2397 struct rq_flags rf; in ttwu_queue() local
2407 rq_lock(rq, &rf); in ttwu_queue()
2409 ttwu_do_activate(rq, p, wake_flags, &rf); in ttwu_queue()
2410 rq_unlock(rq, &rf); in ttwu_queue()
2945 struct rq_flags rf; in wake_up_new_task() local
2948 raw_spin_lock_irqsave(&p->pi_lock, rf.flags); in wake_up_new_task()
2962 rq = __task_rq_lock(p, &rf); in wake_up_new_task()
2975 rq_unpin_lock(rq, &rf); in wake_up_new_task()
2977 rq_repin_lock(rq, &rf); in wake_up_new_task()
2980 task_rq_unlock(rq, p, &rf); in wake_up_new_task()
3099 prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) in prepare_lock_switch() argument
3107 rq_unpin_lock(rq, rf); in prepare_lock_switch()
3334 struct task_struct *next, struct rq_flags *rf) in context_switch() argument
3381 prepare_lock_switch(rq, next, rf); in context_switch()
3548 struct rq_flags rf; in task_sched_runtime() local
3568 rq = task_rq_lock(p, &rf); in task_sched_runtime()
3580 task_rq_unlock(rq, p, &rf); in task_sched_runtime()
3594 struct rq_flags rf; in scheduler_tick() local
3598 rq_lock(rq, &rf); in scheduler_tick()
3605 rq_unlock(rq, &rf); in scheduler_tick()
3659 struct rq_flags rf; in sched_tick_remote() local
3673 rq_lock_irq(rq, &rf); in sched_tick_remote()
3689 rq_unlock_irq(rq, &rf); in sched_tick_remote()
3905 pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task() argument
3920 p = fair_sched_class.pick_next_task(rq, prev, rf); in pick_next_task()
3926 p = idle_sched_class.pick_next_task(rq, prev, rf); in pick_next_task()
3942 if (class->balance(rq, prev, rf)) in pick_next_task()
4002 struct rq_flags rf; in __schedule() local
4026 rq_lock(rq, &rf); in __schedule()
4048 next = pick_next_task(rq, prev, &rf); in __schedule()
4078 rq = context_switch(rq, prev, next, &rf); in __schedule()
4081 rq_unlock_irq(rq, &rf); in __schedule()
4376 struct rq_flags rf; in rt_mutex_setprio() local
4388 rq = __task_rq_lock(p, &rf); in rt_mutex_setprio()
4482 __task_rq_unlock(rq, &rf); in rt_mutex_setprio()
4498 struct rq_flags rf; in set_user_nice() local
4507 rq = task_rq_lock(p, &rf); in set_user_nice()
4545 task_rq_unlock(rq, p, &rf); in set_user_nice()
4759 struct rq_flags rf; in __sched_setscheduler() local
4870 rq = task_rq_lock(p, &rf); in __sched_setscheduler()
4936 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4997 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
5011 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
5583 struct rq_flags rf; in do_sched_yield() local
5586 rq = this_rq_lock_irq(&rf); in do_sched_yield()
5596 rq_unlock(rq, &rf); in do_sched_yield()
5851 struct rq_flags rf; in sched_rr_get_interval() local
5868 rq = task_rq_lock(p, &rf); in sched_rr_get_interval()
5872 task_rq_unlock(rq, p, &rf); in sched_rr_get_interval()
6149 struct rq_flags rf; in sched_setnuma() local
6152 rq = task_rq_lock(p, &rf); in sched_setnuma()
6167 task_rq_unlock(rq, p, &rf); in sched_setnuma()
6231 static void migrate_tasks(struct rq *dead_rq, struct rq_flags *rf) in migrate_tasks() argument
6235 struct rq_flags orf = *rf; in migrate_tasks()
6275 rq_unlock(rq, rf); in migrate_tasks()
6277 rq_relock(rq, rf); in migrate_tasks()
6291 rq = __migrate_task(rq, rf, next, dest_cpu); in migrate_tasks()
6293 rq_unlock(rq, rf); in migrate_tasks()
6295 *rf = orf; in migrate_tasks()
6296 rq_relock(rq, rf); in migrate_tasks()
6386 struct rq_flags rf; in sched_cpu_activate() local
6411 rq_lock_irqsave(rq, &rf); in sched_cpu_activate()
6416 rq_unlock_irqrestore(rq, &rf); in sched_cpu_activate()
6474 struct rq_flags rf; in sched_cpu_dying() local
6480 rq_lock_irqsave(rq, &rf); in sched_cpu_dying()
6485 migrate_tasks(rq, &rf); in sched_cpu_dying()
6487 rq_unlock_irqrestore(rq, &rf); in sched_cpu_dying()
7038 struct rq_flags rf; in sched_move_task() local
7041 rq = task_rq_lock(tsk, &rf); in sched_move_task()
7059 task_rq_unlock(rq, tsk, &rf); in sched_move_task()
7119 struct rq_flags rf; in cpu_cgroup_fork() local
7122 rq = task_rq_lock(task, &rf); in cpu_cgroup_fork()
7127 task_rq_unlock(rq, task, &rf); in cpu_cgroup_fork()
7424 struct rq_flags rf; in tg_set_cfs_bandwidth() local
7426 rq_lock_irq(rq, &rf); in tg_set_cfs_bandwidth()
7432 rq_unlock_irq(rq, &rf); in tg_set_cfs_bandwidth()