Lines Matching refs:rf
180 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock() argument
191 rq_pin_lock(rq, rf); in __task_rq_lock()
204 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock() argument
211 raw_spin_lock_irqsave(&p->pi_lock, rf->flags); in task_rq_lock()
232 rq_pin_lock(rq, rf); in task_rq_lock()
236 raw_spin_unlock_irqrestore(&p->pi_lock, rf->flags); in task_rq_lock()
349 struct rq_flags rf; in hrtick() local
353 rq_lock(rq, &rf); in hrtick()
356 rq_unlock(rq, &rf); in hrtick()
376 struct rq_flags rf; in __hrtick_start() local
378 rq_lock(rq, &rf); in __hrtick_start()
380 rq_unlock(rq, &rf); in __hrtick_start()
1026 struct rq_flags rf; in uclamp_update_util_min_rt_default() local
1033 rq = task_rq_lock(p, &rf); in uclamp_update_util_min_rt_default()
1035 task_rq_unlock(rq, p, &rf); in uclamp_update_util_min_rt_default()
1284 struct rq_flags rf; in uclamp_update_active() local
1295 rq = task_rq_lock(p, &rf); in uclamp_update_active()
1308 task_rq_unlock(rq, p, &rf); in uclamp_update_active()
1733 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task() argument
1740 rq_unlock(rq, rf); in move_queued_task()
1744 rq_lock(rq, rf); in move_queued_task()
1766 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task() argument
1774 rq = move_queued_task(rq, rf, p, dest_cpu); in __migrate_task()
1789 struct rq_flags rf; in migration_cpu_stop() local
1804 rq_lock(rq, &rf); in migration_cpu_stop()
1812 rq = __migrate_task(rq, &rf, p, arg->dest_cpu); in migration_cpu_stop()
1816 rq_unlock(rq, &rf); in migration_cpu_stop()
1876 struct rq_flags rf; in __set_cpus_allowed_ptr() local
1880 rq = task_rq_lock(p, &rf); in __set_cpus_allowed_ptr()
1932 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
1940 rq = move_queued_task(rq, &rf, p, dest_cpu); in __set_cpus_allowed_ptr()
1943 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
2142 struct rq_flags rf; in wait_task_inactive() local
2177 rq = task_rq_lock(p, &rf); in wait_task_inactive()
2184 task_rq_unlock(rq, p, &rf); in wait_task_inactive()
2459 struct rq_flags *rf) in ttwu_do_wakeup() argument
2471 rq_unpin_lock(rq, rf); in ttwu_do_wakeup()
2473 rq_repin_lock(rq, rf); in ttwu_do_wakeup()
2492 struct rq_flags *rf) in ttwu_do_activate() argument
2512 ttwu_do_wakeup(rq, p, wake_flags, rf); in ttwu_do_activate()
2542 struct rq_flags rf; in ttwu_runnable() local
2546 rq = __task_rq_lock(p, &rf); in ttwu_runnable()
2550 ttwu_do_wakeup(rq, p, wake_flags, &rf); in ttwu_runnable()
2553 __task_rq_unlock(rq, &rf); in ttwu_runnable()
2564 struct rq_flags rf; in sched_ttwu_pending() local
2576 rq_lock_irqsave(rq, &rf); in sched_ttwu_pending()
2586 ttwu_do_activate(rq, p, p->sched_remote_wakeup ? WF_MIGRATED : 0, &rf); in sched_ttwu_pending()
2589 rq_unlock_irqrestore(rq, &rf); in sched_ttwu_pending()
2621 struct rq_flags rf; in wake_up_if_idle() local
2631 rq_lock_irqsave(rq, &rf); in wake_up_if_idle()
2635 rq_unlock_irqrestore(rq, &rf); in wake_up_if_idle()
2694 struct rq_flags rf; in ttwu_queue() local
2699 rq_lock(rq, &rf); in ttwu_queue()
2701 ttwu_do_activate(rq, p, wake_flags, &rf); in ttwu_queue()
2702 rq_unlock(rq, &rf); in ttwu_queue()
3011 struct rq_flags rf; in try_invoke_on_locked_down_task() local
3017 rq = __task_rq_lock(p, &rf); in try_invoke_on_locked_down_task()
3020 rq_unlock(rq, &rf); in try_invoke_on_locked_down_task()
3336 struct rq_flags rf; in wake_up_new_task() local
3339 raw_spin_lock_irqsave(&p->pi_lock, rf.flags); in wake_up_new_task()
3354 rq = __task_rq_lock(p, &rf); in wake_up_new_task()
3367 rq_unpin_lock(rq, &rf); in wake_up_new_task()
3369 rq_repin_lock(rq, &rf); in wake_up_new_task()
3372 task_rq_unlock(rq, p, &rf); in wake_up_new_task()
3494 prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) in prepare_lock_switch() argument
3502 rq_unpin_lock(rq, rf); in prepare_lock_switch()
3729 struct task_struct *next, struct rq_flags *rf) in context_switch() argument
3776 prepare_lock_switch(rq, next, rf); in context_switch()
3943 struct rq_flags rf; in task_sched_runtime() local
3963 rq = task_rq_lock(p, &rf); in task_sched_runtime()
3975 task_rq_unlock(rq, p, &rf); in task_sched_runtime()
3989 struct rq_flags rf; in scheduler_tick() local
3995 rq_lock(rq, &rf); in scheduler_tick()
4004 rq_unlock(rq, &rf); in scheduler_tick()
4058 struct rq_flags rf; in sched_tick_remote() local
4072 rq_lock_irq(rq, &rf); in sched_tick_remote()
4091 rq_unlock_irq(rq, &rf); in sched_tick_remote()
4306 struct rq_flags *rf) in put_prev_task_balance() argument
4319 if (class->balance(rq, prev, rf)) in put_prev_task_balance()
4331 pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task() argument
4345 p = pick_next_task_fair(rq, prev, rf); in pick_next_task()
4359 put_prev_task_balance(rq, prev, rf); in pick_next_task()
4415 struct rq_flags rf; in __schedule() local
4446 rq_lock(rq, &rf); in __schedule()
4496 next = pick_next_task(rq, prev, &rf); in __schedule()
4528 rq = context_switch(rq, prev, next, &rf); in __schedule()
4531 rq_unlock_irq(rq, &rf); in __schedule()
4838 struct rq_flags rf; in rt_mutex_setprio() local
4850 rq = __task_rq_lock(p, &rf); in rt_mutex_setprio()
4946 __task_rq_unlock(rq, &rf); in rt_mutex_setprio()
4962 struct rq_flags rf; in set_user_nice() local
4971 rq = task_rq_lock(p, &rf); in set_user_nice()
5008 task_rq_unlock(rq, p, &rf); in set_user_nice()
5222 struct rq_flags rf; in __sched_setscheduler() local
5333 rq = task_rq_lock(p, &rf); in __sched_setscheduler()
5399 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
5460 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
5474 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
6095 struct rq_flags rf; in do_sched_yield() local
6098 rq = this_rq_lock_irq(&rf); in do_sched_yield()
6108 rq_unlock(rq, &rf); in do_sched_yield()
6363 struct rq_flags rf; in sched_rr_get_interval() local
6380 rq = task_rq_lock(p, &rf); in sched_rr_get_interval()
6384 task_rq_unlock(rq, p, &rf); in sched_rr_get_interval()
6655 struct rq_flags rf; in sched_setnuma() local
6658 rq = task_rq_lock(p, &rf); in sched_setnuma()
6673 task_rq_unlock(rq, p, &rf); in sched_setnuma()
6738 static void migrate_tasks(struct rq *dead_rq, struct rq_flags *rf) in migrate_tasks() argument
6742 struct rq_flags orf = *rf; in migrate_tasks()
6782 rq_unlock(rq, rf); in migrate_tasks()
6784 rq_relock(rq, rf); in migrate_tasks()
6798 rq = __migrate_task(rq, rf, next, dest_cpu); in migrate_tasks()
6800 rq_unlock(rq, rf); in migrate_tasks()
6802 *rf = orf; in migrate_tasks()
6803 rq_relock(rq, rf); in migrate_tasks()
6893 struct rq_flags rf; in sched_cpu_activate() local
6918 rq_lock_irqsave(rq, &rf); in sched_cpu_activate()
6923 rq_unlock_irqrestore(rq, &rf); in sched_cpu_activate()
6981 struct rq_flags rf; in sched_cpu_dying() local
6986 rq_lock_irqsave(rq, &rf); in sched_cpu_dying()
6991 migrate_tasks(rq, &rf); in sched_cpu_dying()
6993 rq_unlock_irqrestore(rq, &rf); in sched_cpu_dying()
7552 struct rq_flags rf; in sched_move_task() local
7555 rq = task_rq_lock(tsk, &rf); in sched_move_task()
7580 task_rq_unlock(rq, tsk, &rf); in sched_move_task()
7646 struct rq_flags rf; in cpu_cgroup_fork() local
7649 rq = task_rq_lock(task, &rf); in cpu_cgroup_fork()
7654 task_rq_unlock(rq, task, &rf); in cpu_cgroup_fork()
7961 struct rq_flags rf; in tg_set_cfs_bandwidth() local
7963 rq_lock_irq(rq, &rf); in tg_set_cfs_bandwidth()
7969 rq_unlock_irq(rq, &rf); in tg_set_cfs_bandwidth()