Lines Matching refs:rf
66 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf) in __task_rq_lock() argument
77 rq_pin_lock(rq, rf); in __task_rq_lock()
90 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf) in task_rq_lock() argument
97 raw_spin_lock_irqsave(&p->pi_lock, rf->flags); in task_rq_lock()
117 rq_pin_lock(rq, rf); in task_rq_lock()
121 raw_spin_unlock_irqrestore(&p->pi_lock, rf->flags); in task_rq_lock()
227 struct rq_flags rf; in hrtick() local
231 rq_lock(rq, &rf); in hrtick()
234 rq_unlock(rq, &rf); in hrtick()
254 struct rq_flags rf; in __hrtick_start() local
256 rq_lock(rq, &rf); in __hrtick_start()
259 rq_unlock(rq, &rf); in __hrtick_start()
908 static struct rq *move_queued_task(struct rq *rq, struct rq_flags *rf, in move_queued_task() argument
916 rq_unlock(rq, rf); in move_queued_task()
920 rq_lock(rq, rf); in move_queued_task()
943 static struct rq *__migrate_task(struct rq *rq, struct rq_flags *rf, in __migrate_task() argument
951 rq = move_queued_task(rq, rf, p, dest_cpu); in __migrate_task()
966 struct rq_flags rf; in migration_cpu_stop() local
981 rq_lock(rq, &rf); in migration_cpu_stop()
989 rq = __migrate_task(rq, &rf, p, arg->dest_cpu); in migration_cpu_stop()
993 rq_unlock(rq, &rf); in migration_cpu_stop()
1053 struct rq_flags rf; in __set_cpus_allowed_ptr() local
1057 rq = task_rq_lock(p, &rf); in __set_cpus_allowed_ptr()
1104 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
1113 rq = move_queued_task(rq, &rf, p, dest_cpu); in __set_cpus_allowed_ptr()
1116 task_rq_unlock(rq, p, &rf); in __set_cpus_allowed_ptr()
1317 struct rq_flags rf; in wait_task_inactive() local
1352 rq = task_rq_lock(p, &rf); in wait_task_inactive()
1359 task_rq_unlock(rq, p, &rf); in wait_task_inactive()
1650 struct rq_flags *rf) in ttwu_do_wakeup() argument
1662 rq_unpin_lock(rq, rf); in ttwu_do_wakeup()
1664 rq_repin_lock(rq, rf); in ttwu_do_wakeup()
1683 struct rq_flags *rf) in ttwu_do_activate() argument
1698 ttwu_do_wakeup(rq, p, wake_flags, rf); in ttwu_do_activate()
1709 struct rq_flags rf; in ttwu_remote() local
1713 rq = __task_rq_lock(p, &rf); in ttwu_remote()
1717 ttwu_do_wakeup(rq, p, wake_flags, &rf); in ttwu_remote()
1720 __task_rq_unlock(rq, &rf); in ttwu_remote()
1731 struct rq_flags rf; in sched_ttwu_pending() local
1736 rq_lock_irqsave(rq, &rf); in sched_ttwu_pending()
1740 ttwu_do_activate(rq, p, p->sched_remote_wakeup ? WF_MIGRATED : 0, &rf); in sched_ttwu_pending()
1742 rq_unlock_irqrestore(rq, &rf); in sched_ttwu_pending()
1800 struct rq_flags rf; in wake_up_if_idle() local
1810 rq_lock_irqsave(rq, &rf); in wake_up_if_idle()
1814 rq_unlock_irqrestore(rq, &rf); in wake_up_if_idle()
1830 struct rq_flags rf; in ttwu_queue() local
1840 rq_lock(rq, &rf); in ttwu_queue()
1842 ttwu_do_activate(rq, p, wake_flags, &rf); in ttwu_queue()
1843 rq_unlock(rq, &rf); in ttwu_queue()
2069 static void try_to_wake_up_local(struct task_struct *p, struct rq_flags *rf) in try_to_wake_up_local() argument
2086 rq_unlock(rq, rf); in try_to_wake_up_local()
2088 rq_relock(rq, rf); in try_to_wake_up_local()
2104 ttwu_do_wakeup(rq, p, 0, rf); in try_to_wake_up_local()
2396 struct rq_flags rf; in wake_up_new_task() local
2399 raw_spin_lock_irqsave(&p->pi_lock, rf.flags); in wake_up_new_task()
2413 rq = __task_rq_lock(p, &rf); in wake_up_new_task()
2427 rq_unpin_lock(rq, &rf); in wake_up_new_task()
2429 rq_repin_lock(rq, &rf); in wake_up_new_task()
2432 task_rq_unlock(rq, p, &rf); in wake_up_new_task()
2551 prepare_lock_switch(struct rq *rq, struct task_struct *next, struct rq_flags *rf) in prepare_lock_switch() argument
2559 rq_unpin_lock(rq, rf); in prepare_lock_switch()
2786 struct task_struct *next, struct rq_flags *rf) in context_switch() argument
2822 prepare_lock_switch(rq, next, rf); in context_switch()
2997 struct rq_flags rf; in task_sched_runtime() local
3017 rq = task_rq_lock(p, &rf); in task_sched_runtime()
3029 task_rq_unlock(rq, p, &rf); in task_sched_runtime()
3043 struct rq_flags rf; in scheduler_tick() local
3047 rq_lock(rq, &rf); in scheduler_tick()
3054 rq_unlock(rq, &rf); in scheduler_tick()
3080 struct rq_flags rf; in sched_tick_remote() local
3093 rq_lock_irq(rq, &rf); in sched_tick_remote()
3109 rq_unlock_irq(rq, &rf); in sched_tick_remote()
3304 pick_next_task(struct rq *rq, struct task_struct *prev, struct rq_flags *rf) in pick_next_task() argument
3319 p = fair_sched_class.pick_next_task(rq, prev, rf); in pick_next_task()
3325 p = idle_sched_class.pick_next_task(rq, prev, rf); in pick_next_task()
3332 p = class->pick_next_task(rq, prev, rf); in pick_next_task()
3387 struct rq_flags rf; in __schedule() local
3411 rq_lock(rq, &rf); in __schedule()
3441 try_to_wake_up_local(to_wakeup, &rf); in __schedule()
3447 next = pick_next_task(rq, prev, &rf); in __schedule()
3473 rq = context_switch(rq, prev, next, &rf); in __schedule()
3476 rq_unlock_irq(rq, &rf); in __schedule()
3748 struct rq_flags rf; in rt_mutex_setprio() local
3760 rq = __task_rq_lock(p, &rf); in rt_mutex_setprio()
3854 __task_rq_unlock(rq, &rf); in rt_mutex_setprio()
3870 struct rq_flags rf; in set_user_nice() local
3879 rq = task_rq_lock(p, &rf); in set_user_nice()
3917 task_rq_unlock(rq, p, &rf); in set_user_nice()
4124 struct rq_flags rf; in __sched_setscheduler() local
4225 rq = task_rq_lock(p, &rf); in __sched_setscheduler()
4232 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4249 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4263 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4279 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4289 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4299 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4346 task_rq_unlock(rq, p, &rf); in __sched_setscheduler()
4932 struct rq_flags rf; in do_sched_yield() local
4937 rq_lock(rq, &rf); in do_sched_yield()
4947 rq_unlock(rq, &rf); in do_sched_yield()
5202 struct rq_flags rf; in sched_rr_get_interval() local
5219 rq = task_rq_lock(p, &rf); in sched_rr_get_interval()
5223 task_rq_unlock(rq, p, &rf); in sched_rr_get_interval()
5499 struct rq_flags rf; in sched_setnuma() local
5502 rq = task_rq_lock(p, &rf); in sched_setnuma()
5517 task_rq_unlock(rq, p, &rf); in sched_setnuma()
5580 static void migrate_tasks(struct rq *dead_rq, struct rq_flags *rf) in migrate_tasks() argument
5584 struct rq_flags orf = *rf; in migrate_tasks()
5616 next = pick_next_task(rq, &fake_task, rf); in migrate_tasks()
5629 rq_unlock(rq, rf); in migrate_tasks()
5631 rq_relock(rq, rf); in migrate_tasks()
5645 rq = __migrate_task(rq, rf, next, dest_cpu); in migrate_tasks()
5647 rq_unlock(rq, rf); in migrate_tasks()
5649 *rf = orf; in migrate_tasks()
5650 rq_relock(rq, rf); in migrate_tasks()
5740 struct rq_flags rf; in sched_cpu_activate() local
5770 rq_lock_irqsave(rq, &rf); in sched_cpu_activate()
5775 rq_unlock_irqrestore(rq, &rf); in sched_cpu_activate()
5827 struct rq_flags rf; in sched_cpu_dying() local
5833 rq_lock_irqsave(rq, &rf); in sched_cpu_dying()
5838 migrate_tasks(rq, &rf); in sched_cpu_dying()
5840 rq_unlock_irqrestore(rq, &rf); in sched_cpu_dying()
6347 struct rq_flags rf; in sched_move_task() local
6350 rq = task_rq_lock(tsk, &rf); in sched_move_task()
6368 task_rq_unlock(rq, tsk, &rf); in sched_move_task()
6428 struct rq_flags rf; in cpu_cgroup_fork() local
6431 rq = task_rq_lock(task, &rf); in cpu_cgroup_fork()
6436 task_rq_unlock(rq, task, &rf); in cpu_cgroup_fork()
6563 struct rq_flags rf; in tg_set_cfs_bandwidth() local
6565 rq_lock_irq(rq, &rf); in tg_set_cfs_bandwidth()
6571 rq_unlock_irq(rq, &rf); in tg_set_cfs_bandwidth()