Lines Matching refs:waiter
35 static inline int __ww_mutex_add_waiter(struct rt_mutex_waiter *waiter, in __ww_mutex_add_waiter() argument
53 struct rt_mutex_waiter *waiter, in __ww_mutex_check_kill() argument
340 waiter_update_prio(struct rt_mutex_waiter *waiter, struct task_struct *task) in waiter_update_prio() argument
342 lockdep_assert_held(&waiter->lock->wait_lock); in waiter_update_prio()
343 lockdep_assert(RB_EMPTY_NODE(&waiter->tree.entry)); in waiter_update_prio()
345 waiter->tree.prio = __waiter_prio(task); in waiter_update_prio()
346 waiter->tree.deadline = task->dl.deadline; in waiter_update_prio()
353 waiter_clone_prio(struct rt_mutex_waiter *waiter, struct task_struct *task) in waiter_clone_prio() argument
355 lockdep_assert_held(&waiter->lock->wait_lock); in waiter_clone_prio()
357 lockdep_assert(RB_EMPTY_NODE(&waiter->pi_tree.entry)); in waiter_clone_prio()
359 waiter->pi_tree.prio = waiter->tree.prio; in waiter_clone_prio()
360 waiter->pi_tree.deadline = waiter->tree.deadline; in waiter_clone_prio()
407 static inline bool rt_mutex_steal(struct rt_mutex_waiter *waiter, in rt_mutex_steal() argument
410 if (rt_waiter_node_less(&waiter->tree, &top_waiter->tree)) in rt_mutex_steal()
418 if (rt_prio(waiter->tree.prio) || dl_prio(waiter->tree.prio)) in rt_mutex_steal()
421 return rt_waiter_node_equal(&waiter->tree, &top_waiter->tree); in rt_mutex_steal()
457 rt_mutex_enqueue(struct rt_mutex_base *lock, struct rt_mutex_waiter *waiter) in rt_mutex_enqueue() argument
461 rb_add_cached(&waiter->tree.entry, &lock->waiters, __waiter_less); in rt_mutex_enqueue()
465 rt_mutex_dequeue(struct rt_mutex_base *lock, struct rt_mutex_waiter *waiter) in rt_mutex_dequeue() argument
469 if (RB_EMPTY_NODE(&waiter->tree.entry)) in rt_mutex_dequeue()
472 rb_erase_cached(&waiter->tree.entry, &lock->waiters); in rt_mutex_dequeue()
473 RB_CLEAR_NODE(&waiter->tree.entry); in rt_mutex_dequeue()
485 rt_mutex_enqueue_pi(struct task_struct *task, struct rt_mutex_waiter *waiter) in rt_mutex_enqueue_pi() argument
489 rb_add_cached(&waiter->pi_tree.entry, &task->pi_waiters, __pi_waiter_less); in rt_mutex_enqueue_pi()
493 rt_mutex_dequeue_pi(struct task_struct *task, struct rt_mutex_waiter *waiter) in rt_mutex_dequeue_pi() argument
497 if (RB_EMPTY_NODE(&waiter->pi_tree.entry)) in rt_mutex_dequeue_pi()
500 rb_erase_cached(&waiter->pi_tree.entry, &task->pi_waiters); in rt_mutex_dequeue_pi()
501 RB_CLEAR_NODE(&waiter->pi_tree.entry); in rt_mutex_dequeue_pi()
569 rt_mutex_cond_detect_deadlock(struct rt_mutex_waiter *waiter, in rt_mutex_cond_detect_deadlock() argument
573 return waiter != NULL; in rt_mutex_cond_detect_deadlock()
662 struct rt_mutex_waiter *waiter, *top_waiter = orig_waiter; in rt_mutex_adjust_prio_chain() local
714 waiter = task->pi_blocked_on; in rt_mutex_adjust_prio_chain()
725 if (!waiter) in rt_mutex_adjust_prio_chain()
744 if (next_lock != waiter->lock) in rt_mutex_adjust_prio_chain()
769 if (IS_ENABLED(CONFIG_PREEMPT_RT) && waiter->ww_ctx && detect_deadlock) in rt_mutex_adjust_prio_chain()
801 if (rt_waiter_node_equal(&waiter->tree, task_to_waiter_node(task))) { in rt_mutex_adjust_prio_chain()
812 lock = waiter->lock; in rt_mutex_adjust_prio_chain()
912 rt_mutex_dequeue(lock, waiter); in rt_mutex_adjust_prio_chain()
925 waiter_update_prio(waiter, task); in rt_mutex_adjust_prio_chain()
927 rt_mutex_enqueue(lock, waiter); in rt_mutex_adjust_prio_chain()
969 if (waiter == rt_mutex_top_waiter(lock)) { in rt_mutex_adjust_prio_chain()
977 waiter_clone_prio(waiter, task); in rt_mutex_adjust_prio_chain()
978 rt_mutex_enqueue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
981 } else if (prerequeue_top_waiter == waiter) { in rt_mutex_adjust_prio_chain()
992 rt_mutex_dequeue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
993 waiter = rt_mutex_top_waiter(lock); in rt_mutex_adjust_prio_chain()
994 waiter_clone_prio(waiter, task); in rt_mutex_adjust_prio_chain()
995 rt_mutex_enqueue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
1040 if (!detect_deadlock && waiter != top_waiter) in rt_mutex_adjust_prio_chain()
1065 struct rt_mutex_waiter *waiter) in try_to_take_rt_mutex() argument
1099 if (waiter) { in try_to_take_rt_mutex()
1106 if (waiter == top_waiter || rt_mutex_steal(waiter, top_waiter)) { in try_to_take_rt_mutex()
1111 rt_mutex_dequeue(lock, waiter); in try_to_take_rt_mutex()
1181 struct rt_mutex_waiter *waiter, in task_blocks_on_rt_mutex() argument
1187 struct rt_mutex_waiter *top_waiter = waiter; in task_blocks_on_rt_mutex()
1209 waiter->task = task; in task_blocks_on_rt_mutex()
1210 waiter->lock = lock; in task_blocks_on_rt_mutex()
1211 waiter_update_prio(waiter, task); in task_blocks_on_rt_mutex()
1212 waiter_clone_prio(waiter, task); in task_blocks_on_rt_mutex()
1217 rt_mutex_enqueue(lock, waiter); in task_blocks_on_rt_mutex()
1219 task->pi_blocked_on = waiter; in task_blocks_on_rt_mutex()
1228 res = __ww_mutex_add_waiter(waiter, rtm, ww_ctx); in task_blocks_on_rt_mutex()
1231 rt_mutex_dequeue(lock, waiter); in task_blocks_on_rt_mutex()
1242 if (waiter == rt_mutex_top_waiter(lock)) { in task_blocks_on_rt_mutex()
1244 rt_mutex_enqueue_pi(owner, waiter); in task_blocks_on_rt_mutex()
1249 } else if (rt_mutex_cond_detect_deadlock(waiter, chwalk)) { in task_blocks_on_rt_mutex()
1275 next_lock, waiter, task); in task_blocks_on_rt_mutex()
1291 struct rt_mutex_waiter *waiter; in mark_wakeup_next_waiter() local
1297 waiter = rt_mutex_top_waiter(lock); in mark_wakeup_next_waiter()
1306 rt_mutex_dequeue_pi(current, waiter); in mark_wakeup_next_waiter()
1330 rt_mutex_wake_q_add(wqh, waiter); in mark_wakeup_next_waiter()
1458 struct rt_mutex_waiter *waiter, in rtmutex_spin_on_owner() argument
1484 !rt_mutex_waiter_is_top_waiter(lock, waiter)) { in rtmutex_spin_on_owner()
1495 struct rt_mutex_waiter *waiter, in rtmutex_spin_on_owner() argument
1516 struct rt_mutex_waiter *waiter) in remove_waiter() argument
1518 bool is_top_waiter = (waiter == rt_mutex_top_waiter(lock)); in remove_waiter()
1525 rt_mutex_dequeue(lock, waiter); in remove_waiter()
1538 rt_mutex_dequeue_pi(owner, waiter); in remove_waiter()
1583 struct rt_mutex_waiter *waiter) in rt_mutex_slowlock_block() argument
1591 if (try_to_take_rt_mutex(lock, current, waiter)) in rt_mutex_slowlock_block()
1604 ret = __ww_mutex_check_kill(rtm, waiter, ww_ctx); in rt_mutex_slowlock_block()
1609 if (waiter == rt_mutex_top_waiter(lock)) in rt_mutex_slowlock_block()
1615 if (!owner || !rtmutex_spin_on_owner(lock, waiter, owner)) in rt_mutex_slowlock_block()
1661 struct rt_mutex_waiter *waiter) in __rt_mutex_slowlock() argument
1682 ret = task_blocks_on_rt_mutex(lock, waiter, current, ww_ctx, chwalk); in __rt_mutex_slowlock()
1684 ret = rt_mutex_slowlock_block(lock, ww_ctx, state, NULL, waiter); in __rt_mutex_slowlock()
1695 remove_waiter(lock, waiter); in __rt_mutex_slowlock()
1696 rt_mutex_handle_deadlock(ret, chwalk, waiter); in __rt_mutex_slowlock()
1714 struct rt_mutex_waiter waiter; in __rt_mutex_slowlock_locked() local
1717 rt_mutex_init_waiter(&waiter); in __rt_mutex_slowlock_locked()
1718 waiter.ww_ctx = ww_ctx; in __rt_mutex_slowlock_locked()
1721 &waiter); in __rt_mutex_slowlock_locked()
1723 debug_rt_mutex_free_waiter(&waiter); in __rt_mutex_slowlock_locked()
1776 struct rt_mutex_waiter waiter; in rtlock_slowlock_locked() local
1784 rt_mutex_init_rtlock_waiter(&waiter); in rtlock_slowlock_locked()
1791 task_blocks_on_rt_mutex(lock, &waiter, current, NULL, RT_MUTEX_MIN_CHAINWALK); in rtlock_slowlock_locked()
1795 if (try_to_take_rt_mutex(lock, current, &waiter)) in rtlock_slowlock_locked()
1798 if (&waiter == rt_mutex_top_waiter(lock)) in rtlock_slowlock_locked()
1804 if (!owner || !rtmutex_spin_on_owner(lock, &waiter, owner)) in rtlock_slowlock_locked()
1819 debug_rt_mutex_free_waiter(&waiter); in rtlock_slowlock_locked()