Lines Matching refs:waiter
35 static inline int __ww_mutex_add_waiter(struct rt_mutex_waiter *waiter, in __ww_mutex_add_waiter() argument
53 struct rt_mutex_waiter *waiter, in __ww_mutex_check_kill() argument
300 waiter_update_prio(struct rt_mutex_waiter *waiter, struct task_struct *task) in waiter_update_prio() argument
302 waiter->prio = __waiter_prio(task); in waiter_update_prio()
303 waiter->deadline = task->dl.deadline; in waiter_update_prio()
348 static inline bool rt_mutex_steal(struct rt_mutex_waiter *waiter, in rt_mutex_steal() argument
351 if (rt_mutex_waiter_less(waiter, top_waiter)) in rt_mutex_steal()
359 if (rt_prio(waiter->prio) || dl_prio(waiter->prio)) in rt_mutex_steal()
362 return rt_mutex_waiter_equal(waiter, top_waiter); in rt_mutex_steal()
398 rt_mutex_enqueue(struct rt_mutex_base *lock, struct rt_mutex_waiter *waiter) in rt_mutex_enqueue() argument
400 rb_add_cached(&waiter->tree_entry, &lock->waiters, __waiter_less); in rt_mutex_enqueue()
404 rt_mutex_dequeue(struct rt_mutex_base *lock, struct rt_mutex_waiter *waiter) in rt_mutex_dequeue() argument
406 if (RB_EMPTY_NODE(&waiter->tree_entry)) in rt_mutex_dequeue()
409 rb_erase_cached(&waiter->tree_entry, &lock->waiters); in rt_mutex_dequeue()
410 RB_CLEAR_NODE(&waiter->tree_entry); in rt_mutex_dequeue()
423 rt_mutex_enqueue_pi(struct task_struct *task, struct rt_mutex_waiter *waiter) in rt_mutex_enqueue_pi() argument
425 rb_add_cached(&waiter->pi_tree_entry, &task->pi_waiters, __pi_waiter_less); in rt_mutex_enqueue_pi()
429 rt_mutex_dequeue_pi(struct task_struct *task, struct rt_mutex_waiter *waiter) in rt_mutex_dequeue_pi() argument
431 if (RB_EMPTY_NODE(&waiter->pi_tree_entry)) in rt_mutex_dequeue_pi()
434 rb_erase_cached(&waiter->pi_tree_entry, &task->pi_waiters); in rt_mutex_dequeue_pi()
435 RB_CLEAR_NODE(&waiter->pi_tree_entry); in rt_mutex_dequeue_pi()
500 rt_mutex_cond_detect_deadlock(struct rt_mutex_waiter *waiter, in rt_mutex_cond_detect_deadlock() argument
504 return waiter != NULL; in rt_mutex_cond_detect_deadlock()
583 struct rt_mutex_waiter *waiter, *top_waiter = orig_waiter; in rt_mutex_adjust_prio_chain() local
635 waiter = task->pi_blocked_on; in rt_mutex_adjust_prio_chain()
646 if (!waiter) in rt_mutex_adjust_prio_chain()
665 if (next_lock != waiter->lock) in rt_mutex_adjust_prio_chain()
690 if (IS_ENABLED(CONFIG_PREEMPT_RT) && waiter->ww_ctx && detect_deadlock) in rt_mutex_adjust_prio_chain()
722 if (rt_mutex_waiter_equal(waiter, task_to_waiter(task))) { in rt_mutex_adjust_prio_chain()
732 lock = waiter->lock; in rt_mutex_adjust_prio_chain()
828 rt_mutex_dequeue(lock, waiter); in rt_mutex_adjust_prio_chain()
846 waiter_update_prio(waiter, task); in rt_mutex_adjust_prio_chain()
848 rt_mutex_enqueue(lock, waiter); in rt_mutex_adjust_prio_chain()
868 wake_up_state(waiter->task, waiter->wake_state); in rt_mutex_adjust_prio_chain()
878 if (waiter == rt_mutex_top_waiter(lock)) { in rt_mutex_adjust_prio_chain()
886 rt_mutex_enqueue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
889 } else if (prerequeue_top_waiter == waiter) { in rt_mutex_adjust_prio_chain()
900 rt_mutex_dequeue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
901 waiter = rt_mutex_top_waiter(lock); in rt_mutex_adjust_prio_chain()
902 rt_mutex_enqueue_pi(task, waiter); in rt_mutex_adjust_prio_chain()
947 if (!detect_deadlock && waiter != top_waiter) in rt_mutex_adjust_prio_chain()
972 struct rt_mutex_waiter *waiter) in try_to_take_rt_mutex() argument
1006 if (waiter) { in try_to_take_rt_mutex()
1013 if (waiter == top_waiter || rt_mutex_steal(waiter, top_waiter)) { in try_to_take_rt_mutex()
1018 rt_mutex_dequeue(lock, waiter); in try_to_take_rt_mutex()
1088 struct rt_mutex_waiter *waiter, in task_blocks_on_rt_mutex() argument
1094 struct rt_mutex_waiter *top_waiter = waiter; in task_blocks_on_rt_mutex()
1116 waiter->task = task; in task_blocks_on_rt_mutex()
1117 waiter->lock = lock; in task_blocks_on_rt_mutex()
1118 waiter_update_prio(waiter, task); in task_blocks_on_rt_mutex()
1123 rt_mutex_enqueue(lock, waiter); in task_blocks_on_rt_mutex()
1125 task->pi_blocked_on = waiter; in task_blocks_on_rt_mutex()
1134 res = __ww_mutex_add_waiter(waiter, rtm, ww_ctx); in task_blocks_on_rt_mutex()
1137 rt_mutex_dequeue(lock, waiter); in task_blocks_on_rt_mutex()
1148 if (waiter == rt_mutex_top_waiter(lock)) { in task_blocks_on_rt_mutex()
1150 rt_mutex_enqueue_pi(owner, waiter); in task_blocks_on_rt_mutex()
1155 } else if (rt_mutex_cond_detect_deadlock(waiter, chwalk)) { in task_blocks_on_rt_mutex()
1181 next_lock, waiter, task); in task_blocks_on_rt_mutex()
1197 struct rt_mutex_waiter *waiter; in mark_wakeup_next_waiter() local
1201 waiter = rt_mutex_top_waiter(lock); in mark_wakeup_next_waiter()
1210 rt_mutex_dequeue_pi(current, waiter); in mark_wakeup_next_waiter()
1234 rt_mutex_wake_q_add(wqh, waiter); in mark_wakeup_next_waiter()
1362 struct rt_mutex_waiter *waiter, in rtmutex_spin_on_owner() argument
1388 !rt_mutex_waiter_is_top_waiter(lock, waiter)) { in rtmutex_spin_on_owner()
1399 struct rt_mutex_waiter *waiter, in rtmutex_spin_on_owner() argument
1420 struct rt_mutex_waiter *waiter) in remove_waiter() argument
1422 bool is_top_waiter = (waiter == rt_mutex_top_waiter(lock)); in remove_waiter()
1429 rt_mutex_dequeue(lock, waiter); in remove_waiter()
1442 rt_mutex_dequeue_pi(owner, waiter); in remove_waiter()
1487 struct rt_mutex_waiter *waiter) in rt_mutex_slowlock_block() argument
1495 if (try_to_take_rt_mutex(lock, current, waiter)) in rt_mutex_slowlock_block()
1508 ret = __ww_mutex_check_kill(rtm, waiter, ww_ctx); in rt_mutex_slowlock_block()
1513 if (waiter == rt_mutex_top_waiter(lock)) in rt_mutex_slowlock_block()
1519 if (!owner || !rtmutex_spin_on_owner(lock, waiter, owner)) in rt_mutex_slowlock_block()
1565 struct rt_mutex_waiter *waiter) in __rt_mutex_slowlock() argument
1586 ret = task_blocks_on_rt_mutex(lock, waiter, current, ww_ctx, chwalk); in __rt_mutex_slowlock()
1588 ret = rt_mutex_slowlock_block(lock, ww_ctx, state, NULL, waiter); in __rt_mutex_slowlock()
1599 remove_waiter(lock, waiter); in __rt_mutex_slowlock()
1600 rt_mutex_handle_deadlock(ret, chwalk, waiter); in __rt_mutex_slowlock()
1618 struct rt_mutex_waiter waiter; in __rt_mutex_slowlock_locked() local
1621 rt_mutex_init_waiter(&waiter); in __rt_mutex_slowlock_locked()
1622 waiter.ww_ctx = ww_ctx; in __rt_mutex_slowlock_locked()
1625 &waiter); in __rt_mutex_slowlock_locked()
1627 debug_rt_mutex_free_waiter(&waiter); in __rt_mutex_slowlock_locked()
1680 struct rt_mutex_waiter waiter; in rtlock_slowlock_locked() local
1688 rt_mutex_init_rtlock_waiter(&waiter); in rtlock_slowlock_locked()
1695 task_blocks_on_rt_mutex(lock, &waiter, current, NULL, RT_MUTEX_MIN_CHAINWALK); in rtlock_slowlock_locked()
1699 if (try_to_take_rt_mutex(lock, current, &waiter)) in rtlock_slowlock_locked()
1702 if (&waiter == rt_mutex_top_waiter(lock)) in rtlock_slowlock_locked()
1708 if (!owner || !rtmutex_spin_on_owner(lock, &waiter, owner)) in rtlock_slowlock_locked()
1723 debug_rt_mutex_free_waiter(&waiter); in rtlock_slowlock_locked()