Lines Matching refs:locked
93 (cmpxchg_acquire(&lock->locked, 0, _Q_LOCKED_VAL) == 0)) { in pv_hybrid_queued_unfair_trylock()
123 return !READ_ONCE(lock->locked) && in trylock_clear_pending()
304 if (READ_ONCE(node->locked)) in pv_wait_node()
324 if (!READ_ONCE(node->locked)) { in pv_wait_node()
344 qstat_inc(qstat_pv_spurious_wakeup, !READ_ONCE(node->locked)); in pv_wait_node()
392 WRITE_ONCE(lock->locked, _Q_SLOW_VAL); in pv_kick_node()
457 if (xchg(&lock->locked, _Q_SLOW_VAL) == 0) { in pv_wait_head_or_lock()
463 WRITE_ONCE(lock->locked, _Q_LOCKED_VAL); in pv_wait_head_or_lock()
471 pv_wait(&lock->locked, _Q_SLOW_VAL); in pv_wait_head_or_lock()
494 __pv_queued_spin_unlock_slowpath(struct qspinlock *lock, u8 locked) in __pv_queued_spin_unlock_slowpath() argument
498 if (unlikely(locked != _Q_SLOW_VAL)) { in __pv_queued_spin_unlock_slowpath()
524 smp_store_release(&lock->locked, 0); in __pv_queued_spin_unlock_slowpath()
550 u8 locked; in __pv_queued_spin_unlock() local
557 locked = cmpxchg_release(&lock->locked, _Q_LOCKED_VAL, 0); in __pv_queued_spin_unlock()
558 if (likely(locked == _Q_LOCKED_VAL)) in __pv_queued_spin_unlock()
561 __pv_queued_spin_unlock_slowpath(lock, locked); in __pv_queued_spin_unlock()