Lines Matching refs:sk_lock
569 mutex_acquire(&sk->sk_lock.dep_map, 0, 1, _RET_IP_); in __sk_receive_skb()
573 mutex_release(&sk->sk_lock.dep_map, _RET_IP_); in __sk_receive_skb()
2950 __releases(&sk->sk_lock.slock) in __lock_sock()
2951 __acquires(&sk->sk_lock.slock) in __lock_sock()
2956 prepare_to_wait_exclusive(&sk->sk_lock.wq, &wait, in __lock_sock()
2958 spin_unlock_bh(&sk->sk_lock.slock); in __lock_sock()
2960 spin_lock_bh(&sk->sk_lock.slock); in __lock_sock()
2964 finish_wait(&sk->sk_lock.wq, &wait); in __lock_sock()
2968 __releases(&sk->sk_lock.slock) in __release_sock()
2969 __acquires(&sk->sk_lock.slock) in __release_sock()
2976 spin_unlock_bh(&sk->sk_lock.slock); in __release_sock()
2990 spin_lock_bh(&sk->sk_lock.slock); in __release_sock()
3002 spin_lock_bh(&sk->sk_lock.slock); in __sk_flush_backlog()
3004 spin_unlock_bh(&sk->sk_lock.slock); in __sk_flush_backlog()
3505 mutex_acquire(&sk->sk_lock.dep_map, subclass, 0, _RET_IP_); in lock_sock_nested()
3508 spin_lock_bh(&sk->sk_lock.slock); in lock_sock_nested()
3511 sk->sk_lock.owned = 1; in lock_sock_nested()
3512 spin_unlock_bh(&sk->sk_lock.slock); in lock_sock_nested()
3518 spin_lock_bh(&sk->sk_lock.slock); in release_sock()
3529 if (waitqueue_active(&sk->sk_lock.wq)) in release_sock()
3530 wake_up(&sk->sk_lock.wq); in release_sock()
3531 spin_unlock_bh(&sk->sk_lock.slock); in release_sock()
3535 bool __lock_sock_fast(struct sock *sk) __acquires(&sk->sk_lock.slock) in __lock_sock_fast()
3538 spin_lock_bh(&sk->sk_lock.slock); in __lock_sock_fast()
3560 sk->sk_lock.owned = 1; in __lock_sock_fast()
3561 __acquire(&sk->sk_lock.slock); in __lock_sock_fast()
3562 spin_unlock_bh(&sk->sk_lock.slock); in __lock_sock_fast()