Lines Matching defs:sk_lock
242 * @sk_lock: synchronizer
461 socket_lock_t sk_lock;
1653 * Mark both the sk_lock and the sk_lock.slock as a
1659 sk->sk_lock.owned = 0; \
1660 init_waitqueue_head(&sk->sk_lock.wq); \
1661 spin_lock_init(&(sk)->sk_lock.slock); \
1662 debug_check_no_locks_freed((void *)&(sk)->sk_lock, \
1663 sizeof((sk)->sk_lock)); \
1664 lockdep_set_class_and_name(&(sk)->sk_lock.slock, \
1666 lockdep_init_map(&(sk)->sk_lock.dep_map, (name), (key), 0); \
1671 return lockdep_is_held(&sk->sk_lock) ||
1672 lockdep_is_held(&sk->sk_lock.slock);
1687 #define bh_lock_sock(__sk) spin_lock(&((__sk)->sk_lock.slock))
1689 spin_lock_nested(&((__sk)->sk_lock.slock), \
1691 #define bh_unlock_sock(__sk) spin_unlock(&((__sk)->sk_lock.slock))
1693 bool __lock_sock_fast(struct sock *sk) __acquires(&sk->sk_lock.slock);
1702 * sk_lock.slock locked, owned = 0, BH disabled
1706 * sk_lock.slock unlocked, owned = 1, BH enabled
1710 /* The sk_lock has mutex_lock() semantics here. */
1711 mutex_acquire(&sk->sk_lock.dep_map, 0, 0, _RET_IP_);
1719 mutex_acquire(&sk->sk_lock.dep_map, SINGLE_DEPTH_NESTING, 0, _RET_IP_);
1733 __releases(&sk->sk_lock.slock)
1737 __release(&sk->sk_lock.slock);
1739 mutex_release(&sk->sk_lock.dep_map, _RET_IP_);
1740 spin_unlock_bh(&sk->sk_lock.slock);
1780 return sk->sk_lock.owned;
1785 return sk->sk_lock.owned;
1791 sk->sk_lock.owned = 0;
1793 /* The sk_lock has mutex_unlock() semantics: */
1794 mutex_release(&sk->sk_lock.dep_map, _RET_IP_);
1803 !spin_is_locked(&sk->sk_lock.slock);