Lines Matching defs:nsk
487 static void smc_adjust_sock_bufsizes(struct sock *nsk, struct sock *osk,
490 nsk->sk_userlocks = osk->sk_userlocks;
492 nsk->sk_sndbuf = osk->sk_sndbuf;
494 nsk->sk_rcvbuf = osk->sk_rcvbuf;
497 static void smc_copy_sock_settings(struct sock *nsk, struct sock *osk,
501 nsk->sk_type = osk->sk_type;
502 nsk->sk_sndtimeo = READ_ONCE(osk->sk_sndtimeo);
503 nsk->sk_rcvtimeo = READ_ONCE(osk->sk_rcvtimeo);
504 nsk->sk_mark = READ_ONCE(osk->sk_mark);
505 nsk->sk_priority = READ_ONCE(osk->sk_priority);
506 nsk->sk_rcvlowat = osk->sk_rcvlowat;
507 nsk->sk_bound_dev_if = osk->sk_bound_dev_if;
508 nsk->sk_err = osk->sk_err;
510 nsk->sk_flags &= ~mask;
511 nsk->sk_flags |= osk->sk_flags & mask;
513 smc_adjust_sock_bufsizes(nsk, osk, mask);
2704 struct sock *sk = sock->sk, *nsk;
2723 while (!(nsk = smc_accept_dequeue(sk, new_sock))) {
2743 rc = sock_error(nsk);
2751 if (smc_sk(nsk)->use_fallback) {
2752 struct sock *clcsk = smc_sk(nsk)->clcsock->sk;
2758 } else if (!atomic_read(&smc_sk(nsk)->conn.bytes_to_rcv)) {
2759 lock_sock(nsk);
2760 smc_rx_wait(smc_sk(nsk), &timeo, 0, smc_rx_data_available);
2761 release_sock(nsk);