| /linux/kernel/bpf/ |
| H A D | reuseport_array.c | 189 const struct sock *nsk, in reuseport_array_update_check() argument 200 if (nsk->sk_protocol != IPPROTO_UDP && nsk->sk_protocol != IPPROTO_TCP) in reuseport_array_update_check() 203 if (nsk->sk_family != AF_INET && nsk->sk_family != AF_INET6) in reuseport_array_update_check() 206 if (nsk->sk_type != SOCK_STREAM && nsk->sk_type != SOCK_DGRAM) in reuseport_array_update_check() 217 if (!sock_flag(nsk, SOCK_RCU_FREE) || !sk_hashed(nsk) || !nsk_reuse) in reuseport_array_update_check() 221 if (READ_ONCE(nsk->sk_user_data)) in reuseport_array_update_check() 236 struct sock *free_osk = NULL, *osk, *nsk; in bpf_fd_reuseport_array_update_elem() local 263 nsk = socket->sk; in bpf_fd_reuseport_array_update_elem() 264 if (!nsk) { in bpf_fd_reuseport_array_update_elem() 270 err = reuseport_array_update_check(array, nsk, in bpf_fd_reuseport_array_update_elem() [all …]
|
| /linux/net/core/ |
| H A D | sock_reuseport.c | 625 struct sock *nsk = NULL; in reuseport_migrate_sock() local 659 nsk = bpf_run_sk_reuseport(reuse, sk, prog, skb, migrating_sk, hash); in reuseport_migrate_sock() 665 if (!nsk) in reuseport_migrate_sock() 666 nsk = reuseport_select_sock_by_hash(reuse, hash, socks); in reuseport_migrate_sock() 668 if (IS_ERR_OR_NULL(nsk) || unlikely(!refcount_inc_not_zero(&nsk->sk_refcnt))) { in reuseport_migrate_sock() 669 nsk = NULL; in reuseport_migrate_sock() 675 return nsk; in reuseport_migrate_sock()
|
| H A D | sock.c | 2203 static void sock_copy(struct sock *nsk, const struct sock *osk) in sock_copy() argument 2207 void *sptr = nsk->sk_security; in sock_copy() 2219 memcpy(nsk, osk, offsetof(struct sock, sk_dontcopy_begin)); in sock_copy() 2221 unsafe_memcpy(&nsk->sk_dontcopy_end, &osk->sk_dontcopy_end, in sock_copy() 2226 nsk->sk_security = sptr; in sock_copy() 2227 security_sk_clone(osk, nsk); in sock_copy()
|
| /linux/net/iucv/ |
| H A D | af_iucv.c | 802 struct sock *sk = sock->sk, *nsk; in iucv_sock_accept() local 817 while (!(nsk = iucv_accept_dequeue(sk, newsock))) { in iucv_sock_accept() 1591 struct sock *sk, *nsk; in iucv_callback_connreq() local 1636 nsk = iucv_sock_alloc(NULL, sk->sk_protocol, GFP_ATOMIC, 0); in iucv_callback_connreq() 1637 if (!nsk) { in iucv_callback_connreq() 1643 niucv = iucv_sk(nsk); in iucv_callback_connreq() 1644 iucv_sock_init(nsk, sk); in iucv_callback_connreq() 1646 nsk->sk_allocation |= GFP_DMA; in iucv_callback_connreq() 1664 err = pr_iucv->path_accept(path, &af_iucv_handler, nuser_data, nsk); in iucv_callback_connreq() 1666 iucv_sever_path(nsk, 1); in iucv_callback_connreq() [all …]
|
| /linux/net/smc/ |
| H A D | af_smc.c | 489 static void smc_adjust_sock_bufsizes(struct sock *nsk, struct sock *osk, in smc_adjust_sock_bufsizes() argument 492 nsk->sk_userlocks = osk->sk_userlocks; in smc_adjust_sock_bufsizes() 494 nsk->sk_sndbuf = osk->sk_sndbuf; in smc_adjust_sock_bufsizes() 496 nsk->sk_rcvbuf = osk->sk_rcvbuf; in smc_adjust_sock_bufsizes() 499 static void smc_copy_sock_settings(struct sock *nsk, struct sock *osk, in smc_copy_sock_settings() argument 503 nsk->sk_type = osk->sk_type; in smc_copy_sock_settings() 504 nsk->sk_sndtimeo = READ_ONCE(osk->sk_sndtimeo); in smc_copy_sock_settings() 505 nsk->sk_rcvtimeo = READ_ONCE(osk->sk_rcvtimeo); in smc_copy_sock_settings() 506 nsk->sk_mark = READ_ONCE(osk->sk_mark); in smc_copy_sock_settings() 507 nsk->sk_priority = READ_ONCE(osk->sk_priority); in smc_copy_sock_settings() [all …]
|
| /linux/net/ipv6/ |
| H A D | tcp_ipv6.c | 1605 struct sock *nsk = tcp_v6_cookie_check(sk, skb); local 1607 if (nsk != sk) { 1608 if (nsk) { 1609 reason = tcp_child_process(sk, nsk, skb); 1751 struct sock *nsk; local 1770 nsk = reuseport_migrate_sock(sk, req_to_sk(req), skb); 1771 if (!nsk) { 1775 sk = nsk; 1783 nsk = NULL; 1788 nsk = tcp_check_req(sk, skb, req, false, &req_stolen, [all …]
|
| /linux/net/dccp/ |
| H A D | ipv4.c | |
| H A D | ipv6.c | |
| /linux/net/ipv4/ |
| H A D | tcp_ipv4.c | 1894 struct sock *nsk = tcp_v4_cookie_check(sk, skb); local 1896 if (!nsk) 1898 if (nsk != sk) { 1899 reason = tcp_child_process(sk, nsk, skb); 1901 rsk = nsk; 2204 struct sock *nsk; local 2223 nsk = reuseport_migrate_sock(sk, req_to_sk(req), skb); 2224 if (!nsk) { 2228 sk = nsk; 2239 nsk = NULL; [all …]
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 3506 struct sock *nsk = sk_clone_lock(sk, GFP_ATOMIC); in mptcp_sk_clone_init() local 3510 if (!nsk) in mptcp_sk_clone_init() 3514 if (nsk->sk_family == AF_INET6) in mptcp_sk_clone_init() 3515 inet_sk(nsk)->pinet6 = mptcp_inet6_sk(nsk); in mptcp_sk_clone_init() 3518 __mptcp_init_sock(nsk); in mptcp_sk_clone_init() 3521 if (nsk->sk_family == AF_INET6) in mptcp_sk_clone_init() 3522 mptcp_copy_ip6_options(nsk, sk); in mptcp_sk_clone_init() 3525 mptcp_copy_ip_options(nsk, sk); in mptcp_sk_clone_init() 3527 msk = mptcp_sk(nsk); in mptcp_sk_clone_init() 3545 sock_reset_flag(nsk, SOCK_RCU_FREE); in mptcp_sk_clone_init() [all …]
|
| /linux/net/bluetooth/rfcomm/ |
| H A D | sock.c | 476 struct sock *sk = sock->sk, *nsk; in rfcomm_sock_accept() 499 nsk = bt_accept_dequeue(sk, newsock); in rfcomm_sock_accept() 500 if (nsk) in rfcomm_sock_accept() 526 BT_DBG("new socket %p", nsk); in rfcomm_sock_accept() 475 struct sock *sk = sock->sk, *nsk; rfcomm_sock_accept() local
|
| /linux/net/bluetooth/ |
| H A D | l2cap_sock.c | 333 struct sock *sk = sock->sk, *nsk; in l2cap_sock_accept() local 351 nsk = bt_accept_dequeue(sk, newsock); in l2cap_sock_accept() 352 if (nsk) in l2cap_sock_accept() 378 BT_DBG("new socket %p", nsk); in l2cap_sock_accept()
|
| /linux/arch/arm64/kvm/ |
| H A D | pmu-emul.c | 648 bool nsk = evtreg & ARMV8_PMU_EXCLUDE_NS_EL1; in kvm_pmc_counts_at_el1() local 651 return p == nsk; in kvm_pmc_counts_at_el1()
|
| /linux/include/net/ |
| H A D | sock.h | 1030 static inline void sock_copy_flags(struct sock *nsk, const struct sock *osk) in sock_copy_flags() argument 1032 nsk->sk_flags = osk->sk_flags; in sock_copy_flags()
|