| /linux/net/ipv4/ |
| H A D | tcp_fastopen.c | 412 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check() 429 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie() 444 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen() 450 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen() 469 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 487 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 492 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 497 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen() 584 struct net *net = sock_net(sk); in tcp_fastopen_active_disable() 586 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout)) in tcp_fastopen_active_disable() [all …]
|
| H A D | tcp_input.c | 217 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn() 634 min(sndmem, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[2]))); in tcp_sndbuf_expand() 669 int window = tcp_win_from_space(sk, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])) >> 1; in __tcp_grow_window() 742 int tcp_app_win = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_app_win); in tcp_init_buffer_space() 782 struct net *net = sock_net(sk); in tcp_clamp_window() 913 const struct net *net = sock_net(sk); in tcp_rcvbuf_grow() 1155 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio); in tcp_update_pacing_rate() 1157 rate *= READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio); in tcp_update_pacing_rate() 1298 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering))); in tcp_check_sack_reordering() 1303 NET_INC_STATS(sock_net(sk), in tcp_check_sack_reordering() [all …]
|
| H A D | tcp_minisocks.c | 330 struct net *net = sock_net(sk); in tcp_time_wait() 667 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child() 705 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req() 753 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 853 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req() 859 NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req() 862 NET_INC_STATS(sock_net(sk), LINUX_MIB_TSECRREJECTED); in tcp_check_req() 881 TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req() 916 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req() 949 __NET_INC_STATS(sock_net(s in tcp_check_req() [all...] |
| H A D | inet_hashtables.c | 57 return inet6_ehashfn(sock_net(sk), in sk_ehashfn() 61 return inet_ehashfn(sock_net(sk), in sk_ehashfn() 221 struct net *net = sock_net(sk); in __inet_put_port() 263 struct net *net = sock_net(sk); in __inet_inherit_port() 343 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 348 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk() 360 if (net_eq(sock_net(sk), net) && READ_ONCE(sk->sk_num) == hnum && in compute_score() 591 struct net *net = sock_net(sk); in __inet_check_established() 643 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in __inet_check_established() 677 struct net *net = sock_net(sk); in inet_ehash_lookup_by_sk() [all …]
|
| H A D | datagram.c | 41 if (!oif || netif_index_is_l3_master(sock_net(sk), oif)) in __ip4_datagram_connect() 55 IP_INC_STATS(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); in __ip4_datagram_connect() 118 rt = ip_route_output_flow(sock_net(sk), &fl4, sk); in ip4_datagram_release_cb()
|
| H A D | inet_connection_sock.c | 136 inet_get_local_port_range(sock_net(sk), &lo, &hi); in inet_sk_get_local_port_range() 276 const struct net *net = sock_net(sk); in inet_bhash2_addr_any_conflict() 322 struct net *net = sock_net(sk); in inet_csk_find_open_port() 511 struct net *net = sock_net(sk); in inet_csk_get_port() 898 write_pnet(&ireq->ireq_net, sock_net(sk_listener)); in inet_reqsk_alloc() 924 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQFAILURE); in inet_reqsk_clone() 1059 net = sock_net(sk_listener); in reqsk_timer_handler() 1429 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQSUCCESS); in inet_csk_complete_hashdance() 1435 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPMIGRATEREQFAILURE); in inet_csk_complete_hashdance() 1484 __NET_INC_STATS(sock_net(nsk), in inet_csk_listen_stop() [all …]
|
| H A D | udp.c | 142 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse() 179 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse2() 202 struct net *net = sock_net(sk); in udp_reuseport_add_sock() 207 if (net_eq(sock_net(sk2), net) && in udp_reuseport_add_sock() 235 struct net *net = sock_net(sk); in udp_lib_get_port() 327 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in udp_lib_get_port() 352 ipv4_portaddr_hash(sock_net(sk), htonl(INADDR_ANY), snum); in udp_v4_get_port() 354 ipv4_portaddr_hash(sock_net(sk), inet_sk(sk)->inet_rcv_saddr, 0); in udp_v4_get_port() 370 if (!net_eq(sock_net(sk), net) || in compute_score() 616 struct net *net = sock_net(sk); in udp_lib_hash4() [all …]
|
| /linux/net/vmw_vsock/ |
| H A D | diag.c | 59 net = sock_net(skb->sk); in vsock_diag_dump() 79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump() 144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
|
| /linux/net/tls/ |
| H A D | tls_main.c | 350 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in tls_sk_proto_cleanup() 353 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in tls_sk_proto_cleanup() 358 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXSW); in tls_sk_proto_cleanup() 361 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXDEVICE); in tls_sk_proto_cleanup() 666 TLS_INC_STATS(sock_net(sk), tx ? LINUX_MIB_TLSTXREKEYERROR in do_tls_setsockopt_conf() 717 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXDEVICE); in do_tls_setsockopt_conf() 718 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in do_tls_setsockopt_conf() 726 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXREKEYOK); in do_tls_setsockopt_conf() 728 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXSW); in do_tls_setsockopt_conf() 729 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in do_tls_setsockopt_conf() [all …]
|
| /linux/net/mctp/ |
| H A D | af_mctp.c | 56 struct net *net = sock_net(&msk->sk); in mctp_bind() 136 struct net *net = sock_net(&msk->sk); in mctp_connect() 226 addr->smctp_network = mctp_default_net(sock_net(sk)); in mctp_sendmsg() 236 rc = mctp_dst_from_extaddr(&dst, sock_net(sk), in mctp_sendmsg() 244 rc = mctp_route_lookup(sock_net(sk), addr->smctp_network, in mctp_sendmsg() 505 struct net *net = sock_net(&msk->sk); in mctp_ioctl_alloctag() 548 struct net *net = sock_net(&msk->sk); in mctp_ioctl_droptag() 655 struct net *net = sock_net(&msk->sk); in mctp_sk_expire_keys() 710 struct net *net = sock_net(sk); in mctp_sk_hash() 758 struct net *net = sock_net(s in mctp_sk_unhash() [all...] |
| /linux/net/netlink/ |
| H A D | diag.c | 94 struct net *net = sock_net(skb->sk); in __netlink_diag_dump() 134 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 160 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump() 228 struct net *net = sock_net(skb->sk); in netlink_diag_handler_dump()
|
| H A D | af_netlink.c | 292 if (!net_eq(dev_net(dev), sock_net(sk))) in __netlink_deliver_tap_skb() 347 netlink_deliver_tap(sock_net(dst), skb); in netlink_deliver_tap_kernel() 481 !net_eq(sock_net(&nlk->sk), read_pnet(&x->pnet)); in netlink_compare() 506 netlink_compare_arg_init(&arg, sock_net(sk), nlk_sk(sk)->portid); in __netlink_insert() 744 nlk->netlink_unbind(sock_net(sk), i + 1); in netlink_release() 757 .net = sock_net(sk), in netlink_release() 794 sock_prot_inuse_add(sock_net(sk), &netlink_proto, -1); in netlink_release() 803 struct net *net = sock_net(sk); in netlink_autobind() 900 return netlink_ns_capable(skb, sock_net(skb->sk)->user_ns, cap); in netlink_net_capable() 907 ns_capable(sock_net(soc in netlink_allowed() [all...] |
| /linux/net/handshake/ |
| H A D | request.c | 94 trace_handshake_destruct(sock_net(sk), req, sk); in handshake_sk_destruct() 243 net = sock_net(req->hr_sk); in handshake_req_submit() 291 struct net *net = sock_net(sk); in handshake_complete() 320 net = sock_net(sk); in handshake_req_cancel()
|
| /linux/net/smc/ |
| H A D | smc_stats.h | 113 struct net *_net = sock_net(&__smc->sk); \ 128 struct net *_net = sock_net(&__smc->sk); \ 166 struct net *_net = sock_net(&(_smc)->sk); \ 185 struct net *net = sock_net(&(_smc)->sk); \ 231 struct net *net = sock_net(&(__smc)->sk); \
|
| H A D | smc_tracepoint.h | 35 __entry->net_cookie = sock_net(sk)->net_cookie; 61 __entry->net_cookie = sock_net(sk)->net_cookie;
|
| /linux/net/xfrm/ |
| H A D | espintcp.c | 17 XFRM_INC_STATS(sock_net(sk), LINUX_MIB_XFRMINERROR); in handle_nonesp() 39 skb->dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in handle_esp() 66 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 79 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 87 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv() 94 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv() 100 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv()
|
| /linux/net/mptcp/ |
| H A D | subflow.c | 35 MPTCP_INC_STATS(sock_net(req_to_sk(req)), field); in SUBFLOW_REQ_INC_STATS() 91 msk = mptcp_token_get_sock(sock_net(req_to_sk(req)), subflow_req->token); in subflow_token_join_request() 114 subflow_req->csum_reqd = mptcp_is_checksum_enabled(sock_net(sk_listener)); in subflow_init_req() 115 subflow_req->allow_join_id0 = mptcp_allow_join_id0(sock_net(sk_listener)); in subflow_init_req() 550 MPTCP_INC_STATS(sock_net(sk), in subflow_finish_connect() 563 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_MPCAPABLEACTIVEACK); in subflow_finish_connect() 584 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_JOINSYNACKMAC); in subflow_finish_connect() 599 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_JOINSYNACKRX); in subflow_finish_connect() 602 MPTCP_INC_STATS(sock_net(sk), MPTCP_MIB_JOINSYNACKBACKUPRX); in subflow_finish_connect() 608 MPTCP_INC_STATS(sock_net(s in subflow_finish_connect() [all...] |
| H A D | pm.c | 309 const struct net *net = sock_net((struct sock *)msk); in mptcp_adjust_add_addr_timeout() 495 MPTCP_INC_STATS(sock_net((struct sock *)msk), in mptcp_pm_remove_addr() 518 MPTCP_ADD_STATS(sock_net((struct sock *)msk), in mptcp_pm_new_connection() 700 __MPTCP_INC_STATS(sock_net((struct sock *)msk), MPTCP_MIB_ADDADDRDROP); in mptcp_pm_add_addr_echoed() 713 __MPTCP_INC_STATS(sock_net((struct sock *)msk), MPTCP_MIB_ADDADDRDROP); in mptcp_pm_add_addr_echoed() 797 __MPTCP_INC_STATS(sock_net(sk), rm_type); in mptcp_pm_rm_addr_received() 801 __MPTCP_INC_STATS(sock_net(sk), rm_type); in mptcp_pm_rm_addr_received() 834 __MPTCP_INC_STATS(sock_net((struct sock *)msk), MPTCP_MIB_RMADDRDROP); in mptcp_pm_mp_fail_received() 998 struct net *net = sock_net(sk); in mptcp_pm_subflows_chk_stale() 1084 u8 pm_type = mptcp_get_pm_type(sock_net((struc in mptcp_pm_data_reset() [all...] |
| /linux/net/tipc/ |
| H A D | socket.c | 288 u32 onode = tipc_own_addr(sock_net(sk)); in tipc_sk_respond() 296 tipc_node_xmit_skb(sock_net(sk), skb, dnode, selector); in tipc_sk_respond() 337 u32 self = tipc_own_addr(sock_net(sk)); in tsk_peer_msg() 552 struct net *net = sock_net(sk); in __tipc_shutdown() 760 addr->addr.id.node = tipc_own_addr(sock_net(sk)); in tipc_getname() 847 struct net *net = sock_net(sk); in tipc_sendmcast() 962 struct net *net = sock_net(sk); in tipc_send_group_unicast() 1010 struct net *net = sock_net(sk); in tipc_send_group_anycast() 1085 struct net *net = sock_net(sk); in tipc_send_group_bcast() 1162 struct net *net = sock_net(sk); in tipc_send_group_mcast() [all …]
|
| H A D | net.c | 204 struct net *net = sock_net(skb->sk); in tipc_nl_net_dump() 230 struct net *net = sock_net(skb->sk); in __tipc_nl_net_set() 327 struct net *net = sock_net(skb->sk); in tipc_nl_net_addr_legacy_get()
|
| /linux/net/rds/ |
| H A D | tcp_listen.c | 220 if (!rds_tcp_laddr_check(sock_net(listen_sock->sk), peer_addr, dev_if)) { in rds_tcp_accept_one() 226 conn = rds_conn_create(sock_net(listen_sock->sk), in rds_tcp_accept_one() 372 rds_tcp_accept_work(net_generic(sock_net(sk), rds_tcp_netid)); in rds_tcp_listen_data_ready() 374 ready = rds_tcp_listen_sock_def_readable(sock_net(sk)); in rds_tcp_listen_data_ready()
|
| /linux/net/netfilter/ |
| H A D | xt_cgroup.c | 130 if (!sk || !sk_fullsock(sk) || !net_eq(xt_net(par), sock_net(sk))) in cgroup_mt_v0() 146 if (!sk || !sk_fullsock(sk) || !net_eq(xt_net(par), sock_net(sk))) in cgroup_mt_v1() 167 if (!sk || !sk_fullsock(sk) || !net_eq(xt_net(par), sock_net(sk))) in cgroup_mt_v2()
|
| /linux/net/openvswitch/ |
| H A D | datapath.c | 613 struct net *net = sock_net(skb->sk); in ovs_packet_cmd_execute() 997 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_new() 1244 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_set() 1348 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_get() 1374 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_get() 1407 struct net *net = sock_net(skb->sk); in ovs_flow_cmd_del() 1429 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_del() 1470 netlink_set_err(sock_net(skb->sk)->genl_sock, 0, 0, in ovs_flow_cmd_del() 1499 dp = get_dp_rcu(sock_net(skb->sk), ovs_header->dp_ifindex); in ovs_flow_cmd_dump() 1674 dp = lookup_datapath(sock_net(sk in ovs_dp_reset_user_features() [all...] |
| /linux/net/phonet/ |
| H A D | pn_netlink.c | 55 struct net *net = sock_net(skb->sk); in addr_doit() 136 pndevs = phonet_device_list(sock_net(skb->sk)); in getaddr_dumpit() 234 struct net *net = sock_net(skb->sk); in route_doit() 295 struct net *net = sock_net(skb->sk); in route_dumpit()
|
| /linux/net/can/ |
| H A D | bcm.c | 309 dev = dev_get_by_index(sock_net(op->sk), op->ifindex); in bcm_can_tx() 859 dev = dev_get_by_index(sock_net(op->sk), in bcm_delete_rx_op() 867 can_rx_unregister(sock_net(op->sk), NULL, in bcm_delete_rx_op() 1293 dev = dev_get_by_index(sock_net(sk), ifindex); in bcm_rx_setup() 1295 err = can_rx_register(sock_net(sk), dev, in bcm_rx_setup() 1306 err = can_rx_register(sock_net(sk), NULL, op->can_id, in bcm_rx_setup() 1351 dev = dev_get_by_index(sock_net(sk), ifindex); in bcm_tx_send() 1415 dev = dev_get_by_index(sock_net(sk), ifindex); in bcm_sendmsg() 1494 if (!net_eq(dev_net(dev), sock_net(sk))) in bcm_notify() 1510 if (sock_net(sk)->can.bcmproc_dir && bo->bcm_proc_read) { in bcm_notify() [all …]
|