| /linux/include/net/ |
| H A D | hotdata.h | 16 struct net_hotdata { struct 53 #define inet_ehash_secret net_hotdata.tcp_protocol.secret argument 54 #define udp_ehash_secret net_hotdata.udp_protocol.secret 55 #define inet6_ehash_secret net_hotdata.tcpv6_protocol.secret 56 #define tcp_ipv6_hash_secret net_hotdata.tcpv6_offload.secret 57 #define udp6_ehash_secret net_hotdata.udpv6_protocol.secret 58 #define udp_ipv6_hash_secret net_hotdata.udpv6_offload.secret 60 extern struct net_hotdata net_hotdata;
|
| /linux/net/core/ |
| H A D | hotdata.c | 10 struct net_hotdata net_hotdata __cacheline_aligned = { 11 .offload_base = LIST_HEAD_INIT(net_hotdata.offload_base), 27 EXPORT_SYMBOL(net_hotdata);
|
| H A D | sysctl_net_core.c | 154 net_hotdata.rps_sock_flow_table, in rps_sock_flow_sysctl() 175 net_hotdata.rps_cpu_mask = in rps_sock_flow_sysctl() 187 rcu_assign_pointer(net_hotdata.rps_sock_flow_table, in rps_sock_flow_sysctl() 318 WRITE_ONCE(net_hotdata.dev_rx_weight, weight * dev_weight_rx_bias); in proc_do_dev_weight() 319 WRITE_ONCE(net_hotdata.dev_tx_weight, weight * dev_weight_tx_bias); in proc_do_dev_weight() 402 .data = &net_hotdata.sysctl_mem_pcpu_rsv, 434 .data = &net_hotdata.max_backlog, 441 .data = &net_hotdata.qdisc_max_burst, 500 .data = &net_hotdata.tstamp_prequeue, 569 .data = &net_hotdata.netdev_budget, [all …]
|
| H A D | gso.c | 20 list_for_each_entry_rcu(ptype, &net_hotdata.offload_base, list) { in skb_eth_gso_segment() 51 list_for_each_entry_rcu(ptype, &net_hotdata.offload_base, list) { in skb_mac_gso_segment()
|
| H A D | gro.c | 30 list_for_each_entry(elem, &net_hotdata.offload_base, list) { in dev_add_offload() 54 struct list_head *head = &net_hotdata.offload_base; in __dev_remove_offload() 254 struct list_head *head = &net_hotdata.offload_base; in gro_complete() 466 struct list_head *head = &net_hotdata.offload_base; in dev_gro_receive() 572 struct list_head *offload_head = &net_hotdata.offload_base; in gro_find_receive_by_type() 586 struct list_head *offload_head = &net_hotdata.offload_base; in gro_find_complete_by_type()
|
| H A D | skbuff.c | 292 nc->skb_count = kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, in napi_skb_cache_get() 356 nc->skb_count += kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, in napi_skb_cache_get_bulk() 363 n -= kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, in napi_skb_cache_get_bulk() 440 skb = kmem_cache_alloc(net_hotdata.skbuff_cache, in slab_build_skb() 492 skb = kmem_cache_alloc(net_hotdata.skbuff_cache, in __build_skb() 592 return kmem_cache_alloc_node(net_hotdata.skb_small_head_cache, in kmalloc_pfmemalloc() 613 obj = kmem_cache_alloc_node(net_hotdata.skb_small_head_cache, in kmalloc_reserve() 683 cache = net_hotdata.skbuff_fclone_cache; in __alloc_skb() 686 cache = net_hotdata.skbuff_cache; in __alloc_skb() 1087 kmem_cache_free(net_hotdata.skb_small_head_cache, head); in skb_kfree_head() [all …]
|
| H A D | gro_cells.c | 34 if (skb_queue_len(&cell->napi_skbs) > READ_ONCE(net_hotdata.max_backlog)) { in gro_cells_receive()
|
| H A D | dev.c | 4222 if (unlikely(defer_count > READ_ONCE(net_hotdata.qdisc_max_burst))) { in __dev_xmit_skb() 5120 sock_flow_table = rcu_dereference(net_hotdata.rps_sock_flow_table); in get_rps_cpu() 5130 if ((ident ^ hash) & ~net_hotdata.rps_cpu_mask) in get_rps_cpu() 5133 next_cpu = ident & net_hotdata.rps_cpu_mask; in get_rps_cpu() 5357 max_backlog = READ_ONCE(net_hotdata.max_backlog); in enqueue_to_backlog() 5663 net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue), skb); in netif_rx_internal() 5958 net_timestamp_check(!READ_ONCE(net_hotdata.tstamp_prequeue), skb); in __netif_receive_skb_core() 6357 net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue), skb); in netif_receive_skb_internal() 6386 net_timestamp_check(READ_ONCE(net_hotdata.tstamp_prequeue), in netif_receive_skb_list_internal() 6632 napi->weight = READ_ONCE(net_hotdata in process_backlog() [all...] |
| H A D | xdp.c | 854 skb = kmem_cache_alloc(net_hotdata.skbuff_cache, GFP_ATOMIC); in xdp_build_skb_from_frame()
|
| /linux/net/ipv6/ |
| H A D | tcpv6_offload.c | 196 net_hotdata.tcpv6_offload = (struct net_offload) { in tcpv6_offload_init() 203 return inet6_add_offload(&net_hotdata.tcpv6_offload, IPPROTO_TCP); in tcpv6_offload_init()
|
| H A D | ip6_offload.c | 347 if (likely(ops == &net_hotdata.tcpv6_offload)) in ipv6_gro_complete() 350 if (ops == &net_hotdata.udpv6_offload) in ipv6_gro_complete() 443 net_hotdata.ipv6_packet_offload = (struct packet_offload) { 451 dev_add_offload(&net_hotdata.ipv6_packet_offload);
|
| H A D | tcp_ipv6.c | 2368 net_hotdata.tcpv6_protocol = (struct inet6_protocol) { 2373 ret = inet6_add_protocol(&net_hotdata.tcpv6_protocol, IPPROTO_TCP); 2398 inet6_del_protocol(&net_hotdata.tcpv6_protocol, IPPROTO_TCP); 2406 inet6_del_protocol(&net_hotdata.tcpv6_protocol, IPPROTO_TCP);
|
| /linux/net/ipv4/ |
| H A D | af_inet.c | 1876 net_hotdata.ip_packet_offload = (struct packet_offload) { in ipv4_offload_init() 1884 dev_add_offload(&net_hotdata.ip_packet_offload); in ipv4_offload_init() 1939 net_hotdata.udp_protocol = (struct net_protocol) { in inet_init() 1944 if (inet_add_protocol(&net_hotdata.udp_protocol, IPPROTO_UDP) < 0) in inet_init() 1947 net_hotdata.tcp_protocol = (struct net_protocol) { in inet_init() 1953 if (inet_add_protocol(&net_hotdata.tcp_protocol, IPPROTO_TCP) < 0) in inet_init()
|
| H A D | tcp_offload.c | 470 net_hotdata.tcpv4_offload = (struct net_offload) { in tcpv4_offload_init() 477 return inet_add_offload(&net_hotdata.tcpv4_offload, IPPROTO_TCP); in tcpv4_offload_init()
|
| H A D | udp_offload.c | 990 net_hotdata.udpv4_offload = (struct net_offload) { in udpv4_offload_init() 998 return inet_add_offload(&net_hotdata.udpv4_offload, IPPROTO_UDP);
|
| H A D | tcp.c | 1305 if (i >= READ_ONCE(net_hotdata.sysctl_max_skb_frags)) { in tcp_sendmsg_locked()
|
| /linux/net/xfrm/ |
| H A D | espintcp.c | 174 READ_ONCE(net_hotdata.max_backlog)) { in espintcp_queue_out()
|
| H A D | xfrm_input.c | 804 if (skb_queue_len(&trans->queue) >= READ_ONCE(net_hotdata.max_backlog)) in xfrm_trans_queue_net()
|
| /linux/drivers/net/ovpn/ |
| H A D | tcp.c | 355 READ_ONCE(net_hotdata.max_backlog)) { in ovpn_tcp_send_skb()
|
| /linux/net/sched/ |
| H A D | sch_generic.c | 417 int quota = READ_ONCE(net_hotdata.dev_tx_weight); in __qdisc_run()
|
| /linux/net/bpf/ |
| H A D | test_run.c | 249 n = kmem_cache_alloc_bulk(net_hotdata.skbuff_cache, gfp, nframes, in xdp_recv_frames()
|
| /linux/net/mptcp/ |
| H A D | protocol.c | 1318 if (!can_coalesce && i >= READ_ONCE(net_hotdata.sysctl_max_skb_frags)) { in mptcp_sendmsg_frag()
|