/linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
H A D | cn10k.h | 13 static inline int mtu_to_dwrr_weight(struct otx2_nic *pfvf, int mtu) in mtu_to_dwrr_weight() argument 20 weight = mtu / pfvf->hw.dwrr_mtu; in mtu_to_dwrr_weight() 21 if (mtu % pfvf->hw.dwrr_mtu) in mtu_to_dwrr_weight()
|
/linux/net/batman-adv/ |
H A D | fragmentation.c | 368 if (total_size > neigh_node->if_incoming->net_dev->mtu) { in batadv_frag_skb_fwd() 405 unsigned int mtu = fragment_size + header_size; in batadv_frag_create() local 407 skb_fragment = dev_alloc_skb(ll_reserved + mtu + tailroom); in batadv_frag_create() 442 unsigned int mtu = net_dev->mtu; in batadv_frag_send_packet() local 450 mtu = min_t(unsigned int, mtu, BATADV_FRAG_MAX_FRAG_SIZE); in batadv_frag_send_packet() 451 max_fragment_size = mtu - header_size; in batadv_frag_send_packet()
|
/linux/drivers/infiniband/hw/qib/ |
H A D | qib_qp.c | 265 static int mtu_to_enum(u32 mtu) in mtu_to_enum() argument 269 switch (mtu) { in mtu_to_enum() 294 int mtu, pmtu, pidx = qp->port_num - 1; in qib_get_pmtu_from_attr() local 298 mtu = ib_mtu_enum_to_int(attr->path_mtu); in qib_get_pmtu_from_attr() 299 if (mtu == -1) in qib_get_pmtu_from_attr() 302 if (mtu > dd->pport[pidx].ibmtu) in qib_get_pmtu_from_attr() 309 int qib_mtu_to_path_mtu(u32 mtu) in qib_mtu_to_path_mtu() argument 311 return mtu_to_enum(mtu); in qib_mtu_to_path_mtu()
|
/linux/drivers/net/ |
H A D | ntb_netdev.c | 130 skb = netdev_alloc_skb(ndev, ndev->mtu + ETH_HLEN); in ntb_netdev_rx_handler() 138 rc = ntb_transport_rx_enqueue(qp, skb, skb->data, ndev->mtu + ETH_HLEN); in ntb_netdev_rx_handler() 255 skb = netdev_alloc_skb(ndev, ndev->mtu + ETH_HLEN); in ntb_netdev_open() 262 ndev->mtu + ETH_HLEN); in ntb_netdev_open() 309 WRITE_ONCE(ndev->mtu, new_mtu); in ntb_netdev_change_mtu() 316 if (ndev->mtu < new_mtu) { in ntb_netdev_change_mtu() 338 WRITE_ONCE(ndev->mtu, new_mtu); in ntb_netdev_change_mtu() 446 ndev->mtu = ntb_transport_max_size(dev->qp) - ETH_HLEN; in ntb_netdev_probe()
|
/linux/net/bridge/ |
H A D | br_nf_core.c | 25 struct sk_buff *skb, u32 mtu, in fake_update_pmtu() argument 49 return dst->dev->mtu; in fake_mtu() 58 .mtu = fake_mtu,
|
/linux/drivers/net/ethernet/ti/icssg/ |
H A D | icssg_mii_cfg.c | 28 void icssg_mii_update_mtu(struct regmap *mii_rt, int mii, int mtu) in icssg_mii_update_mtu() argument 30 mtu += (ETH_HLEN + ETH_FCS_LEN); in icssg_mii_update_mtu() 35 (mtu - 1) << PRUSS_MII_RT_RX_FRMS_MAX_FRM_SHIFT); in icssg_mii_update_mtu() 40 (mtu - 1) << PRUSS_MII_RT_RX_FRMS_MAX_FRM_SHIFT); in icssg_mii_update_mtu()
|
/linux/net/ipv4/ |
H A D | ip_tunnel_core.c | 196 static int iptunnel_pmtud_build_icmp(struct sk_buff *skb, int mtu) in iptunnel_pmtud_build_icmp() argument 226 .un.frag.mtu = htons(mtu), in iptunnel_pmtud_build_icmp() 262 static int iptunnel_pmtud_check_icmp(struct sk_buff *skb, int mtu) in iptunnel_pmtud_check_icmp() argument 267 if (mtu < 576 || iph->frag_off != htons(IP_DF)) in iptunnel_pmtud_check_icmp() 278 return iptunnel_pmtud_build_icmp(skb, mtu); in iptunnel_pmtud_check_icmp() 289 static int iptunnel_pmtud_build_icmpv6(struct sk_buff *skb, int mtu) in iptunnel_pmtud_build_icmpv6() argument 319 .icmp6_mtu = htonl(mtu), in iptunnel_pmtud_build_icmpv6() 355 static int iptunnel_pmtud_check_icmpv6(struct sk_buff *skb, int mtu) in iptunnel_pmtud_check_icmpv6() argument 363 if (mtu < IPV6_MIN_MTU) in iptunnel_pmtud_check_icmpv6() 388 return iptunnel_pmtud_build_icmpv6(skb, mtu); in iptunnel_pmtud_check_icmpv6() [all …]
|
/linux/include/linux/can/ |
H A D | dev.h | 140 dev->mtu = CANFD_MTU; in can_set_static_ctrlmode() 150 static inline bool can_is_canxl_dev_mtu(unsigned int mtu) in can_is_canxl_dev_mtu() argument 152 return (mtu >= CANXL_MIN_MTU && mtu <= CANXL_MAX_MTU); in can_is_canxl_dev_mtu()
|
/linux/drivers/net/ethernet/cisco/enic/ |
H A D | enic_res.c | 56 GET_CONFIG(mtu); in enic_get_vnic_config() 75 if (c->mtu == 0) in enic_get_vnic_config() 76 c->mtu = 1500; in enic_get_vnic_config() 77 c->mtu = min_t(u16, ENIC_MAX_MTU, in enic_get_vnic_config() 79 c->mtu)); in enic_get_vnic_config() 86 enic->mac_addr, c->wq_desc_count, c->rq_desc_count, c->mtu); in enic_get_vnic_config()
|
/linux/tools/testing/selftests/net/mptcp/ |
H A D | simult_flows.sh | 79 ip -net "$ns1" link set ns1eth1 up mtu 1500 85 ip -net "$ns1" link set ns1eth2 up mtu 1500 94 ip -net "$ns2" link set ns2eth1 up mtu 1500 98 ip -net "$ns2" link set ns2eth2 up mtu 1500 102 ip -net "$ns2" link set ns2eth3 up mtu 1500 108 ip -net "$ns3" link set ns3eth1 up mtu 1500
|
/linux/net/ipv6/ |
H A D | ip6_tunnel.c | 527 __u32 mtu; in ip6_tnl_err() local 531 mtu = *info - offset; in ip6_tnl_err() 532 if (mtu < IPV6_MIN_MTU) in ip6_tnl_err() 533 mtu = IPV6_MIN_MTU; in ip6_tnl_err() 535 if (len > mtu) { in ip6_tnl_err() 538 rel_info = mtu; in ip6_tnl_err() 1095 int mtu; in ip6_tnl_xmit() local 1190 mtu = dst_mtu(dst) - eth_hlen - psh_hlen - t->tun_hlen; in ip6_tnl_xmit() 1193 mtu -= 8; in ip6_tnl_xmit() 1195 mtu = max(mtu, skb->protocol == htons(ETH_P_IPV6) ? in ip6_tnl_xmit() [all …]
|
H A D | ip6_vti.c | 449 int mtu; in vti6_xmit() local 509 mtu = dst_mtu(dst); in vti6_xmit() 510 if (skb->len > mtu) { in vti6_xmit() 511 skb_dst_update_pmtu_no_confirm(skb, mtu); in vti6_xmit() 514 if (mtu < IPV6_MIN_MTU) in vti6_xmit() 515 mtu = IPV6_MIN_MTU; in vti6_xmit() 517 icmpv6_ndo_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu); in vti6_xmit() 522 htonl(mtu)); in vti6_xmit() 654 int mtu; in vti6_link_config() local 668 if (keep_mtu && dev->mtu) { in vti6_link_config() [all …]
|
H A D | ip6_gre.c | 825 __u32 mtu; in ip6gre_xmit_ipv4() local 839 err = __gre6_xmit(skb, dev, dsfield, &fl6, encap_limit, &mtu, in ip6gre_xmit_ipv4() 845 htonl(mtu)); in ip6gre_xmit_ipv4() 859 __u32 mtu; in ip6gre_xmit_ipv6() local 874 &mtu, skb->protocol); in ip6gre_xmit_ipv6() 877 icmpv6_ndo_send(skb, ICMPV6_PKT_TOOBIG, 0, mtu); in ip6gre_xmit_ipv6() 890 __u32 mtu; in ip6gre_xmit_other() local 901 err = __gre6_xmit(skb, dev, dsfield, &fl6, encap_limit, &mtu, skb->protocol); in ip6gre_xmit_other() 958 __u32 mtu; in ip6erspan_tunnel_xmit() local 970 if (skb->len > dev->mtu + dev->hard_header_len) { in ip6erspan_tunnel_xmit() [all …]
|
/linux/drivers/net/wwan/ |
H A D | mhi_wwan_ctrl.c | 25 size_t mtu; member 86 skb = alloc_skb(mhiwwan->mtu, GFP_KERNEL); in mhi_wwan_ctrl_refill_work() 99 if (mhi_queue_skb(mhi_dev, DMA_FROM_DEVICE, skb, mhiwwan->mtu, MHI_EOT)) { in mhi_wwan_ctrl_refill_work() 147 if (skb->len > mhiwwan->mtu) in mhi_wwan_ctrl_tx() 226 mhiwwan->mtu = MHI_WWAN_MAX_MTU; in mhi_wwan_ctrl_probe()
|
/linux/tools/testing/selftests/wireguard/ |
H A D | netns.sh | 159 [[ $(ip1 link show dev wg0) =~ mtu\ ([0-9]+) ]] && orig_mtu="${BASH_REMATCH[1]}" 179 ip1 link set wg0 mtu $big_mtu 180 ip2 link set wg0 mtu $big_mtu 183 ip1 link set wg0 mtu $orig_mtu 184 ip2 link set wg0 mtu $orig_mtu 190 ip1 link set wg0 mtu $big_mtu 191 ip2 link set wg0 mtu $big_mtu 195 ip1 link set wg0 mtu 1300 196 ip2 link set wg0 mtu 1300 200 n1 ip route add 192.168.241.2/32 dev wg0 mtu 1299 [all …]
|
/linux/drivers/net/slip/ |
H A D | slip.c | 130 static int sl_alloc_bufs(struct slip *sl, int mtu) in sl_alloc_bufs() argument 148 len = mtu * 2; in sl_alloc_bufs() 177 sl->mtu = mtu; in sl_alloc_bufs() 221 static int sl_realloc_bufs(struct slip *sl, int mtu) argument 229 int len = mtu * 2; 251 if (mtu > sl->mtu) { 288 sl->mtu = mtu; 289 WRITE_ONCE(dev->mtu, mtu); 381 if (len > sl->mtu) { /* Sigh, shouldn't occur BUT ... */ 617 dev->mtu = sl->mtu;
|
/linux/drivers/infiniband/sw/rxe/ |
H A D | rxe_req.c | 23 int to_send = (wqe->dma.resid > qp->mtu) ? in retry_first_write_send() 24 qp->mtu : wqe->dma.resid; in retry_first_write_send() 90 qp->mtu; in req_retry() 91 wqe->iova += npsn * qp->mtu; in req_retry() 357 int fits = (wqe->dma.resid <= qp->mtu); in next_opcode() 409 return qp->mtu; in get_mtu() 562 int num_pkt = (wqe->dma.resid + payload + qp->mtu - 1) / qp->mtu; in update_wqe_psn() 648 int mtu; in rxe_requester() local 740 mtu = get_mtu(qp); in rxe_requester() 743 if (payload > mtu) { in rxe_requester() [all …]
|
/linux/net/tipc/ |
H A D | udp_media.h | 49 static inline bool tipc_udp_mtu_bad(u32 mtu) in tipc_udp_mtu_bad() argument 51 if (mtu >= (TIPC_MIN_BEARER_MTU + sizeof(struct iphdr) + in tipc_udp_mtu_bad()
|
H A D | bearer.c | 467 b->mtu = dev->mtu; in tipc_enable_l2_media() 535 int mtu = 0; in tipc_bearer_mtu() local 541 mtu = b->mtu; in tipc_bearer_mtu() 543 return mtu; in tipc_bearer_mtu() 548 int mtu = TIPC_MIN_BEARER_MTU; in tipc_bearer_min_mtu() local 554 mtu += b->encap_hlen; in tipc_bearer_min_mtu() 556 return mtu; in tipc_bearer_min_mtu() 717 b->mtu = dev->mtu; in tipc_l2_device_event() 852 if (nla_put_u32(msg->skb, TIPC_NLA_PROP_MTU, bearer->mtu)) in __tipc_nl_add_bearer() 1165 b->mtu = nla_get_u32(props[TIPC_NLA_PROP_MTU]); in __tipc_nl_bearer_set() [all …]
|
/linux/tools/testing/selftests/net/netfilter/ |
H A D | conntrack_ipip_mtu.sh | 144 ip -net "$r_w" link set dev veth0 up mtu 1400 145 ip -net "$r_w" link set dev veth1 up mtu 1400 147 ip -net "$r_a" link set dev veth0 mtu 1400 148 ip -net "$r_b" link set dev veth0 mtu 1400
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/tc/act/ |
H A D | police.c | 69 } else if (act->police.mtu) { in fill_meter_params_from_act() 70 params->mtu = act->police.mtu; in fill_meter_params_from_act() 92 if (params->mtu) { in tc_act_parse_police()
|
/linux/net/netfilter/ |
H A D | nf_flow_table_ip.c | 246 static bool nf_flow_exceeds_mtu(const struct sk_buff *skb, unsigned int mtu) in nf_flow_exceeds_mtu() argument 248 if (skb->len <= mtu) in nf_flow_exceeds_mtu() 251 if (skb_is_gso(skb) && skb_gso_validate_network_len(skb, mtu)) in nf_flow_exceeds_mtu() 373 unsigned int thoff, mtu; in nf_flow_offload_forward() local 379 mtu = flow->tuplehash[dir].tuple.mtu + ctx->offset; in nf_flow_offload_forward() 380 if (unlikely(nf_flow_exceeds_mtu(skb, mtu))) in nf_flow_offload_forward() 652 unsigned int thoff, mtu; in nf_flow_offload_ipv6_forward() local 658 mtu = flow->tuplehash[dir].tuple.mtu + ctx->offset; in nf_flow_offload_ipv6_forward() 659 if (unlikely(nf_flow_exceeds_mtu(skb, mtu))) in nf_flow_offload_ipv6_forward()
|
/linux/drivers/net/wireguard/ |
H A D | send.c | 145 if (unlikely(!PACKET_CB(skb)->mtu)) in calculate_skb_padding() 154 if (unlikely(last_unit > PACKET_CB(skb)->mtu)) in calculate_skb_padding() 155 last_unit %= PACKET_CB(skb)->mtu; in calculate_skb_padding() 157 padded_size = min(PACKET_CB(skb)->mtu, in calculate_skb_padding() 232 PACKET_CB(skb)->mtu = skb->dev->mtu; in wg_packet_send_keepalive()
|
/linux/tools/testing/selftests/net/ |
H A D | udpgso.c | 313 unsigned int mtu; in get_path_mtu() local 316 vallen = sizeof(mtu); in get_path_mtu() 318 ret = getsockopt(fd, SOL_IP, IP_MTU, &mtu, &vallen); in get_path_mtu() 320 ret = getsockopt(fd, SOL_IPV6, IPV6_MTU, &mtu, &vallen); in get_path_mtu() 326 fprintf(stderr, "path mtu (read): %u\n", mtu); in get_path_mtu() 327 return mtu; in get_path_mtu()
|
/linux/include/net/sctp/ |
H A D | sctp.h | 574 __u32 mtu, __u32 extra) in __sctp_mtu_payload() argument 586 if (WARN_ON_ONCE(mtu && mtu <= overhead)) in __sctp_mtu_payload() 587 mtu = overhead; in __sctp_mtu_payload() 589 return mtu ? mtu - overhead : overhead; in __sctp_mtu_payload() 593 __u32 mtu, __u32 extra) in sctp_mtu_payload() argument 595 return __sctp_mtu_payload(sp, NULL, mtu, extra); in sctp_mtu_payload()
|