Home
last modified time | relevance | path

Searched refs:gso_skb (Results 1 – 5 of 5) sorted by relevance

/linux/net/ipv4/
H A Dudp_offload.c480 struct sk_buff *__udp_gso_segment(struct sk_buff *gso_skb, in __udp_gso_segment() argument
483 struct sock *sk = gso_skb->sk; in __udp_gso_segment()
493 mss = skb_shinfo(gso_skb)->gso_size; in __udp_gso_segment()
494 if (gso_skb->len <= sizeof(*uh) + mss) in __udp_gso_segment()
497 if (unlikely(skb_checksum_start(gso_skb) != in __udp_gso_segment()
498 skb_transport_header(gso_skb) && in __udp_gso_segment()
499 !(skb_shinfo(gso_skb)->gso_type & SKB_GSO_FRAGLIST))) in __udp_gso_segment()
505 if (gso_skb->ip_summed != CHECKSUM_PARTIAL) in __udp_gso_segment()
508 if (skb_gso_ok(gso_skb, features | NETIF_F_GSO_ROBUST)) { in __udp_gso_segment()
510 skb_shinfo(gso_skb)->gso_segs = DIV_ROUND_UP(gso_skb->len - sizeof(*uh), in __udp_gso_segment()
[all …]
H A Dtcp_offload.c16 static void tcp_gso_tstamp(struct sk_buff *skb, struct sk_buff *gso_skb, in tcp_gso_tstamp() argument
19 u32 flags = skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP; in tcp_gso_tstamp()
20 u32 ts_seq = skb_shinfo(gso_skb)->tskey; in tcp_gso_tstamp()
143 struct sk_buff *gso_skb = skb; in tcp_gso_segment() local
176 copy_destructor = gso_skb->destructor == tcp_wfree; in tcp_gso_segment()
177 ooo_okay = gso_skb->ooo_okay; in tcp_gso_segment()
201 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP)) in tcp_gso_segment()
202 tcp_gso_tstamp(segs, gso_skb, seq, mss); in tcp_gso_segment()
206 ecn_cwr_mask = !!(skb_shinfo(gso_skb)->gso_type & SKB_GSO_TCP_ACCECN); in tcp_gso_segment()
219 skb->destructor = gso_skb->destructor; in tcp_gso_segment()
[all …]
/linux/net/sched/
H A Dsch_generic.c156 __skb_queue_tail(&q->gso_skb, skb); in dev_requeue_skb()
240 if (unlikely(!skb_queue_empty(&q->gso_skb))) { in dequeue_skb()
248 skb = skb_peek(&q->gso_skb); in dequeue_skb()
266 skb = __skb_dequeue(&q->gso_skb); in dequeue_skb()
672 .gso_skb = {
673 .next = (struct sk_buff *)&noop_qdisc.gso_skb,
674 .prev = (struct sk_buff *)&noop_qdisc.gso_skb,
676 .lock = __SPIN_LOCK_UNLOCKED(noop_qdisc.gso_skb.lock),
955 __skb_queue_head_init(&sch->gso_skb); in qdisc_alloc()
1036 __skb_queue_purge(&qdisc->gso_skb); in qdisc_reset()
/linux/include/net/
H A Dsch_generic.h110 struct sk_buff_head gso_skb;
1065 skb = __skb_dequeue(&sch->gso_skb);
1189 struct sk_buff *skb = skb_peek(&sch->gso_skb); in qdisc_dequeue_peeked()
1191 /* we can reuse ->gso_skb because peek isn't called for root qdiscs */ in qdisc_dequeue_peeked()
1196 __skb_queue_head(&sch->gso_skb, skb); in qdisc_dequeue_peeked()
1235 struct sk_buff *skb = skb_peek(&sch->gso_skb); in qdisc_replace()
1238 skb = __skb_dequeue(&sch->gso_skb); in qdisc_replace()
/linux/net/netfilter/
H A Dnfnetlink_queue.c978 struct sk_buff *gso_skb = entry->skb; in __nfqnl_enqueue_packet_gso() local
982 entry->skb = gso_skb; in __nfqnl_enqueue_packet_gso()