Searched refs:gso_skb (Results 1 – 5 of 5) sorted by relevance
| /linux/net/ipv4/ |
| H A D | udp_offload.c | 480 struct sk_buff *__udp_gso_segment(struct sk_buff *gso_skb, in __udp_gso_segment() argument 483 struct sock *sk = gso_skb->sk; in __udp_gso_segment() 493 mss = skb_shinfo(gso_skb)->gso_size; in __udp_gso_segment() 494 if (gso_skb->len <= sizeof(*uh) + mss) in __udp_gso_segment() 497 if (unlikely(skb_checksum_start(gso_skb) != in __udp_gso_segment() 498 skb_transport_header(gso_skb) && in __udp_gso_segment() 499 !(skb_shinfo(gso_skb)->gso_type & SKB_GSO_FRAGLIST))) in __udp_gso_segment() 505 if (gso_skb->ip_summed != CHECKSUM_PARTIAL) in __udp_gso_segment() 508 if (skb_gso_ok(gso_skb, features | NETIF_F_GSO_ROBUST)) { in __udp_gso_segment() 510 skb_shinfo(gso_skb)->gso_segs = DIV_ROUND_UP(gso_skb->len - sizeof(*uh), in __udp_gso_segment() [all …]
|
| H A D | tcp_offload.c | 16 static void tcp_gso_tstamp(struct sk_buff *skb, struct sk_buff *gso_skb, in tcp_gso_tstamp() argument 19 u32 flags = skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP; in tcp_gso_tstamp() 20 u32 ts_seq = skb_shinfo(gso_skb)->tskey; in tcp_gso_tstamp() 143 struct sk_buff *gso_skb = skb; in tcp_gso_segment() local 176 copy_destructor = gso_skb->destructor == tcp_wfree; in tcp_gso_segment() 177 ooo_okay = gso_skb->ooo_okay; in tcp_gso_segment() 201 if (unlikely(skb_shinfo(gso_skb)->tx_flags & SKBTX_ANY_TSTAMP)) in tcp_gso_segment() 202 tcp_gso_tstamp(segs, gso_skb, seq, mss); in tcp_gso_segment() 206 ecn_cwr_mask = !!(skb_shinfo(gso_skb)->gso_type & SKB_GSO_TCP_ACCECN); in tcp_gso_segment() 219 skb->destructor = gso_skb->destructor; in tcp_gso_segment() [all …]
|
| /linux/net/sched/ |
| H A D | sch_generic.c | 156 __skb_queue_tail(&q->gso_skb, skb); in dev_requeue_skb() 240 if (unlikely(!skb_queue_empty(&q->gso_skb))) { in dequeue_skb() 248 skb = skb_peek(&q->gso_skb); in dequeue_skb() 266 skb = __skb_dequeue(&q->gso_skb); in dequeue_skb() 672 .gso_skb = { 673 .next = (struct sk_buff *)&noop_qdisc.gso_skb, 674 .prev = (struct sk_buff *)&noop_qdisc.gso_skb, 676 .lock = __SPIN_LOCK_UNLOCKED(noop_qdisc.gso_skb.lock), 955 __skb_queue_head_init(&sch->gso_skb); in qdisc_alloc() 1036 __skb_queue_purge(&qdisc->gso_skb); in qdisc_reset()
|
| /linux/include/net/ |
| H A D | sch_generic.h | 110 struct sk_buff_head gso_skb; 1065 skb = __skb_dequeue(&sch->gso_skb); 1189 struct sk_buff *skb = skb_peek(&sch->gso_skb); in qdisc_dequeue_peeked() 1191 /* we can reuse ->gso_skb because peek isn't called for root qdiscs */ in qdisc_dequeue_peeked() 1196 __skb_queue_head(&sch->gso_skb, skb); in qdisc_dequeue_peeked() 1235 struct sk_buff *skb = skb_peek(&sch->gso_skb); in qdisc_replace() 1238 skb = __skb_dequeue(&sch->gso_skb); in qdisc_replace()
|
| /linux/net/netfilter/ |
| H A D | nfnetlink_queue.c | 978 struct sk_buff *gso_skb = entry->skb; in __nfqnl_enqueue_packet_gso() local 982 entry->skb = gso_skb; in __nfqnl_enqueue_packet_gso()
|