/linux/net/core/ |
H A D | gro.c | 91 int skb_gro_receive(struct sk_buff *p, struct sk_buff *skb) in skb_gro_receive() argument 93 struct skb_shared_info *pinfo, *skbinfo = skb_shinfo(skb); in skb_gro_receive() 94 unsigned int offset = skb_gro_offset(skb); in skb_gro_receive() 95 unsigned int headlen = skb_headlen(skb); in skb_gro_receive() 96 unsigned int len = skb_gro_len(skb); in skb_gro_receive() 108 if (p->pp_recycle != skb->pp_recycle) in skb_gro_receive() 112 NAPI_GRO_CB(skb)->flush)) in skb_gro_receive() 116 if (NAPI_GRO_CB(skb)->proto != IPPROTO_TCP || in skb_gro_receive() 123 segs = NAPI_GRO_CB(skb)->count; in skb_gro_receive() 150 new_truesize = SKB_TRUESIZE(skb_end_offset(skb)); in skb_gro_receive() [all …]
|
H A D | skbuff.c | 202 static void skb_panic(struct sk_buff *skb, unsigned int sz, void *addr, in skb_panic() argument 206 msg, addr, skb->len, sz, skb->head, skb->data, in skb_panic() 207 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic() 208 skb->dev ? skb->dev->name : "<NULL>"); in skb_panic() 212 static void skb_over_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_over_panic() argument 214 skb_panic(skb, sz, addr, __func__); in skb_over_panic() 217 static void skb_under_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_under_panic() argument 219 skb_panic(skb, sz, addr, __func__); in skb_under_panic() 277 struct sk_buff *skb; in napi_skb_cache_get() local 291 skb = nc->skb_cache[--nc->skb_count]; in napi_skb_cache_get() [all …]
|
H A D | gso.c | 13 struct sk_buff *skb_eth_gso_segment(struct sk_buff *skb, in skb_eth_gso_segment() argument 22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment() 37 struct sk_buff *skb_mac_gso_segment(struct sk_buff *skb, in skb_mac_gso_segment() argument 42 int vlan_depth = skb->mac_len; in skb_mac_gso_segment() 43 __be16 type = skb_network_protocol(skb, &vlan_depth); in skb_mac_gso_segment() 48 __skb_pull(skb, vlan_depth); in skb_mac_gso_segment() 53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment() 59 __skb_push(skb, skb->data - skb_mac_header(skb)); in skb_mac_gso_segment() 66 static bool skb_needs_check(const struct sk_buff *skb, bool tx_path) in skb_needs_check() argument 69 return skb->ip_summed != CHECKSUM_PARTIAL && in skb_needs_check() [all …]
|
/linux/net/xfrm/ |
H A D | xfrm_output.c | 28 static int xfrm_output2(struct net *net, struct sock *sk, struct sk_buff *skb); 29 static int xfrm_inner_extract_output(struct xfrm_state *x, struct sk_buff *skb); 31 static int xfrm_skb_check_space(struct sk_buff *skb) in xfrm_skb_check_space() argument 33 struct dst_entry *dst = skb_dst(skb); in xfrm_skb_check_space() 35 - skb_headroom(skb); in xfrm_skb_check_space() 36 int ntail = dst->dev->needed_tailroom - skb_tailroom(skb); in xfrm_skb_check_space() 45 return pskb_expand_head(skb, nhead, ntail, GFP_ATOMIC); in xfrm_skb_check_space() 52 static struct dst_entry *skb_dst_pop(struct sk_buff *skb) in skb_dst_pop() argument 54 struct dst_entry *child = dst_clone(xfrm_dst_child(skb_dst(skb))); in skb_dst_pop() 56 skb_dst_drop(skb); in skb_dst_pop() [all …]
|
/linux/drivers/net/can/dev/ |
H A D | skb.c | 47 int can_put_echo_skb(struct sk_buff *skb, struct net_device *dev, in can_put_echo_skb() argument 60 (skb->protocol != htons(ETH_P_CAN) && in can_put_echo_skb() 61 skb->protocol != htons(ETH_P_CANFD) && in can_put_echo_skb() 62 skb->protocol != htons(ETH_P_CANXL))) { in can_put_echo_skb() 63 kfree_skb(skb); in can_put_echo_skb() 68 skb = can_create_echo_skb(skb); in can_put_echo_skb() 69 if (!skb) in can_put_echo_skb() 73 skb->ip_summed = CHECKSUM_UNNECESSARY; in can_put_echo_skb() 74 skb->dev = dev; in can_put_echo_skb() 77 can_skb_prv(skb)->frame_len = frame_len; in can_put_echo_skb() [all …]
|
/linux/net/ipv6/ |
H A D | exthdrs.c | 65 static bool ip6_tlvopt_unknown(struct sk_buff *skb, int optoff, in ip6_tlvopt_unknown() argument 79 switch ((skb_network_header(skb)[optoff] & 0xC0) >> 6) { in ip6_tlvopt_unknown() 90 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) in ip6_tlvopt_unknown() 94 icmpv6_param_prob_reason(skb, ICMPV6_UNK_OPTION, optoff, in ip6_tlvopt_unknown() 100 kfree_skb_reason(skb, SKB_DROP_REASON_UNHANDLED_PROTO); in ip6_tlvopt_unknown() 104 static bool ipv6_hop_ra(struct sk_buff *skb, int optoff); 105 static bool ipv6_hop_ioam(struct sk_buff *skb, int optoff); 106 static bool ipv6_hop_jumbo(struct sk_buff *skb, int optoff); 107 static bool ipv6_hop_calipso(struct sk_buff *skb, int optoff); 109 static bool ipv6_dest_hao(struct sk_buff *skb, int optoff); [all …]
|
H A D | ip6_input.c | 49 struct sk_buff *skb) in ip6_rcv_finish_core() argument 52 !skb_dst(skb) && !skb->sk) { in ip6_rcv_finish_core() 53 switch (ipv6_hdr(skb)->nexthdr) { in ip6_rcv_finish_core() 56 tcp_v6_early_demux(skb); in ip6_rcv_finish_core() 60 udp_v6_early_demux(skb); in ip6_rcv_finish_core() 65 if (!skb_valid_dst(skb)) in ip6_rcv_finish_core() 66 ip6_route_input(skb); in ip6_rcv_finish_core() 69 int ip6_rcv_finish(struct net *net, struct sock *sk, struct sk_buff *skb) in ip6_rcv_finish() argument 74 skb = l3mdev_ip6_rcv(skb); in ip6_rcv_finish() 75 if (!skb) in ip6_rcv_finish() [all …]
|
H A D | ip6_offload.c | 33 #define indirect_call_gro_receive_l4(f2, f1, cb, head, skb) \ argument 35 unlikely(gro_recursion_inc_test(skb)) ? \ 36 NAPI_GRO_CB(skb)->flush |= 1, NULL : \ 37 INDIRECT_CALL_L4(cb, f2, f1, head, skb); \ 40 static int ipv6_gro_pull_exthdrs(struct sk_buff *skb, int off, int proto) in ipv6_gro_pull_exthdrs() argument 56 opth = skb_gro_header(skb, off + sizeof(*opth), off); in ipv6_gro_pull_exthdrs() 62 opth = skb_gro_header(skb, off + len, off); in ipv6_gro_pull_exthdrs() 70 skb_gro_pull(skb, off - skb_gro_receive_network_offset(skb)); in ipv6_gro_pull_exthdrs() 74 static int ipv6_gso_pull_exthdrs(struct sk_buff *skb, in argument 107 ipv6_gso_segment(struct sk_buff * skb,netdev_features_t features) ipv6_gso_segment() argument 221 ipv6_gro_receive(struct list_head * head,struct sk_buff * skb) ipv6_gro_receive() argument 309 sit_ip6ip6_gro_receive(struct list_head * head,struct sk_buff * skb) sit_ip6ip6_gro_receive() argument 324 ip4ip6_gro_receive(struct list_head * head,struct sk_buff * skb) ip4ip6_gro_receive() argument 338 ipv6_gro_complete(struct sk_buff * skb,int nhoff) ipv6_gro_complete() argument 390 sit_gro_complete(struct sk_buff * skb,int nhoff) sit_gro_complete() argument 397 ip6ip6_gro_complete(struct sk_buff * skb,int nhoff) ip6ip6_gro_complete() argument 404 ip4ip6_gro_complete(struct sk_buff * skb,int nhoff) ip4ip6_gro_complete() argument 412 sit_gso_segment(struct sk_buff * skb,netdev_features_t features) sit_gso_segment() argument 421 ip4ip6_gso_segment(struct sk_buff * skb,netdev_features_t features) ip4ip6_gso_segment() argument 430 ip6ip6_gso_segment(struct sk_buff * skb,netdev_features_t features) ip6ip6_gso_segment() argument [all...] |
H A D | udp_offload.c | 19 static struct sk_buff *udp6_ufo_fragment(struct sk_buff *skb, in udp6_ufo_fragment() argument 33 if (skb->encapsulation && skb_shinfo(skb)->gso_type & in udp6_ufo_fragment() 35 segs = skb_udp_tunnel_segment(skb, features, true); in udp6_ufo_fragment() 40 if (!(skb_shinfo(skb)->gso_type & (SKB_GSO_UDP | SKB_GSO_UDP_L4))) in udp6_ufo_fragment() 43 if (!pskb_may_pull(skb, sizeof(struct udphdr))) in udp6_ufo_fragment() 46 if (skb_shinfo(skb)->gso_type & SKB_GSO_UDP_L4) in udp6_ufo_fragment() 47 return __udp_gso_segment(skb, features, true); in udp6_ufo_fragment() 49 mss = skb_shinfo(skb)->gso_size; in udp6_ufo_fragment() 50 if (unlikely(skb->len <= mss)) in udp6_ufo_fragment() 57 uh = udp_hdr(skb); in udp6_ufo_fragment() [all …]
|
/linux/net/devlink/ |
H A D | netlink_gen.h | 21 int devlink_nl_pre_doit(const struct genl_split_ops *ops, struct sk_buff *skb, 24 struct sk_buff *skb, struct genl_info *info); 26 struct sk_buff *skb, struct genl_info *info); 28 struct sk_buff *skb, 31 devlink_nl_post_doit(const struct genl_split_ops *ops, struct sk_buff *skb, 35 struct sk_buff *skb, struct genl_info *info); 37 int devlink_nl_get_doit(struct sk_buff *skb, struct genl_info *info); 38 int devlink_nl_get_dumpit(struct sk_buff *skb, struct netlink_callback *cb); 39 int devlink_nl_port_get_doit(struct sk_buff *skb, struct genl_info *info); 40 int devlink_nl_port_get_dumpit(struct sk_buff *skb, [all …]
|
/linux/include/net/ |
H A D | gro.h | 103 #define NAPI_GRO_CB(skb) ((struct napi_gro_cb *)(skb)->cb) argument 106 static inline int gro_recursion_inc_test(struct sk_buff *skb) in gro_recursion_inc_test() argument 108 return ++NAPI_GRO_CB(skb)->recursion_counter == GRO_RECURSION_LIMIT; in gro_recursion_inc_test() 114 struct sk_buff *skb) in call_gro_receive() argument 116 if (unlikely(gro_recursion_inc_test(skb))) { in call_gro_receive() 117 NAPI_GRO_CB(skb)->flush |= 1; in call_gro_receive() 121 return cb(head, skb); in call_gro_receive() 129 struct sk_buff *skb) in call_gro_receive_sk() argument 131 if (unlikely(gro_recursion_inc_test(skb))) { in call_gro_receive_sk() 132 NAPI_GRO_CB(skb)->flush |= 1; in call_gro_receive_sk() [all …]
|
H A D | llc_c_ev.h | 123 static __inline__ struct llc_conn_state_ev *llc_conn_ev(struct sk_buff *skb) in llc_conn_ev() argument 125 return (struct llc_conn_state_ev *)skb->cb; in llc_conn_ev() 128 typedef int (*llc_conn_ev_t)(struct sock *sk, struct sk_buff *skb); 129 typedef int (*llc_conn_ev_qfyr_t)(struct sock *sk, struct sk_buff *skb); 131 int llc_conn_ev_conn_req(struct sock *sk, struct sk_buff *skb); 132 int llc_conn_ev_data_req(struct sock *sk, struct sk_buff *skb); 133 int llc_conn_ev_disc_req(struct sock *sk, struct sk_buff *skb); 134 int llc_conn_ev_rst_req(struct sock *sk, struct sk_buff *skb); 135 int llc_conn_ev_local_busy_detected(struct sock *sk, struct sk_buff *skb); 136 int llc_conn_ev_local_busy_cleared(struct sock *sk, struct sk_buff *skb); [all …]
|
H A D | llc_c_ac.h | 97 typedef int (*llc_conn_action_t)(struct sock *sk, struct sk_buff *skb); 99 int llc_conn_ac_clear_remote_busy(struct sock *sk, struct sk_buff *skb); 100 int llc_conn_ac_conn_ind(struct sock *sk, struct sk_buff *skb); 101 int llc_conn_ac_conn_confirm(struct sock *sk, struct sk_buff *skb); 102 int llc_conn_ac_data_ind(struct sock *sk, struct sk_buff *skb); 103 int llc_conn_ac_disc_ind(struct sock *sk, struct sk_buff *skb); 104 int llc_conn_ac_rst_ind(struct sock *sk, struct sk_buff *skb); 105 int llc_conn_ac_rst_confirm(struct sock *sk, struct sk_buff *skb); 107 struct sk_buff *skb); 109 struct sk_buff *skb); [all …]
|
/linux/net/ipv4/ |
H A D | udp_offload.c | 16 static struct sk_buff *__skb_udp_tunnel_segment(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument 18 struct sk_buff *(*gso_inner_segment)(struct sk_buff *skb, in __skb_udp_tunnel_segment() argument 22 int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb); in __skb_udp_tunnel_segment() 25 struct udphdr *uh = udp_hdr(skb); in __skb_udp_tunnel_segment() 26 u16 mac_offset = skb->mac_header; in __skb_udp_tunnel_segment() 27 __be16 protocol = skb->protocol; in __skb_udp_tunnel_segment() 28 u16 mac_len = skb->mac_len; in __skb_udp_tunnel_segment() 33 if (unlikely(!pskb_may_pull(skb, tnl_hlen))) in __skb_udp_tunnel_segment() 42 if (skb_shinfo(skb)->gso_type & SKB_GSO_PARTIAL) in __skb_udp_tunnel_segment() 45 partial = (__force __wsum)htonl(skb->len); in __skb_udp_tunnel_segment() [all …]
|
H A D | tcp_offload.c | 16 static void tcp_gso_tstamp(struct sk_buff *skb, struct sk_buff *gso_skb, in tcp_gso_tstamp() argument 22 while (skb) { in tcp_gso_tstamp() 24 skb_shinfo(skb)->tx_flags |= flags; in tcp_gso_tstamp() 25 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 29 skb = skb->next; in tcp_gso_tstamp() 88 static struct sk_buff *__tcp4_gso_segment_list(struct sk_buff *skb, in __tcp4_gso_segment_list() argument 91 skb = skb_segment_list(skb, features, skb_mac_header_len(skb)); in __tcp4_gso_segment_list() 92 if (IS_ERR(skb)) in __tcp4_gso_segment_list() 93 return skb; in __tcp4_gso_segment_list() 95 return __tcpv4_gso_segment_list_csum(skb); in __tcp4_gso_segment_list() [all …]
|
H A D | gre_offload.c | 16 static struct sk_buff *gre_gso_segment(struct sk_buff *skb, in gre_gso_segment() argument 19 int tnl_hlen = skb_inner_mac_header(skb) - skb_transport_header(skb); in gre_gso_segment() 22 u16 mac_offset = skb->mac_header; in gre_gso_segment() 23 __be16 protocol = skb->protocol; in gre_gso_segment() 24 u16 mac_len = skb->mac_len; in gre_gso_segment() 27 if (!skb->encapsulation) in gre_gso_segment() 33 if (unlikely(!pskb_may_pull(skb, tnl_hlen))) in gre_gso_segment() 37 skb->encapsulation = 0; in gre_gso_segment() 38 SKB_GSO_CB(skb)->encap_level = 0; in gre_gso_segment() 39 __skb_pull(skb, tnl_hlen); in gre_gso_segment() [all …]
|
/linux/drivers/net/ethernet/qualcomm/rmnet/ |
H A D | rmnet_handlers.c | 22 static void rmnet_set_skb_proto(struct sk_buff *skb) in rmnet_set_skb_proto() argument 24 switch (skb->data[0] & 0xF0) { in rmnet_set_skb_proto() 26 skb->protocol = htons(ETH_P_IP); in rmnet_set_skb_proto() 29 skb->protocol = htons(ETH_P_IPV6); in rmnet_set_skb_proto() 32 skb->protocol = htons(ETH_P_MAP); in rmnet_set_skb_proto() 40 rmnet_deliver_skb(struct sk_buff *skb) in rmnet_deliver_skb() argument 42 struct rmnet_priv *priv = netdev_priv(skb->dev); in rmnet_deliver_skb() 44 skb_reset_transport_header(skb); in rmnet_deliver_skb() 45 skb_reset_network_header(skb); in rmnet_deliver_skb() 46 rmnet_vnd_rx_fixup(skb, skb->dev); in rmnet_deliver_skb() [all …]
|
/linux/drivers/net/wireless/ath/ath10k/ |
H A D | wmi-ops.h | 15 void (*rx)(struct ath10k *ar, struct sk_buff *skb); 19 int (*pull_scan)(struct ath10k *ar, struct sk_buff *skb, 21 int (*pull_mgmt_rx)(struct ath10k *ar, struct sk_buff *skb, 23 int (*pull_mgmt_tx_compl)(struct ath10k *ar, struct sk_buff *skb, 26 struct ath10k *ar, struct sk_buff *skb, 28 int (*pull_ch_info)(struct ath10k *ar, struct sk_buff *skb, 30 int (*pull_vdev_start)(struct ath10k *ar, struct sk_buff *skb, 32 int (*pull_peer_kick)(struct ath10k *ar, struct sk_buff *skb, 34 int (*pull_swba)(struct ath10k *ar, struct sk_buff *skb, 36 int (*pull_phyerr_hdr)(struct ath10k *ar, struct sk_buff *skb, [all …]
|
/linux/net/sched/ |
H A D | sch_frag.c | 18 int (*xmit)(struct sk_buff *skb); 23 static int sch_frag_xmit(struct net *net, struct sock *sk, struct sk_buff *skb) in sch_frag_xmit() argument 27 if (skb_cow_head(skb, data->l2_len) < 0) { in sch_frag_xmit() 28 kfree_skb(skb); in sch_frag_xmit() 32 __skb_dst_copy(skb, data->dst); in sch_frag_xmit() 33 *qdisc_skb_cb(skb) = data->cb; in sch_frag_xmit() 34 skb->inner_protocol = data->inner_protocol; in sch_frag_xmit() 36 __vlan_hwaccel_put_tag(skb, data->vlan_proto, in sch_frag_xmit() 39 __vlan_hwaccel_clear_tag(skb); in sch_frag_xmit() 42 skb_push(skb, data->l2_len); in sch_frag_xmit() [all …]
|
/linux/net/bridge/netfilter/ |
H A D | nf_conntrack_bridge.c | 28 struct sk_buff *skb, in nf_br_ip_fragment() argument 34 int frag_max_size = BR_INPUT_SKB_CB(skb)->frag_max_size; in nf_br_ip_fragment() 35 u8 tstamp_type = skb->tstamp_type; in nf_br_ip_fragment() 37 ktime_t tstamp = skb->tstamp; in nf_br_ip_fragment() 43 if (skb->ip_summed == CHECKSUM_PARTIAL && in nf_br_ip_fragment() 44 (err = skb_checksum_help(skb))) in nf_br_ip_fragment() 47 iph = ip_hdr(skb); in nf_br_ip_fragment() 55 ll_rs = LL_RESERVED_SPACE(skb->dev); in nf_br_ip_fragment() 56 mtu = skb->dev->mtu; in nf_br_ip_fragment() 58 if (skb_has_frag_list(skb)) { in nf_br_ip_fragment() [all …]
|
/linux/net/bridge/ |
H A D | br_netfilter_ipv6.c | 43 int br_validate_ipv6(struct net *net, struct sk_buff *skb) in br_validate_ipv6() argument 46 struct inet6_dev *idev = __in6_dev_get(skb->dev); in br_validate_ipv6() 50 if (!pskb_may_pull(skb, ip6h_len)) in br_validate_ipv6() 53 if (skb->len < ip6h_len) in br_validate_ipv6() 56 hdr = ipv6_hdr(skb); in br_validate_ipv6() 62 if (hdr->nexthdr == NEXTHDR_HOP && nf_ip6_check_hbh_len(skb, &pkt_len)) in br_validate_ipv6() 65 if (pkt_len + ip6h_len > skb->len) { in br_validate_ipv6() 70 if (pskb_trim_rcsum(skb, pkt_len + ip6h_len)) { in br_validate_ipv6() 76 memset(IP6CB(skb), 0, sizeof(struct inet6_skb_parm)); in br_validate_ipv6() 89 br_nf_ipv6_daddr_was_changed(const struct sk_buff *skb, in br_nf_ipv6_daddr_was_changed() argument [all …]
|
/linux/net/ieee802154/6lowpan/ |
H A D | rx.c | 21 static int lowpan_give_skb_to_device(struct sk_buff *skb) in lowpan_give_skb_to_device() argument 23 skb->protocol = htons(ETH_P_IPV6); in lowpan_give_skb_to_device() 24 skb->dev->stats.rx_packets++; in lowpan_give_skb_to_device() 25 skb->dev->stats.rx_bytes += skb->len; in lowpan_give_skb_to_device() 27 return netif_rx(skb); in lowpan_give_skb_to_device() 30 static int lowpan_rx_handlers_result(struct sk_buff *skb, lowpan_rx_result res) in lowpan_rx_handlers_result() argument 40 kfree_skb(skb); in lowpan_rx_handlers_result() 46 return lowpan_give_skb_to_device(skb); in lowpan_rx_handlers_result() 64 static lowpan_rx_result lowpan_rx_h_frag(struct sk_buff *skb) in lowpan_rx_h_frag() argument 68 if (!(lowpan_is_frag1(*skb_network_header(skb)) || in lowpan_rx_h_frag() [all …]
|
/linux/drivers/net/wireguard/ |
H A D | receive.c | 26 #define SKB_TYPE_LE32(skb) (((struct message_header *)(skb)->data)->type) argument 28 static size_t validate_header_len(struct sk_buff *skb) in validate_header_len() argument 30 if (unlikely(skb->len < sizeof(struct message_header))) in validate_header_len() 32 if (SKB_TYPE_LE32(skb) == cpu_to_le32(MESSAGE_DATA) && in validate_header_len() 33 skb->len >= MESSAGE_MINIMUM_LENGTH) in validate_header_len() 35 if (SKB_TYPE_LE32(skb) == cpu_to_le32(MESSAGE_HANDSHAKE_INITIATION) && in validate_header_len() 36 skb->len == sizeof(struct message_handshake_initiation)) in validate_header_len() 38 if (SKB_TYPE_LE32(skb) == cpu_to_le32(MESSAGE_HANDSHAKE_RESPONSE) && in validate_header_len() 39 skb->len == sizeof(struct message_handshake_response)) in validate_header_len() 41 if (SKB_TYPE_LE32(skb) == cpu_to_le32(MESSAGE_HANDSHAKE_COOKIE) && in validate_header_len() [all …]
|
/linux/net/lapb/ |
H A D | lapb_subr.c | 48 struct sk_buff *skb; in lapb_frames_acked() local 58 skb = skb_dequeue(&lapb->ack_queue); in lapb_frames_acked() 59 kfree_skb(skb); in lapb_frames_acked() 66 struct sk_buff *skb, *skb_prev = NULL; in lapb_requeue_frames() local 73 while ((skb = skb_dequeue(&lapb->ack_queue)) != NULL) { in lapb_requeue_frames() 75 skb_queue_head(&lapb->write_queue, skb); in lapb_requeue_frames() 77 skb_append(skb_prev, skb, &lapb->write_queue); in lapb_requeue_frames() 78 skb_prev = skb; in lapb_requeue_frames() 106 int lapb_decode(struct lapb_cb *lapb, struct sk_buff *skb, in lapb_decode() argument 111 lapb_dbg(2, "(%p) S%d RX %3ph\n", lapb->dev, lapb->state, skb->data); in lapb_decode() [all …]
|
/linux/net/ax25/ |
H A D | ax25_in.c | 34 static int ax25_rx_fragment(ax25_cb *ax25, struct sk_buff *skb) in ax25_rx_fragment() argument 39 if (!(*skb->data & AX25_SEG_FIRST)) { in ax25_rx_fragment() 40 if ((ax25->fragno - 1) == (*skb->data & AX25_SEG_REM)) { in ax25_rx_fragment() 42 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment() 43 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment() 44 ax25->fraglen += skb->len; in ax25_rx_fragment() 45 skb_queue_tail(&ax25->frag_queue, skb); in ax25_rx_fragment() 82 if (*skb->data & AX25_SEG_FIRST) { in ax25_rx_fragment() 84 ax25->fragno = *skb->data & AX25_SEG_REM; in ax25_rx_fragment() 85 skb_pull(skb, 1); /* skip fragno */ in ax25_rx_fragment() [all …]
|