| /linux/net/xfrm/ |
| H A D | xfrm_output.c | 28 static int xfrm_output2(struct net *net, struct sock *sk, struct sk_buff *skb); 29 static int xfrm_inner_extract_output(struct xfrm_state *x, struct sk_buff *skb); 31 static int xfrm_skb_check_space(struct sk_buff *skb) in xfrm_skb_check_space() argument 33 struct dst_entry *dst = skb_dst(skb); in xfrm_skb_check_space() 35 - skb_headroom(skb); in xfrm_skb_check_space() 36 int ntail = dst->dev->needed_tailroom - skb_tailroom(skb); in xfrm_skb_check_space() 45 return pskb_expand_head(skb, nhead, ntail, GFP_ATOMIC); in xfrm_skb_check_space() 52 static struct dst_entry *skb_dst_pop(struct sk_buff *skb) in skb_dst_pop() argument 54 struct dst_entry *child = dst_clone(xfrm_dst_child(skb_dst(skb))); in skb_dst_pop() 56 skb_dst_drop(skb); in skb_dst_pop() [all …]
|
| H A D | xfrm_input.c | 41 int (*finish)(struct net *net, struct sock *sk, struct sk_buff *skb); 103 static int xfrm_rcv_cb(struct sk_buff *skb, unsigned int family, u8 protocol, in xfrm_rcv_cb() argument 114 ret = afinfo->callback(skb, protocol, err); in xfrm_rcv_cb() 120 struct sec_path *secpath_set(struct sk_buff *skb) in secpath_set() argument 122 struct sec_path *sp, *tmp = skb_ext_find(skb, SKB_EXT_SEC_PATH); in secpath_set() 124 sp = skb_ext_add(skb, SKB_EXT_SEC_PATH); in secpath_set() 143 int xfrm_parse_spi(struct sk_buff *skb, u8 nexthdr, __be32 *spi, __be32 *seq) in xfrm_parse_spi() argument 160 if (!pskb_may_pull(skb, sizeof(struct ip_comp_hdr))) in xfrm_parse_spi() 162 *spi = htonl(ntohs(*(__be16 *)(skb_transport_header(skb) + 2))); in xfrm_parse_spi() 169 if (!pskb_may_pull(skb, hlen)) in xfrm_parse_spi() [all …]
|
| /linux/net/core/ |
| H A D | gro.c | 92 int skb_gro_receive(struct sk_buff *p, struct sk_buff *skb) in skb_gro_receive() argument 94 struct skb_shared_info *pinfo, *skbinfo = skb_shinfo(skb); in skb_gro_receive() 95 unsigned int offset = skb_gro_offset(skb); in skb_gro_receive() 96 unsigned int headlen = skb_headlen(skb); in skb_gro_receive() 97 unsigned int len = skb_gro_len(skb); in skb_gro_receive() 109 if (p->pp_recycle != skb->pp_recycle) in skb_gro_receive() 113 NAPI_GRO_CB(skb)->flush)) in skb_gro_receive() 117 if (NAPI_GRO_CB(skb)->proto != IPPROTO_TCP || in skb_gro_receive() 124 segs = NAPI_GRO_CB(skb)->count; in skb_gro_receive() 151 new_truesize = SKB_TRUESIZE(skb_end_offset(skb)); in skb_gro_receive() [all …]
|
| H A D | skbuff.c | 19 * Ray VanTassle : Fixed --skb->lock in free 86 #include <trace/events/skb.h> 195 * @skb: buffer 205 static void skb_panic(struct sk_buff *skb, unsigned int sz, void *addr, in skb_panic() argument 209 msg, addr, skb->len, sz, skb->head, skb->data, in skb_panic() 210 (unsigned long)skb->tail, (unsigned long)skb->end, in skb_panic() 211 skb->dev ? skb->dev->name : "<NULL>"); in skb_panic() 215 static void skb_over_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_over_panic() argument 217 skb_panic(skb, sz, addr, __func__); in skb_over_panic() 220 static void skb_under_panic(struct sk_buff *skb, unsigned int sz, void *addr) in skb_under_panic() argument [all …]
|
| H A D | gso.c | 9 * @skb: buffer to segment 13 struct sk_buff *skb_eth_gso_segment(struct sk_buff *skb, in skb_eth_gso_segment() argument 22 segs = ptype->callbacks.gso_segment(skb, features); in skb_eth_gso_segment() 34 * @skb: buffer to segment 37 struct sk_buff *skb_mac_gso_segment(struct sk_buff *skb, in skb_mac_gso_segment() argument 42 int vlan_depth = skb->mac_len; in skb_mac_gso_segment() 43 __be16 type = skb_network_protocol(skb, &vlan_depth); in skb_mac_gso_segment() 48 __skb_pull(skb, vlan_depth); in skb_mac_gso_segment() 53 segs = ptype->callbacks.gso_segment(skb, features); in skb_mac_gso_segment() 59 __skb_push(skb, skb->data - skb_mac_header(skb)); in skb_mac_gso_segment() [all …]
|
| /linux/include/linux/ |
| H A D | skbuff.h | 44 * DOC: skb checksums 105 * not in skb->csum. Thus, skb->csum is undefined in this case. 133 * checksum is bad, skb->csum_level would be set to zero (TCP checksum is 145 * skb->csum, it MUST use CHECKSUM_COMPLETE, not CHECKSUM_UNNECESSARY. 155 * referred to by skb->csum_start + skb->csum_offset and any preceding 194 * The skb was already checksummed by the protocol, or a checksum is not 253 /* Maximum value in skb->csum_level */ 260 /* For X bytes available in skb->head, what is the minimal 272 /* return minimum truesize of one skb containing X bytes of data */ 310 * skb is out in neigh layer. [all …]
|
| /linux/drivers/net/wireless/ath/ath10k/ |
| H A D | wmi-ops.h | 15 void (*rx)(struct ath10k *ar, struct sk_buff *skb); 19 int (*pull_scan)(struct ath10k *ar, struct sk_buff *skb, 21 int (*pull_mgmt_rx)(struct ath10k *ar, struct sk_buff *skb, 23 int (*pull_mgmt_tx_compl)(struct ath10k *ar, struct sk_buff *skb, 26 struct ath10k *ar, struct sk_buff *skb, 28 int (*pull_ch_info)(struct ath10k *ar, struct sk_buff *skb, 30 int (*pull_vdev_start)(struct ath10k *ar, struct sk_buff *skb, 32 int (*pull_peer_kick)(struct ath10k *ar, struct sk_buff *skb, 34 int (*pull_swba)(struct ath10k *ar, struct sk_buff *skb, 36 int (*pull_phyerr_hdr)(struct ath10k *ar, struct sk_buff *skb, [all …]
|
| /linux/net/bridge/ |
| H A D | br_netfilter_hooks.c | 72 #define IS_IP(skb) \ argument 73 (!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IP)) 75 #define IS_IPV6(skb) \ argument 76 (!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_IPV6)) 78 #define IS_ARP(skb) \ argument 79 (!skb_vlan_tag_present(skb) && skb->protocol == htons(ETH_P_ARP)) 81 static inline __be16 vlan_proto(const struct sk_buff *skb) in vlan_proto() argument 83 if (skb_vlan_tag_present(skb)) in vlan_proto() 84 return skb->protocol; in vlan_proto() 85 else if (skb->protocol == htons(ETH_P_8021Q)) in vlan_proto() [all …]
|
| H A D | br_forward.c | 22 const struct sk_buff *skb) in should_deliver() argument 27 return ((p->flags & BR_HAIRPIN_MODE) || skb->dev != p->dev) && in should_deliver() 29 br_allowed_egress(vg, skb) && nbp_switchdev_allowed_egress(p, skb) && in should_deliver() 30 !br_skb_isolated(p, skb); in should_deliver() 33 int br_dev_queue_push_xmit(struct net *net, struct sock *sk, struct sk_buff *skb) in br_dev_queue_push_xmit() argument 35 skb_push(skb, ETH_HLEN); in br_dev_queue_push_xmit() 36 if (!is_skb_forwardable(skb->dev, skb)) in br_dev_queue_push_xmit() 39 br_drop_fake_rtable(skb); in br_dev_queue_push_xmit() 41 if (skb->ip_summed == CHECKSUM_PARTIAL && in br_dev_queue_push_xmit() 42 eth_type_vlan(skb->protocol)) { in br_dev_queue_push_xmit() [all …]
|
| /linux/net/devlink/ |
| H A D | netlink_gen.h | 22 int devlink_nl_pre_doit(const struct genl_split_ops *ops, struct sk_buff *skb, 25 struct sk_buff *skb, struct genl_info *info); 27 struct sk_buff *skb, struct genl_info *info); 29 struct sk_buff *skb, 32 devlink_nl_post_doit(const struct genl_split_ops *ops, struct sk_buff *skb, 36 struct sk_buff *skb, struct genl_info *info); 38 int devlink_nl_get_doit(struct sk_buff *skb, struct genl_info *info); 39 int devlink_nl_get_dumpit(struct sk_buff *skb, struct netlink_callback *cb); 40 int devlink_nl_port_get_doit(struct sk_buff *skb, struct genl_info *info); 41 int devlink_nl_port_get_dumpit(struct sk_buff *skb, [all …]
|
| /linux/include/net/ |
| H A D | gro.h | 20 /* Virtual address of skb_shinfo(skb)->frags[0].page + offset. */ 36 /* This indicates where we are processing relative to skb->data. */ 39 /* This is non-zero if the packet cannot be merged with the new skb. */ 71 /* Free the skb? */ 100 #define NAPI_GRO_CB(skb) ((struct napi_gro_cb *)(skb)->cb) 103 static inline int gro_recursion_inc_test(struct sk_buff *skb) argument 105 return ++NAPI_GRO_CB(skb)->recursion_counter == GRO_RECURSION_LIMIT; 111 struct sk_buff *skb) 113 if (unlikely(gro_recursion_inc_test(skb))) { in call_gro_receive() 106 gro_recursion_inc_test(struct sk_buff * skb) gro_recursion_inc_test() argument 114 call_gro_receive(gro_receive_t cb,struct list_head * head,struct sk_buff * skb) call_gro_receive() argument 129 call_gro_receive_sk(gro_receive_sk_t cb,struct sock * sk,struct list_head * head,struct sk_buff * skb) call_gro_receive_sk() argument 139 skb_gro_offset(const struct sk_buff * skb) skb_gro_offset() argument 144 skb_gro_len(const struct sk_buff * skb) skb_gro_len() argument 149 skb_gro_pull(struct sk_buff * skb,unsigned int len) skb_gro_pull() argument 154 skb_gro_header_fast(const struct sk_buff * skb,unsigned int offset) skb_gro_header_fast() argument 160 skb_gro_may_pull(const struct sk_buff * skb,unsigned int hlen) skb_gro_may_pull() argument 166 skb_gro_header_slow(struct sk_buff * skb,unsigned int hlen,unsigned int offset) skb_gro_header_slow() argument 175 skb_gro_header(struct sk_buff * skb,unsigned int hlen,unsigned int offset) skb_gro_header() argument 186 skb_gro_receive_network_offset(const struct sk_buff * skb) skb_gro_receive_network_offset() argument 191 skb_gro_network_header(const struct sk_buff * skb) skb_gro_network_header() argument 199 inet_gro_compute_pseudo(const struct sk_buff * skb,int proto) inet_gro_compute_pseudo() argument 208 skb_gro_postpull_rcsum(struct sk_buff * skb,const void * start,unsigned int len) skb_gro_postpull_rcsum() argument 223 skb_at_gro_remcsum_start(struct sk_buff * skb) skb_at_gro_remcsum_start() argument 228 __skb_gro_checksum_validate_needed(struct sk_buff * skb,bool zero_okay,__sum16 check) __skb_gro_checksum_validate_needed() argument 240 __skb_gro_checksum_validate_complete(struct sk_buff * skb,__wsum psum) __skb_gro_checksum_validate_complete() argument 252 skb_gro_incr_csum_unnecessary(struct sk_buff * skb) skb_gro_incr_csum_unnecessary() argument 266 __skb_gro_checksum_validate(skb,proto,zero_okay,check,compute_pseudo) global() argument 278 skb_gro_checksum_validate(skb,proto,compute_pseudo) global() argument 281 skb_gro_checksum_validate_zero_check(skb,proto,check,compute_pseudo) global() argument 285 skb_gro_checksum_simple_validate(skb) global() argument 288 __skb_gro_checksum_convert_check(struct sk_buff * skb) __skb_gro_checksum_convert_check() argument 294 __skb_gro_checksum_convert(struct sk_buff * skb,__wsum pseudo) __skb_gro_checksum_convert() argument 301 skb_gro_checksum_try_convert(skb,proto,compute_pseudo) global() argument 319 skb_gro_remcsum_process(struct sk_buff * skb,void * ptr,unsigned int off,size_t hdrlen,int start,int offset,struct gro_remcsum * grc,bool nopartial) skb_gro_remcsum_process() argument 351 skb_gro_remcsum_cleanup(struct sk_buff * skb,struct gro_remcsum * grc) skb_gro_remcsum_cleanup() argument 368 skb_gro_flush_final(struct sk_buff * skb,struct sk_buff * pp,int flush) skb_gro_flush_final() argument 373 skb_gro_flush_final_remcsum(struct sk_buff * skb,struct sk_buff * pp,int flush,struct gro_remcsum * grc) skb_gro_flush_final_remcsum() argument 385 skb_gro_flush_final(struct sk_buff * skb,struct sk_buff * pp,int flush) skb_gro_flush_final() argument 389 skb_gro_flush_final_remcsum(struct sk_buff * skb,struct sk_buff * pp,int flush,struct gro_remcsum * grc) skb_gro_flush_final_remcsum() argument 415 indirect_call_gro_receive_inet(cb,f2,f1,head,skb) global() argument 426 udp_gro_udphdr(struct sk_buff * skb) udp_gro_udphdr() argument 438 ip6_gro_compute_pseudo(const struct sk_buff * skb,int proto) ip6_gro_compute_pseudo() argument 546 gro_normal_one(struct gro_node * gro,struct sk_buff * skb,int segs) gro_normal_one() argument 564 inet_get_iif_sdif(const struct sk_buff * skb,int * iif,int * sdif) inet_get_iif_sdif() argument 585 inet6_get_iif_sdif(const struct sk_buff * skb,int * iif,int * sdif) inet6_get_iif_sdif() argument [all...] |
| H A D | llc_c_ac.h | 97 typedef int (*llc_conn_action_t)(struct sock *sk, struct sk_buff *skb); 99 int llc_conn_ac_clear_remote_busy(struct sock *sk, struct sk_buff *skb); 100 int llc_conn_ac_conn_ind(struct sock *sk, struct sk_buff *skb); 101 int llc_conn_ac_conn_confirm(struct sock *sk, struct sk_buff *skb); 102 int llc_conn_ac_data_ind(struct sock *sk, struct sk_buff *skb); 103 int llc_conn_ac_disc_ind(struct sock *sk, struct sk_buff *skb); 104 int llc_conn_ac_rst_ind(struct sock *sk, struct sk_buff *skb); 105 int llc_conn_ac_rst_confirm(struct sock *sk, struct sk_buff *skb); 107 struct sk_buff *skb); 109 struct sk_buff *skb); [all …]
|
| H A D | llc_c_ev.h | 123 static __inline__ struct llc_conn_state_ev *llc_conn_ev(struct sk_buff *skb) in llc_conn_ev() argument 125 return (struct llc_conn_state_ev *)skb->cb; in llc_conn_ev() 128 typedef int (*llc_conn_ev_t)(struct sock *sk, struct sk_buff *skb); 129 typedef int (*llc_conn_ev_qfyr_t)(struct sock *sk, struct sk_buff *skb); 131 int llc_conn_ev_conn_req(struct sock *sk, struct sk_buff *skb); 132 int llc_conn_ev_data_req(struct sock *sk, struct sk_buff *skb); 133 int llc_conn_ev_disc_req(struct sock *sk, struct sk_buff *skb); 134 int llc_conn_ev_rst_req(struct sock *sk, struct sk_buff *skb); 135 int llc_conn_ev_local_busy_detected(struct sock *sk, struct sk_buff *skb); 136 int llc_conn_ev_local_busy_cleared(struct sock *sk, struct sk_buff *skb); [all …]
|
| /linux/net/ipv4/ |
| H A D | tcp_offload.c | 16 static void tcp_gso_tstamp(struct sk_buff *skb, struct sk_buff *gso_skb, in tcp_gso_tstamp() argument 22 while (skb) { in tcp_gso_tstamp() 24 skb_shinfo(skb)->tx_flags |= flags; in tcp_gso_tstamp() 25 skb_shinfo(skb)->tskey = ts_seq; in tcp_gso_tstamp() 29 skb = skb->next; in tcp_gso_tstamp() 88 static struct sk_buff *__tcp4_gso_segment_list(struct sk_buff *skb, in __tcp4_gso_segment_list() argument 91 skb = skb_segment_list(skb, features, skb_mac_header_len(skb)); in __tcp4_gso_segment_list() 98 tcp4_gso_segment(struct sk_buff * skb,netdev_features_t features) tcp4_gso_segment() argument 132 tcp_gso_segment(struct sk_buff * skb,netdev_features_t features) tcp_gso_segment() argument 285 tcp_gro_pull_header(struct sk_buff * skb) tcp_gro_pull_header() argument 312 tcp_gro_receive(struct list_head * head,struct sk_buff * skb,struct tcphdr * th) tcp_gro_receive() argument 396 tcp_gro_complete(struct sk_buff * skb) tcp_gro_complete() argument 416 tcp4_check_fraglist_gro(struct list_head * head,struct sk_buff * skb,struct tcphdr * th) tcp4_check_fraglist_gro() argument 447 tcp4_gro_receive(struct list_head * head,struct sk_buff * skb) tcp4_gro_receive() argument 470 tcp4_gro_complete(struct sk_buff * skb,int thoff) tcp4_gro_complete() argument [all...] |
| H A D | esp4_offload.c | 31 struct sk_buff *skb) in esp4_gro_receive() argument 33 int offset = skb_gro_offset(skb); in esp4_gro_receive() 40 if (!pskb_pull(skb, offset)) in esp4_gro_receive() 43 if (xfrm_parse_spi(skb, IPPROTO_ESP, &spi, &seq) != 0) in esp4_gro_receive() 46 xo = xfrm_offload(skb); in esp4_gro_receive() 48 struct sec_path *sp = secpath_set(skb); in esp4_gro_receive() 56 x = xfrm_input_state_lookup(dev_net(skb->dev), skb->mark, in esp4_gro_receive() 57 (xfrm_address_t *)&ip_hdr(skb)->daddr, in esp4_gro_receive() 69 skb->mark = xfrm_smark_get(skb->mark, x); in esp4_gro_receive() 74 xo = xfrm_offload(skb); in esp4_gro_receive() [all …]
|
| /linux/drivers/net/can/dev/ |
| H A D | skb.c | 41 /* Put the skb on the stack to be looped backed locally lateron 47 int can_put_echo_skb(struct sk_buff *skb, struct net_device *dev, in can_put_echo_skb() argument 60 (skb->protocol != htons(ETH_P_CAN) && in can_put_echo_skb() 61 skb->protocol != htons(ETH_P_CANFD) && in can_put_echo_skb() 62 skb->protocol != htons(ETH_P_CANXL))) { in can_put_echo_skb() 63 kfree_skb(skb); in can_put_echo_skb() 68 skb = can_create_echo_skb(skb); in can_put_echo_skb() 69 if (!skb) in can_put_echo_skb() 73 skb->ip_summed = CHECKSUM_UNNECESSARY; in can_put_echo_skb() 74 skb->dev = dev; in can_put_echo_skb() [all …]
|
| /linux/net/openvswitch/ |
| H A D | actions.c | 90 static struct deferred_action *add_deferred_actions(struct sk_buff *skb, in add_deferred_actions() argument 100 da->skb = skb; in add_deferred_actions() 119 static int clone_execute(struct datapath *dp, struct sk_buff *skb, 125 static int do_execute_actions(struct datapath *dp, struct sk_buff *skb, 129 static int push_mpls(struct sk_buff *skb, struct sw_flow_key *key, in push_mpls() argument 134 err = skb_mpls_push(skb, mpls_lse, mpls_ethertype, mac_len, !!mac_len); in push_mpls() 145 static int pop_mpls(struct sk_buff *skb, struct sw_flow_key *key, in pop_mpls() argument 150 err = skb_mpls_pop(skb, ethertype, skb->mac_len, in pop_mpls() 162 static int set_mpls(struct sk_buff *skb, struct sw_flow_key *flow_key, in set_mpls() argument 169 if (!pskb_may_pull(skb, skb_network_offset(skb) + MPLS_HLEN)) in set_mpls() [all …]
|
| /linux/net/ipv6/ |
| H A D | ip6_input.c | 49 struct sk_buff *skb) in ip6_rcv_finish_core() argument 52 !skb_dst(skb) && !skb->sk) { in ip6_rcv_finish_core() 53 switch (ipv6_hdr(skb)->nexthdr) { in ip6_rcv_finish_core() 56 tcp_v6_early_demux(skb); in ip6_rcv_finish_core() 60 udp_v6_early_demux(skb); in ip6_rcv_finish_core() 65 if (!skb_valid_dst(skb)) in ip6_rcv_finish_core() 66 ip6_route_input(skb); in ip6_rcv_finish_core() 69 int ip6_rcv_finish(struct net *net, struct sock *sk, struct sk_buff *skb) in ip6_rcv_finish() argument 72 * skb to its handler for processing in ip6_rcv_finish() 74 skb = l3mdev_ip6_rcv(skb); in ip6_rcv_finish() [all …]
|
| H A D | ip6_offload.c | 33 #define indirect_call_gro_receive_l4(f2, f1, cb, head, skb) \ argument 35 unlikely(gro_recursion_inc_test(skb)) ? \ 36 NAPI_GRO_CB(skb)->flush |= 1, NULL : \ 37 INDIRECT_CALL_L4(cb, f2, f1, head, skb); \ 40 static int ipv6_gro_pull_exthdrs(struct sk_buff *skb, int off, int proto) in ipv6_gro_pull_exthdrs() argument 56 opth = skb_gro_header(skb, off + sizeof(*opth), off); in ipv6_gro_pull_exthdrs() 62 opth = skb_gro_header(skb, off + len, off); in ipv6_gro_pull_exthdrs() 70 skb_gro_pull(skb, off - skb_gro_receive_network_offset(skb)); in ipv6_gro_pull_exthdrs() 74 static int ipv6_gso_pull_exthdrs(struct sk_buff *skb, int proto) in ipv6_gso_pull_exthdrs() argument 90 if (unlikely(!pskb_may_pull(skb, 8))) in ipv6_gso_pull_exthdrs() [all …]
|
| H A D | exthdrs.c | 65 static bool ip6_tlvopt_unknown(struct sk_buff *skb, int optoff, in ip6_tlvopt_unknown() argument 79 switch ((skb_network_header(skb)[optoff] & 0xC0) >> 6) { in ip6_tlvopt_unknown() 90 if (ipv6_addr_is_multicast(&ipv6_hdr(skb)->daddr)) in ip6_tlvopt_unknown() 94 icmpv6_param_prob_reason(skb, ICMPV6_UNK_OPTION, optoff, in ip6_tlvopt_unknown() 100 kfree_skb_reason(skb, SKB_DROP_REASON_UNHANDLED_PROTO); in ip6_tlvopt_unknown() 104 static bool ipv6_hop_ra(struct sk_buff *skb, int optoff); 105 static bool ipv6_hop_ioam(struct sk_buff *skb, int optoff); 106 static bool ipv6_hop_jumbo(struct sk_buff *skb, int optoff); 107 static bool ipv6_hop_calipso(struct sk_buff *skb, int optoff); 109 static bool ipv6_dest_hao(struct sk_buff *skb, int optoff); [all …]
|
| H A D | esp6_offload.c | 54 struct sk_buff *skb) in esp6_gro_receive() argument 56 int offset = skb_gro_offset(skb); in esp6_gro_receive() 64 if (NAPI_GRO_CB(skb)->proto == IPPROTO_UDP) in esp6_gro_receive() 67 if (!pskb_pull(skb, offset)) in esp6_gro_receive() 70 if (xfrm_parse_spi(skb, IPPROTO_ESP, &spi, &seq) != 0) in esp6_gro_receive() 73 xo = xfrm_offload(skb); in esp6_gro_receive() 75 struct sec_path *sp = secpath_set(skb); in esp6_gro_receive() 83 x = xfrm_input_state_lookup(dev_net(skb->dev), skb->mark, in esp6_gro_receive() 84 (xfrm_address_t *)&ipv6_hdr(skb)->daddr, in esp6_gro_receive() 96 skb->mark = xfrm_smark_get(skb->mark, x); in esp6_gro_receive() [all …]
|
| /linux/net/bridge/netfilter/ |
| H A D | nf_conntrack_bridge.c | 28 struct sk_buff *skb, in nf_br_ip_fragment() argument 34 int frag_max_size = BR_INPUT_SKB_CB(skb)->frag_max_size; in nf_br_ip_fragment() 35 u8 tstamp_type = skb->tstamp_type; in nf_br_ip_fragment() 37 ktime_t tstamp = skb->tstamp; in nf_br_ip_fragment() 43 if (skb->ip_summed == CHECKSUM_PARTIAL && in nf_br_ip_fragment() 44 (err = skb_checksum_help(skb))) in nf_br_ip_fragment() 47 iph = ip_hdr(skb); in nf_br_ip_fragment() 55 ll_rs = LL_RESERVED_SPACE(skb->dev); in nf_br_ip_fragment() 56 mtu = skb->dev->mtu; in nf_br_ip_fragment() 58 if (skb_has_frag_list(skb)) { in nf_br_ip_fragment() 130 br_skb_cb_save(struct sk_buff * skb,struct br_input_skb_cb * cb,size_t inet_skb_parm_size) br_skb_cb_save() argument 137 br_skb_cb_restore(struct sk_buff * skb,const struct br_input_skb_cb * cb,u16 fragsz) br_skb_cb_restore() argument 145 nf_ct_br_defrag4(struct sk_buff * skb,const struct nf_hook_state * state) nf_ct_br_defrag4() argument 175 nf_ct_br_defrag6(struct sk_buff * skb,const struct nf_hook_state * state) nf_ct_br_defrag6() argument 204 nf_ct_br_ip_check(const struct sk_buff * skb) nf_ct_br_ip_check() argument 223 nf_ct_br_ipv6_check(const struct sk_buff * skb) nf_ct_br_ipv6_check() argument 240 nf_ct_bridge_pre(void * priv,struct sk_buff * skb,const struct nf_hook_state * state) nf_ct_bridge_pre() argument 294 nf_ct_bridge_in(void * priv,struct sk_buff * skb,const struct nf_hook_state * state) nf_ct_bridge_in() argument 324 nf_ct_bridge_frag_save(struct sk_buff * skb,struct nf_bridge_frag_data * data) nf_ct_bridge_frag_save() argument 338 nf_ct_bridge_refrag(struct sk_buff * skb,const struct nf_hook_state * state,int (* output)(struct net *,struct sock * sk,const struct nf_bridge_frag_data * data,struct sk_buff *)) nf_ct_bridge_refrag() argument 365 nf_ct_bridge_frag_restore(struct sk_buff * skb,const struct nf_bridge_frag_data * data) nf_ct_bridge_frag_restore() argument 388 nf_ct_bridge_refrag_post(struct net * net,struct sock * sk,const struct nf_bridge_frag_data * data,struct sk_buff * skb) nf_ct_bridge_refrag_post() argument 399 nf_ct_bridge_post(void * priv,struct sk_buff * skb,const struct nf_hook_state * state) nf_ct_bridge_post() argument [all...] |
| /linux/net/netfilter/ |
| H A D | nf_flow_table_ip.c | 23 struct sk_buff *skb, unsigned int thoff) in nf_flow_state_check() argument 30 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_state_check() 43 static void nf_flow_nat_ip_tcp(struct sk_buff *skb, unsigned int thoff, in nf_flow_nat_ip_tcp() argument 48 tcph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ip_tcp() 49 inet_proto_csum_replace4(&tcph->check, skb, addr, new_addr, true); in nf_flow_nat_ip_tcp() 52 static void nf_flow_nat_ip_udp(struct sk_buff *skb, unsigned int thoff, in nf_flow_nat_ip_udp() argument 57 udph = (void *)(skb_network_header(skb) + thoff); in nf_flow_nat_ip_udp() 58 if (udph->check || skb->ip_summed == CHECKSUM_PARTIAL) { in nf_flow_nat_ip_udp() 59 inet_proto_csum_replace4(&udph->check, skb, addr, in nf_flow_nat_ip_udp() 66 static void nf_flow_nat_ip_l4proto(struct sk_buff *skb, struct iphdr *iph, in nf_flow_nat_ip_l4proto() argument [all …]
|
| /linux/net/sched/ |
| H A D | sch_frag.c | 18 int (*xmit)(struct sk_buff *skb); 26 static int sch_frag_xmit(struct net *net, struct sock *sk, struct sk_buff *skb) in sch_frag_xmit() 31 if (skb_cow_head(skb, data->l2_len) < 0) { in sch_frag_xmit() 32 kfree_skb(skb); in sch_frag_xmit() 36 __skb_dst_copy(skb, data->dst); in sch_frag_xmit() 37 *qdisc_skb_cb(skb) = data->cb; in sch_frag_xmit() 38 skb->inner_protocol = data->inner_protocol; in sch_frag_xmit() 40 __vlan_hwaccel_put_tag(skb, data->vlan_proto, in sch_frag_xmit() 43 __vlan_hwaccel_clear_tag(skb); in sch_frag_xmit() 46 skb_push(skb, dat in sch_frag_xmit() 23 sch_frag_xmit(struct net * net,struct sock * sk,struct sk_buff * skb) sch_frag_xmit() argument 50 sch_frag_prepare_frag(struct sk_buff * skb,int (* xmit)(struct sk_buff * skb)) sch_frag_prepare_frag() argument 51 sch_frag_prepare_frag(struct sk_buff * skb,int (* xmit)(struct sk_buff * skb)) sch_frag_prepare_frag() argument 84 sch_fragment(struct net * net,struct sk_buff * skb,u16 mru,int (* xmit)(struct sk_buff * skb)) sch_fragment() argument 85 sch_fragment(struct net * net,struct sk_buff * skb,u16 mru,int (* xmit)(struct sk_buff * skb)) sch_fragment() argument 140 sch_frag_xmit_hook(struct sk_buff * skb,int (* xmit)(struct sk_buff * skb)) sch_frag_xmit_hook() argument [all...] |
| /linux/drivers/net/ethernet/qualcomm/rmnet/ |
| H A D | rmnet_handlers.c | 22 static void rmnet_set_skb_proto(struct sk_buff *skb) in rmnet_set_skb_proto() argument 24 switch (skb->data[0] & 0xF0) { in rmnet_set_skb_proto() 26 skb->protocol = htons(ETH_P_IP); in rmnet_set_skb_proto() 29 skb->protocol = htons(ETH_P_IPV6); in rmnet_set_skb_proto() 32 skb->protocol = htons(ETH_P_MAP); in rmnet_set_skb_proto() 40 rmnet_deliver_skb(struct sk_buff *skb) in rmnet_deliver_skb() argument 42 struct rmnet_priv *priv = netdev_priv(skb->dev); in rmnet_deliver_skb() 44 skb_reset_transport_header(skb); in rmnet_deliver_skb() 45 skb_reset_network_header(skb); in rmnet_deliver_skb() 46 rmnet_vnd_rx_fixup(skb, skb->dev); in rmnet_deliver_skb() [all …]
|