Searched refs:softnet_data (Results 1 – 8 of 8) sorted by relevance
85 static u32 softnet_input_pkt_queue_len(struct softnet_data *sd) in softnet_input_pkt_queue_len()90 static u32 softnet_process_queue_len(struct softnet_data *sd) in softnet_process_queue_len()95 static struct softnet_data *softnet_get_online(loff_t *pos) in softnet_get_online()97 struct softnet_data *sd = NULL; in softnet_get_online()101 sd = &per_cpu(softnet_data, *pos); in softnet_get_online()125 struct softnet_data *sd = v; in softnet_seq_show()
231 static inline void backlog_lock_irq_save(struct softnet_data *sd, in backlog_lock_irq_save()243 static inline void backlog_lock_irq_disable(struct softnet_data *sd) in backlog_lock_irq_disable()251 static inline void backlog_unlock_irq_restore(struct softnet_data *sd, in backlog_unlock_irq_restore()263 static inline void backlog_unlock_irq_enable(struct softnet_data *sd) in backlog_unlock_irq_enable()462 DEFINE_PER_CPU_ALIGNED(struct softnet_data, softnet_data) = {465 EXPORT_PER_CPU_SYMBOL(softnet_data);3382 struct softnet_data *sd; in __netif_reschedule()3386 sd = this_cpu_ptr(&softnet_data); in __netif_reschedule()3457 skb->next = __this_cpu_read(softnet_data.completion_queue); in dev_kfree_skb_irq_reason()3458 __this_cpu_write(softnet_data.completion_queue, skb); in dev_kfree_skb_irq_reason()[all …]
214 struct softnet_data *sd; in flow_limit_cpu_sysctl()229 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()253 sd = &per_cpu(softnet_data, i); in flow_limit_cpu_sysctl()
246 struct softnet_data *sd = &get_cpu_var(softnet_data); in zap_completion_queue()268 put_cpu_var(softnet_data); in zap_completion_queue()
120 struct softnet_data *sd; in validate_xmit_xfrm()149 sd = this_cpu_ptr(&softnet_data); in validate_xmit_xfrm()472 struct softnet_data *sd; in xfrm_dev_resume()485 sd = this_cpu_ptr(&softnet_data); in xfrm_dev_resume()494 void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()
3515 struct softnet_data { struct3524 struct softnet_data *rps_ipi_list; argument3550 struct softnet_data *rps_ipi_next; argument3571 DECLARE_PER_CPU_ALIGNED(struct softnet_data, softnet_data); argument3582 return this_cpu_read(softnet_data.xmit.recursion); in dev_recursion_level()5268 __this_cpu_write(softnet_data.xmit.more, more); in netdev_xmit_set_more()5273 return __this_cpu_read(softnet_data.xmit.more); in netdev_xmit_more()
435 xmit = this_cpu_ptr(&softnet_data.xmit); in tcf_mirred_act()
2068 void xfrm_dev_backlog(struct softnet_data *sd);2133 static inline void xfrm_dev_backlog(struct softnet_data *sd) in xfrm_dev_backlog()