| /linux/tools/testing/selftests/bpf/progs/ |
| H A D | xdp_features.c | 65 xdp_process_echo_packet(struct xdp_md *xdp, bool dut) in xdp_process_echo_packet() 135 xdp_update_stats(struct xdp_md *xdp, bool tx, bool dut) in xdp_update_stats() 156 int xdp_tester_check_tx(struct xdp_md *xdp) in xdp_tester_check_tx() 164 int xdp_tester_check_rx(struct xdp_md *xdp) in xdp_tester_check_rx() 174 int xdp_do_pass(struct xdp_md *xdp) in xdp_do_pass() 182 int xdp_do_drop(struct xdp_md *xdp) in xdp_do_drop() 191 int xdp_do_aborted(struct xdp_md *xdp) in xdp_do_aborted() 200 int xdp_do_tx(struct xdp_md *xdp) in xdp_do_tx() 217 int xdp_do_redirect(struct xdp_md *xdp) in xdp_do_redirect() 227 const struct bpf_prog *xdp, __u32 act) in BPF_PROG() [all …]
|
| H A D | xsk_xdp_progs.c | 19 xsk_def_prog(struct xdp_md * xdp) xsk_def_prog() argument 24 xsk_xdp_drop(struct xdp_md * xdp) xsk_xdp_drop() argument 33 xsk_xdp_populate_metadata(struct xdp_md * xdp) xsk_xdp_populate_metadata() argument 56 xsk_xdp_shared_umem(struct xdp_md * xdp) xsk_xdp_shared_umem() argument [all...] |
| H A D | xdp_redirect_map.c | 14 xdp_redirect_map_0(struct xdp_md * xdp) xdp_redirect_map_0() argument 20 xdp_redirect_map_1(struct xdp_md * xdp) xdp_redirect_map_1() argument 26 xdp_redirect_map_2(struct xdp_md * xdp) xdp_redirect_map_2() argument [all...] |
| H A D | test_xdp_do_redirect.c | 29 int xdp_redirect(struct xdp_md *xdp) in xdp_redirect() 86 int xdp_count_pkts(struct xdp_md *xdp) in xdp_count_pkts() 102 int xdp_redirect_to_111(struct xdp_md *xdp) in xdp_redirect_to_111() 108 int xdp_redirect_to_222(struct xdp_md *xdp) in xdp_redirect_to_222()
|
| H A D | test_xdp.c | 79 static __always_inline int handle_ipv4(struct xdp_md *xdp) in handle_ipv4() 152 static __always_inline int handle_ipv6(struct xdp_md *xdp) in handle_ipv6() 213 int _xdp_tx_iptunnel(struct xdp_md *xdp) in _xdp_tx_iptunnel()
|
| H A D | test_xdp_loop.c | 75 static __always_inline int handle_ipv4(struct xdp_md *xdp) in handle_ipv4() 148 static __always_inline int handle_ipv6(struct xdp_md *xdp) in handle_ipv6() 209 int _xdp_tx_iptunnel(struct xdp_md *xdp) in _xdp_tx_iptunnel()
|
| H A D | dynptr_success.c | 1023 static __always_inline void test_dynptr_probe_str_xdp(struct xdp_md *xdp, voi argument 611 test_dynptr_copy_xdp(struct xdp_md * xdp) test_dynptr_copy_xdp() argument 810 test_dynptr_memset_xdp_chunks(struct xdp_md * xdp) test_dynptr_memset_xdp_chunks() argument 940 test_dynptr_probe_xdp(struct xdp_md * xdp,void * ptr,bpf_read_dynptr_fn_t bpf_read_dynptr_fn) test_dynptr_probe_xdp() argument 968 test_dynptr_probe_str_xdp(struct xdp_md * xdp,void * ptr,bpf_read_dynptr_fn_t bpf_read_dynptr_fn) test_dynptr_probe_str_xdp() argument 1005 test_probe_read_user_dynptr(struct xdp_md * xdp) test_probe_read_user_dynptr() argument 1014 test_probe_read_kernel_dynptr(struct xdp_md * xdp) test_probe_read_kernel_dynptr() argument 1032 test_probe_read_kernel_str_dynptr(struct xdp_md * xdp) test_probe_read_kernel_str_dynptr() argument [all...] |
| H A D | test_xdp_dynptr.c | 77 static __always_inline int handle_ipv4(struct xdp_md *xdp, struct bpf_dynptr *xdp_ptr) in handle_ipv4() 159 static __always_inline int handle_ipv6(struct xdp_md *xdp, struct bpf_dynptr *xdp_ptr) in handle_ipv6() 231 int _xdp_tx_iptunnel(struct xdp_md *xdp) in _xdp_tx_iptunnel()
|
| H A D | test_parse_tcp_hdr_opt.c | 36 static int parse_hdr_opt(const struct xdp_md *xdp, struct hdr_opt_state *state) in parse_hdr_opt() 82 int xdp_ingress_v6(struct xdp_md *xdp) in xdp_ingress_v6()
|
| H A D | test_xdp_bpf2bpf.c | 45 int BPF_PROG(trace_on_entry, struct xdp_buff *xdp) in BPF_PROG() 62 int BPF_PROG(trace_on_exit, struct xdp_buff *xdp, int ret) in BPF_PROG()
|
| /linux/include/net/libeth/ |
| H A D | xsk.h | 25 struct libeth_xdp_buff *xdp) in libeth_xsk_tx_queue_head() 65 struct libeth_xdp_buff *xdp, in libeth_xsk_tx_queue_bulk() 115 struct libeth_xdp_buff *xdp = frm.xsk; in libeth_xsk_tx_fill_buf() local 352 struct libeth_xdp_buff *xdp, u32 len) in libeth_xsk_process_buff() 390 __libeth_xsk_run_prog(struct libeth_xdp_buff *xdp, in __libeth_xsk_run_prog() 419 #define libeth_xsk_run_prog(xdp, bq, fl) \ argument 441 __libeth_xsk_run_pass(struct libeth_xdp_buff *xdp, in __libeth_xsk_run_pass() 444 void (*prep)(struct libeth_xdp_buff *xdp, in __libeth_xsk_run_pass() 446 u32 (*run)(struct libeth_xdp_buff *xdp, in __libeth_xsk_run_pass() 449 const struct libeth_xdp_buff *xdp, in __libeth_xsk_run_pass() [all …]
|
| H A D | xdp.h | 599 const struct libeth_xdp_buff *xdp) in libeth_xdp_tx_queue_head() 638 struct libeth_xdp_buff *xdp, in libeth_xdp_tx_queue_bulk() 1248 #define libeth_xdp_return_buff(xdp) __libeth_xdp_return_buff(xdp, true) argument 1250 static inline void __libeth_xdp_return_buff(struct libeth_xdp_buff *xdp, in __libeth_xdp_return_buff() 1278 static inline void libeth_xdp_prepare_buff(struct libeth_xdp_buff *xdp, in libeth_xdp_prepare_buff() 1303 static inline bool libeth_xdp_process_buff(struct libeth_xdp_buff *xdp, in libeth_xdp_process_buff() 1330 const struct libeth_xdp_buff *xdp) in libeth_xdp_buff_stats_frags() 1355 __libeth_xdp_run_prog(struct libeth_xdp_buff *xdp, in __libeth_xdp_run_prog() 1403 __libeth_xdp_run_flush(struct libeth_xdp_buff *xdp, in __libeth_xdp_run_flush() 1405 u32 (*run)(struct libeth_xdp_buff *xdp, in __libeth_xdp_run_flush() [all …]
|
| /linux/drivers/net/ethernet/intel/libeth/ |
| H A D | xsk.c | 36 void libeth_xsk_buff_free_slow(struct libeth_xdp_buff *xdp) in libeth_xsk_buff_free_slow() 54 struct libeth_xdp_buff *xdp) in libeth_xsk_buff_add_frag() 78 const struct libeth_xdp_buff *xdp) in libeth_xsk_buff_stats_frags() 98 u32 __libeth_xsk_run_prog_slow(struct libeth_xdp_buff *xdp, in __libeth_xsk_run_prog_slow() 130 u32 __cold libeth_xsk_prog_exception(struct libeth_xdp_buff *xdp, in libeth_xsk_prog_exception()
|
| H A D | xdp.c | 244 void __cold libeth_xdp_return_buff_slow(struct libeth_xdp_buff *xdp) in libeth_xdp_return_buff_slow() 261 bool libeth_xdp_buff_add_frag(struct libeth_xdp_buff *xdp, in libeth_xdp_buff_add_frag() 296 struct libeth_xdp_buff *xdp, in libeth_xdp_prog_exception()
|
| /linux/include/trace/events/ |
| H A D | xdp.h | 159 #define _trace_xdp_redirect(dev, xdp, to) \ argument 162 #define _trace_xdp_redirect_err(dev, xdp, to, err) \ argument 165 #define _trace_xdp_redirect_map(dev, xdp, to, map_type, map_id, index) \ argument 168 #define _trace_xdp_redirect_map_err(dev, xdp, to, map_type, map_id, index, err) \ argument
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 176 static u16 igb_fill_rx_descs(struct xsk_buff_pool *pool, struct xdp_buff **xdp, in igb_fill_rx_descs() 207 struct xdp_buff **xdp; in igb_alloc_rx_buffers_zc() local 256 struct xdp_buff *xdp = rx_ring->rx_buffer_info_zc[ntc]; in igb_clean_rx_ring_zc() local 266 struct xdp_buff *xdp, in igb_construct_skb_zc() 295 struct xdp_buff *xdp, struct xsk_buff_pool *xsk_pool, in igb_run_xdp_zc() 360 struct xdp_buff *xdp; in igb_clean_rx_irq_zc() local
|
| /linux/samples/bpf/ |
| H A D | xdp_sample.bpf.c | 97 const struct bpf_prog *xdp, const void *tgt, int err, in BPF_PROG() 105 const struct bpf_prog *xdp, const void *tgt, int err, in BPF_PROG() 113 const struct bpf_prog *xdp, const void *tgt, int err, in BPF_PROG() 121 const struct bpf_prog *xdp, const void *tgt, int err, in BPF_PROG() 182 const struct bpf_prog *xdp, u32 act) in BPF_PROG()
|
| H A D | xdp_tx_iptunnel_kern.c | 77 static __always_inline int handle_ipv4(struct xdp_md *xdp) in handle_ipv4() 152 static __always_inline int handle_ipv6(struct xdp_md *xdp) in handle_ipv6() 216 int _xdp_tx_iptunnel(struct xdp_md *xdp) in _xdp_tx_iptunnel()
|
| /linux/drivers/net/ethernet/microchip/lan966x/ |
| H A D | lan966x_xdp.c | 9 static int lan966x_xdp_setup(struct net_device *dev, struct netdev_bpf *xdp) in lan966x_xdp_setup() 43 int lan966x_xdp(struct net_device *dev, struct netdev_bpf *xdp) in lan966x_xdp() 79 struct xdp_buff xdp; in lan966x_xdp_run() local
|
| /linux/drivers/net/vmxnet3/ |
| H A D | vmxnet3_xdp.c | 261 vmxnet3_run_xdp(struct vmxnet3_rx_queue *rq, struct xdp_buff *xdp, in vmxnet3_run_xdp() 314 const struct xdp_buff *xdp) in vmxnet3_build_skb() 341 struct xdp_buff xdp; in vmxnet3_process_xdp_small() local 387 struct xdp_buff xdp; in vmxnet3_process_xdp() local
|
| /linux/include/linux/ |
| H A D | if_tun.h | 31 static inline void *tun_xdp_to_ptr(struct xdp_frame *xdp) in tun_xdp_to_ptr() 63 static inline void *tun_xdp_to_ptr(struct xdp_frame *xdp) in tun_xdp_to_ptr()
|
| /linux/drivers/net/ethernet/aquantia/atlantic/ |
| H A D | aq_ring.c | 24 struct xdp_buff *xdp) in aq_get_rxpages_xdp() 386 static struct sk_buff *aq_xdp_build_skb(struct xdp_buff *xdp, in aq_xdp_build_skb() 406 struct xdp_buff *xdp, in aq_xdp_run_prog() 487 struct xdp_buff *xdp) in aq_add_rx_fragment() 725 struct xdp_buff xdp; in __aq_ring_xdp_clean() local
|
| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_xsk.c | 18 struct xdp_buff *xdp) in nfp_net_xsk_rx_bufs_stash() 62 struct xdp_buff *xdp; in nfp_net_xsk_rx_ring_fill_freelist() local
|
| /linux/net/bpf/ |
| H A D | test_run.c | 125 struct xdp_test_data *xdp = arg; in xdp_test_run_init_page() local 150 static int xdp_test_run_setup(struct xdp_test_data *xdp, struct xdp_buff *orig_ctx) in xdp_test_run_setup() 204 static void xdp_test_run_teardown(struct xdp_test_data *xdp) in xdp_test_run_teardown() 274 static int xdp_test_run_batch(struct xdp_test_data *xdp, struct bpf_prog *prog, in xdp_test_run_batch() 365 struct xdp_test_data xdp = { .batch_size = batch_size }; in bpf_test_run_xdp_live() local 390 u32 *retval, u32 *time, bool xdp) in bpf_test_run() 1207 static int xdp_convert_md_to_buff(struct xdp_md *xdp_md, struct xdp_buff *xdp) in xdp_convert_md_to_buff() 1253 static void xdp_convert_buff_to_md(struct xdp_buff *xdp, struct xdp_md *xdp_md) in xdp_convert_buff_to_md() 1277 struct xdp_buff xdp = {}; in bpf_prog_test_run_xdp() local
|
| /linux/net/xdp/ |
| H A D | xsk_buff_pool.c | 586 xp_alloc_new_from_fq(struct xsk_buff_pool * pool,struct xdp_buff ** xdp,u32 max) xp_alloc_new_from_fq() argument 622 xp_alloc_reused(struct xsk_buff_pool * pool,struct xdp_buff ** xdp,u32 nb_entries) xp_alloc_reused() argument 642 xp_alloc_slow(struct xsk_buff_pool * pool,struct xdp_buff ** xdp,u32 max) xp_alloc_slow() argument 660 xp_alloc_batch(struct xsk_buff_pool * pool,struct xdp_buff ** xdp,u32 max) xp_alloc_batch() argument [all...] |