| /linux/drivers/net/ethernet/netronome/nfp/ |
| H A D | nfp_net_dp.c | 76 nfp_net_rx_ring_init(struct nfp_net_rx_ring *rx_ring, in nfp_net_rx_ring_init() argument 81 rx_ring->idx = idx; in nfp_net_rx_ring_init() 82 rx_ring->r_vec = r_vec; in nfp_net_rx_ring_init() 83 u64_stats_init(&rx_ring->r_vec->rx_sync); in nfp_net_rx_ring_init() 85 rx_ring->fl_qcidx = rx_ring->idx * nn->stride_rx; in nfp_net_rx_ring_init() 86 rx_ring->qcp_fl = nn->rx_bar + NFP_QCP_QUEUE_OFF(rx_ring->fl_qcidx); in nfp_net_rx_ring_init() 95 void nfp_net_rx_ring_reset(struct nfp_net_rx_ring *rx_ring) in nfp_net_rx_ring_reset() argument 102 if (rx_ring->wr_p == 0 && rx_ring->rd_p == 0) in nfp_net_rx_ring_reset() 106 wr_idx = D_IDX(rx_ring, rx_ring->wr_p); in nfp_net_rx_ring_reset() 107 last_idx = rx_ring->cnt - 1; in nfp_net_rx_ring_reset() [all …]
|
| H A D | nfp_net_xsk.c | 17 nfp_net_xsk_rx_bufs_stash(struct nfp_net_rx_ring *rx_ring, unsigned int idx, in nfp_net_xsk_rx_bufs_stash() argument 22 headroom = xsk_pool_get_headroom(rx_ring->r_vec->xsk_pool); in nfp_net_xsk_rx_bufs_stash() 24 rx_ring->rxds[idx].fld.reserved = 0; in nfp_net_xsk_rx_bufs_stash() 25 rx_ring->rxds[idx].fld.meta_len_dd = 0; in nfp_net_xsk_rx_bufs_stash() 27 rx_ring->xsk_rxbufs[idx].xdp = xdp; in nfp_net_xsk_rx_bufs_stash() 28 rx_ring->xsk_rxbufs[idx].dma_addr = in nfp_net_xsk_rx_bufs_stash() 46 void nfp_net_xsk_rx_bufs_free(struct nfp_net_rx_ring *rx_ring) in nfp_net_xsk_rx_bufs_free() argument 50 if (!rx_ring->cnt) in nfp_net_xsk_rx_bufs_free() 53 for (i = 0; i < rx_ring->cnt - 1; i++) in nfp_net_xsk_rx_bufs_free() 54 nfp_net_xsk_rx_free(&rx_ring->xsk_rxbufs[i]); in nfp_net_xsk_rx_bufs_free() [all …]
|
| H A D | nfp_net_debugfs.c | 15 struct nfp_net_rx_ring *rx_ring; in nfp_rx_q_show() local 24 if (!r_vec->nfp_net || !r_vec->rx_ring) in nfp_rx_q_show() 27 rx_ring = r_vec->rx_ring; in nfp_rx_q_show() 31 rxd_cnt = rx_ring->cnt; in nfp_rx_q_show() 33 fl_rd_p = nfp_qcp_rd_ptr_read(rx_ring->qcp_fl); in nfp_rx_q_show() 34 fl_wr_p = nfp_qcp_wr_ptr_read(rx_ring->qcp_fl); in nfp_rx_q_show() 37 rx_ring->idx, rx_ring->fl_qcidx, in nfp_rx_q_show() 38 rx_ring->cnt, &rx_ring->dma, rx_ring->rxds, in nfp_rx_q_show() 39 rx_ring->rd_p, rx_ring->wr_p, fl_rd_p, fl_wr_p); in nfp_rx_q_show() 42 rxd = &rx_ring->rxds[i]; in nfp_rx_q_show() [all …]
|
| H A D | nfp_net_xsk.h | 18 static inline int nfp_net_rx_space(struct nfp_net_rx_ring *rx_ring) in nfp_net_rx_space() argument 20 return rx_ring->cnt - rx_ring->wr_p + rx_ring->rd_p - 1; in nfp_net_rx_space() 35 void nfp_net_xsk_rx_bufs_free(struct nfp_net_rx_ring *rx_ring); 37 void nfp_net_xsk_rx_ring_fill_freelist(struct nfp_net_rx_ring *rx_ring);
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 37 struct igb_ring *rx_ring = adapter->rx_ring[qid]; in igb_txrx_ring_disable() local 43 wr32(E1000_RXDCTL(rx_ring->reg_idx), 0); in igb_txrx_ring_disable() 48 napi_disable(&rx_ring->q_vector->napi); in igb_txrx_ring_disable() 51 igb_clean_rx_ring(rx_ring); in igb_txrx_ring_disable() 53 memset(&rx_ring->rx_stats, 0, sizeof(rx_ring->rx_stats)); in igb_txrx_ring_disable() 60 struct igb_ring *rx_ring = adapter->rx_ring[qid]; in igb_txrx_ring_enable() local 63 igb_configure_rx_ring(adapter, rx_ring); in igb_txrx_ring_enable() 73 if (rx_ring->xsk_pool) in igb_txrx_ring_enable() 74 igb_alloc_rx_buffers_zc(rx_ring, rx_ring->xsk_pool, in igb_txrx_ring_enable() 75 igb_desc_unused(rx_ring)); in igb_txrx_ring_enable() [all …]
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 10 void i40e_clear_rx_bi_zc(struct i40e_ring *rx_ring) in i40e_clear_rx_bi_zc() argument 12 memset(rx_ring->rx_bi_zc, 0, in i40e_clear_rx_bi_zc() 13 sizeof(*rx_ring->rx_bi_zc) * rx_ring->count); in i40e_clear_rx_bi_zc() 16 static struct xdp_buff **i40e_rx_bi(struct i40e_ring *rx_ring, u32 idx) in i40e_rx_bi() argument 18 return &rx_ring->rx_bi_zc[idx]; in i40e_rx_bi() 30 static int i40e_realloc_rx_xdp_bi(struct i40e_ring *rx_ring, bool pool_present) in i40e_realloc_rx_xdp_bi() argument 32 size_t elem_size = pool_present ? sizeof(*rx_ring->rx_bi_zc) : in i40e_realloc_rx_xdp_bi() 33 sizeof(*rx_ring->rx_bi); in i40e_realloc_rx_xdp_bi() 34 void *sw_ring = kcalloc(rx_ring->count, elem_size, GFP_KERNEL); in i40e_realloc_rx_xdp_bi() 40 kfree(rx_ring->rx_bi); in i40e_realloc_rx_xdp_bi() [all …]
|
| H A D | i40e_txrx.c | 681 static void i40e_fd_handle_status(struct i40e_ring *rx_ring, u64 qword0_raw, in i40e_fd_handle_status() argument 684 struct i40e_pf *pf = rx_ring->vsi->back; in i40e_fd_handle_status() 1350 static struct i40e_rx_buffer *i40e_rx_bi(struct i40e_ring *rx_ring, u32 idx) in i40e_rx_bi() argument 1352 return &rx_ring->rx_bi[idx]; in i40e_rx_bi() 1362 static void i40e_reuse_rx_page(struct i40e_ring *rx_ring, in i40e_reuse_rx_page() argument 1366 u16 nta = rx_ring->next_to_alloc; in i40e_reuse_rx_page() 1368 new_buff = i40e_rx_bi(rx_ring, nta); in i40e_reuse_rx_page() 1372 rx_ring->next_to_alloc = (nta < rx_ring->count) ? nta : 0; in i40e_reuse_rx_page() 1396 void i40e_clean_programming_status(struct i40e_ring *rx_ring, u64 qword0_raw, in i40e_clean_programming_status() argument 1404 i40e_fd_handle_status(rx_ring, qword0_raw, qword1, id); in i40e_clean_programming_status() [all …]
|
| H A D | i40e_txrx_common.h | 10 void i40e_clean_programming_status(struct i40e_ring *rx_ring, u64 qword0_raw, 12 void i40e_process_skb_fields(struct i40e_ring *rx_ring, 15 void i40e_update_rx_stats(struct i40e_ring *rx_ring, 18 void i40e_finalize_xdp_rx(struct i40e_ring *rx_ring, unsigned int xdp_res); 19 void i40e_release_rx_desc(struct i40e_ring *rx_ring, u32 val); 105 void i40e_xsk_clean_rx_ring(struct i40e_ring *rx_ring);
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_xsk.c | 98 struct ixgbe_ring *rx_ring, in ixgbe_run_xdp_zc() argument 107 xdp_prog = READ_ONCE(rx_ring->xdp_prog); in ixgbe_run_xdp_zc() 111 err = xdp_do_redirect(rx_ring->netdev, xdp, xdp_prog); in ixgbe_run_xdp_zc() 114 if (xsk_uses_need_wakeup(rx_ring->xsk_pool) && err == -ENOBUFS) in ixgbe_run_xdp_zc() 141 bpf_warn_invalid_xdp_action(rx_ring->netdev, xdp_prog, act); in ixgbe_run_xdp_zc() 146 trace_xdp_exception(rx_ring->netdev, xdp_prog, act); in ixgbe_run_xdp_zc() 151 bool ixgbe_alloc_rx_buffers_zc(struct ixgbe_ring *rx_ring, u16 count) in ixgbe_alloc_rx_buffers_zc() argument 155 u16 i = rx_ring->next_to_use; in ixgbe_alloc_rx_buffers_zc() 163 rx_desc = IXGBE_RX_DESC(rx_ring, i); in ixgbe_alloc_rx_buffers_zc() 164 bi = &rx_ring->rx_buffer_info[i]; in ixgbe_alloc_rx_buffers_zc() [all …]
|
| H A D | ixgbe_txrx_common.h | 18 bool ixgbe_cleanup_headers(struct ixgbe_ring *rx_ring, 21 void ixgbe_process_skb_fields(struct ixgbe_ring *rx_ring, 39 bool ixgbe_alloc_rx_buffers_zc(struct ixgbe_ring *rx_ring, u16 cleaned_count); 41 struct ixgbe_ring *rx_ring, 43 void ixgbe_xsk_clean_rx_ring(struct ixgbe_ring *rx_ring); 52 void ixgbe_update_rx_ring_stats(struct ixgbe_ring *rx_ring,
|
| H A D | ixgbe_main.c | 593 struct ixgbe_ring *rx_ring; in ixgbe_dump() local 732 rx_ring = adapter->rx_ring[n]; in ixgbe_dump() 734 n, rx_ring->next_to_use, rx_ring->next_to_clean); in ixgbe_dump() 789 rx_ring = adapter->rx_ring[n]; in ixgbe_dump() 791 pr_info("RX QUEUE INDEX = %d\n", rx_ring->queue_index); in ixgbe_dump() 802 for (i = 0; i < rx_ring->count; i++) { in ixgbe_dump() 805 if (i == rx_ring->next_to_use) in ixgbe_dump() 807 else if (i == rx_ring->next_to_clean) in ixgbe_dump() 812 rx_buffer_info = &rx_ring->rx_buffer_info[i]; in ixgbe_dump() 813 rx_desc = IXGBE_RX_DESC(rx_ring, i); in ixgbe_dump() [all …]
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| H A D | ena_xdp.h | 45 int ena_xdp_register_rxq_info(struct ena_ring *rx_ring); 46 void ena_xdp_unregister_rxq_info(struct ena_ring *rx_ring); 82 static inline int ena_xdp_execute(struct ena_ring *rx_ring, struct xdp_buff *xdp) in ena_xdp_execute() argument 90 xdp_prog = READ_ONCE(rx_ring->xdp_bpf_prog); in ena_xdp_execute() 98 trace_xdp_exception(rx_ring->netdev, xdp_prog, verdict); in ena_xdp_execute() 99 xdp_stat = &rx_ring->rx_stats.xdp_aborted; in ena_xdp_execute() 105 xdp_ring = rx_ring->xdp_ring; in ena_xdp_execute() 110 if (ena_xdp_xmit_frame(xdp_ring, rx_ring->adapter, xdpf, in ena_xdp_execute() 115 xdp_stat = &rx_ring->rx_stats.xdp_tx; in ena_xdp_execute() 119 if (likely(!xdp_do_redirect(rx_ring->netdev, xdp, xdp_prog))) { in ena_xdp_execute() [all …]
|
| H A D | ena_netdev.c | 96 adapter->rx_ring[i].mtu = mtu; in update_rx_ring_mtu() 194 rxr = &adapter->rx_ring[i]; in ena_init_io_rings() 370 struct ena_ring *rx_ring = &adapter->rx_ring[qid]; in ena_setup_rx_resources() local 374 if (rx_ring->rx_buffer_info) { in ena_setup_rx_resources() 383 size = sizeof(struct ena_rx_buffer) * (rx_ring->ring_size + 1); in ena_setup_rx_resources() 386 rx_ring->rx_buffer_info = vzalloc_node(size, node); in ena_setup_rx_resources() 387 if (!rx_ring->rx_buffer_info) { in ena_setup_rx_resources() 388 rx_ring->rx_buffer_info = vzalloc(size); in ena_setup_rx_resources() 389 if (!rx_ring->rx_buffer_info) in ena_setup_rx_resources() 393 size = sizeof(u16) * rx_ring->ring_size; in ena_setup_rx_resources() [all …]
|
| H A D | ena_xdp.c | 135 if (!READ_ONCE(adapter->rx_ring->xdp_bpf_prog)) in ena_xdp_xmit() 196 int ena_xdp_register_rxq_info(struct ena_ring *rx_ring) in ena_xdp_register_rxq_info() argument 200 rc = xdp_rxq_info_reg(&rx_ring->xdp_rxq, rx_ring->netdev, rx_ring->qid, 0); in ena_xdp_register_rxq_info() 202 netif_dbg(rx_ring->adapter, ifup, rx_ring->netdev, "Registering RX info for queue %d", in ena_xdp_register_rxq_info() 203 rx_ring->qid); in ena_xdp_register_rxq_info() 205 netif_err(rx_ring->adapter, ifup, rx_ring->netdev, in ena_xdp_register_rxq_info() 207 rx_ring->qid, rc); in ena_xdp_register_rxq_info() 211 rc = xdp_rxq_info_reg_mem_model(&rx_ring->xdp_rxq, MEM_TYPE_PAGE_SHARED, NULL); in ena_xdp_register_rxq_info() 214 netif_err(rx_ring->adapter, ifup, rx_ring->netdev, in ena_xdp_register_rxq_info() 216 rx_ring->qid, rc); in ena_xdp_register_rxq_info() [all …]
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.c | 711 static void iavf_clean_rx_ring(struct iavf_ring *rx_ring) in iavf_clean_rx_ring() argument 714 if (!rx_ring->rx_fqes) in iavf_clean_rx_ring() 717 if (rx_ring->skb) { in iavf_clean_rx_ring() 718 dev_kfree_skb(rx_ring->skb); in iavf_clean_rx_ring() 719 rx_ring->skb = NULL; in iavf_clean_rx_ring() 723 for (u32 i = rx_ring->next_to_clean; i != rx_ring->next_to_use; ) { in iavf_clean_rx_ring() 724 const struct libeth_fqe *rx_fqes = &rx_ring->rx_fqes[i]; in iavf_clean_rx_ring() 728 if (unlikely(++i == rx_ring->count)) in iavf_clean_rx_ring() 732 rx_ring->next_to_clean = 0; in iavf_clean_rx_ring() 733 rx_ring->next_to_use = 0; in iavf_clean_rx_ring() [all …]
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_xdp.c | 30 igc_disable_rx_ring(adapter->rx_ring[i]); in igc_xdp_set_prog() 32 napi_disable(&adapter->rx_ring[i]->q_vector->napi); in igc_xdp_set_prog() 47 napi_enable(&adapter->rx_ring[i]->q_vector->napi); in igc_xdp_set_prog() 49 igc_enable_rx_ring(adapter->rx_ring[i]); in igc_xdp_set_prog() 61 struct igc_ring *rx_ring, *tx_ring; in igc_xdp_enable_pool() local 89 rx_ring = adapter->rx_ring[queue_id]; in igc_xdp_enable_pool() 92 napi = &rx_ring->q_vector->napi; in igc_xdp_enable_pool() 95 igc_disable_rx_ring(rx_ring); in igc_xdp_enable_pool() 100 set_bit(IGC_RING_FLAG_AF_XDP_ZC, &rx_ring->flags); in igc_xdp_enable_pool() 105 igc_enable_rx_ring(rx_ring); in igc_xdp_enable_pool() [all …]
|
| H A D | igc_dump.c | 119 struct igc_ring *rx_ring; in igc_rings_dump() local 214 rx_ring = adapter->rx_ring[n]; in igc_rings_dump() 215 netdev_info(netdev, "%5d %5X %5X\n", n, rx_ring->next_to_use, in igc_rings_dump() 216 rx_ring->next_to_clean); in igc_rings_dump() 247 rx_ring = adapter->rx_ring[n]; in igc_rings_dump() 250 rx_ring->queue_index); in igc_rings_dump() 255 for (i = 0; i < rx_ring->count; i++) { in igc_rings_dump() 259 buffer_info = &rx_ring->rx_buffer_info[i]; in igc_rings_dump() 260 rx_desc = IGC_RX_DESC(rx_ring, i); in igc_rings_dump() 264 if (i == rx_ring->next_to_use) in igc_rings_dump() [all …]
|
| /linux/drivers/net/ethernet/mscc/ |
| H A D | ocelot_fdma.c | 58 struct ocelot_fdma_rx_ring *rx_ring = &fdma->rx_ring; in ocelot_fdma_rx_ring_free() local 60 if (rx_ring->next_to_use >= rx_ring->next_to_clean) in ocelot_fdma_rx_ring_free() 62 (rx_ring->next_to_use - rx_ring->next_to_clean) - 1; in ocelot_fdma_rx_ring_free() 64 return rx_ring->next_to_clean - rx_ring->next_to_use - 1; in ocelot_fdma_rx_ring_free() 149 struct ocelot_fdma_rx_ring *rx_ring; in ocelot_fdma_alloc_rx_buffs() local 156 rx_ring = &fdma->rx_ring; in ocelot_fdma_alloc_rx_buffs() 157 idx = rx_ring->next_to_use; in ocelot_fdma_alloc_rx_buffs() 160 rxb = &rx_ring->bufs[idx]; in ocelot_fdma_alloc_rx_buffs() 171 dcb = &rx_ring->dcbs[idx]; in ocelot_fdma_alloc_rx_buffs() 177 dcb->llp = ocelot_fdma_idx_dma(rx_ring->dcbs_dma, idx); in ocelot_fdma_alloc_rx_buffs() [all …]
|
| /linux/drivers/net/ethernet/aquantia/atlantic/ |
| H A D | aq_ring.c | 52 static int aq_alloc_rxpages(struct aq_rxpage *rxpage, struct aq_ring_s *rx_ring) in aq_alloc_rxpages() argument 54 struct device *dev = aq_nic_get_dev(rx_ring->aq_nic); in aq_alloc_rxpages() 55 unsigned int order = rx_ring->page_order; in aq_alloc_rxpages() 73 rxpage->pg_off = rx_ring->page_offset; in aq_alloc_rxpages() 407 struct aq_ring_s *rx_ring, in aq_xdp_run_prog() argument 417 u64_stats_update_begin(&rx_ring->stats.rx.syncp); in aq_xdp_run_prog() 418 ++rx_ring->stats.rx.packets; in aq_xdp_run_prog() 419 rx_ring->stats.rx.bytes += xdp_get_buff_len(xdp); in aq_xdp_run_prog() 420 u64_stats_update_end(&rx_ring->stats.rx.syncp); in aq_xdp_run_prog() 422 prog = READ_ONCE(rx_ring->xdp_prog); in aq_xdp_run_prog() [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbevf/ |
| H A D | ixgbevf_main.c | 116 static void ixgbevf_reuse_rx_page(struct ixgbevf_ring *rx_ring, 510 static void ixgbevf_process_skb_fields(struct ixgbevf_ring *rx_ring, in ixgbevf_process_skb_fields() argument 514 ixgbevf_rx_hash(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields() 515 ixgbevf_rx_checksum(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields() 519 unsigned long *active_vlans = netdev_priv(rx_ring->netdev); in ixgbevf_process_skb_fields() 526 ixgbevf_ipsec_rx(rx_ring, rx_desc, skb); in ixgbevf_process_skb_fields() 528 skb->protocol = eth_type_trans(skb, rx_ring->netdev); in ixgbevf_process_skb_fields() 532 struct ixgbevf_rx_buffer *ixgbevf_get_rx_buffer(struct ixgbevf_ring *rx_ring, in ixgbevf_get_rx_buffer() argument 537 rx_buffer = &rx_ring->rx_buffer_info[rx_ring->next_to_clean]; in ixgbevf_get_rx_buffer() 541 dma_sync_single_range_for_cpu(rx_ring->dev, in ixgbevf_get_rx_buffer() [all …]
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_netdev.c | 83 int fm10k_setup_rx_resources(struct fm10k_ring *rx_ring) in fm10k_setup_rx_resources() argument 85 struct device *dev = rx_ring->dev; in fm10k_setup_rx_resources() 88 size = sizeof(struct fm10k_rx_buffer) * rx_ring->count; in fm10k_setup_rx_resources() 90 rx_ring->rx_buffer = vzalloc(size); in fm10k_setup_rx_resources() 91 if (!rx_ring->rx_buffer) in fm10k_setup_rx_resources() 94 u64_stats_init(&rx_ring->syncp); in fm10k_setup_rx_resources() 97 rx_ring->size = rx_ring->count * sizeof(union fm10k_rx_desc); in fm10k_setup_rx_resources() 98 rx_ring->size = ALIGN(rx_ring->size, 4096); in fm10k_setup_rx_resources() 100 rx_ring->desc = dma_alloc_coherent(dev, rx_ring->size, in fm10k_setup_rx_resources() 101 &rx_ring->dma, GFP_KERNEL); in fm10k_setup_rx_resources() [all …]
|
| /linux/drivers/net/ethernet/agere/ |
| H A D | et131x.c | 288 struct rx_ring { struct 492 struct rx_ring rx_ring; member 740 struct rx_ring *rx_ring = &adapter->rx_ring; in et131x_rx_dma_enable() local 742 if (rx_ring->fbr[1]->buffsize == 4096) in et131x_rx_dma_enable() 744 else if (rx_ring->fbr[1]->buffsize == 8192) in et131x_rx_dma_enable() 746 else if (rx_ring->fbr[1]->buffsize == 16384) in et131x_rx_dma_enable() 750 if (rx_ring->fbr[0]->buffsize == 256) in et131x_rx_dma_enable() 752 else if (rx_ring->fbr[0]->buffsize == 512) in et131x_rx_dma_enable() 754 else if (rx_ring->fbr[0]->buffsize == 1024) in et131x_rx_dma_enable() 1539 struct rx_ring *rx_local = &adapter->rx_ring; in et131x_config_rx_dma_regs() [all …]
|
| /linux/drivers/net/ethernet/spacemit/ |
| H A D | k1_emac.c | 108 struct emac_desc_ring rx_ring; member 373 struct emac_desc_ring *rx_ring; in emac_clean_rx_desc_ring() local 376 rx_ring = &priv->rx_ring; in emac_clean_rx_desc_ring() 378 for (i = 0; i < rx_ring->total_cnt; i++) { in emac_clean_rx_desc_ring() 379 rx_buf = &rx_ring->rx_desc_buf[i]; in emac_clean_rx_desc_ring() 391 rx_ring->tail = 0; in emac_clean_rx_desc_ring() 392 rx_ring->head = 0; in emac_clean_rx_desc_ring() 426 struct emac_desc_ring *rx_ring = &priv->rx_ring; in emac_alloc_rx_resources() local 429 rx_ring->rx_desc_buf = kcalloc(rx_ring->total_cnt, in emac_alloc_rx_resources() 430 sizeof(*rx_ring->rx_desc_buf), in emac_alloc_rx_resources() [all …]
|
| /linux/drivers/net/wireless/ath/ath10k/ |
| H A D | htt_rx.c | 38 hash_for_each_possible(ar->htt.rx_ring.skb_table, rxcb, hlist, paddr) in ath10k_htt_rx_find_skb_paddr() 53 if (htt->rx_ring.in_ord_rx) { in ath10k_htt_rx_ring_free() 54 hash_for_each_safe(htt->rx_ring.skb_table, i, n, rxcb, hlist) { in ath10k_htt_rx_ring_free() 63 for (i = 0; i < htt->rx_ring.size; i++) { in ath10k_htt_rx_ring_free() 64 skb = htt->rx_ring.netbufs_ring[i]; in ath10k_htt_rx_ring_free() 76 htt->rx_ring.fill_cnt = 0; in ath10k_htt_rx_ring_free() 77 hash_init(htt->rx_ring.skb_table); in ath10k_htt_rx_ring_free() 78 memset(htt->rx_ring.netbufs_ring, 0, in ath10k_htt_rx_ring_free() 79 htt->rx_ring.size * sizeof(htt->rx_ring.netbufs_ring[0])); in ath10k_htt_rx_ring_free() 84 return htt->rx_ring.size * sizeof(htt->rx_ring.paddrs_ring_32); in ath10k_htt_get_rx_ring_size_32() [all …]
|
| /linux/drivers/net/ethernet/intel/e1000/ |
| H A D | e1000_main.c | 81 struct e1000_rx_ring *rx_ring); 100 struct e1000_rx_ring *rx_ring); 114 struct e1000_rx_ring *rx_ring, 117 struct e1000_rx_ring *rx_ring, 120 struct e1000_rx_ring *rx_ring, in e1000_alloc_dummy_rx_buffers() argument 125 struct e1000_rx_ring *rx_ring, 128 struct e1000_rx_ring *rx_ring, 375 struct e1000_rx_ring *ring = &adapter->rx_ring[i]; in e1000_configure() 1226 kfree(adapter->rx_ring); in e1000_probe() 1271 kfree(adapter->rx_ring); in e1000_remove() [all …]
|