Searched refs:rxr (Results 1 – 8 of 8) sorted by relevance
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt.c | 399 static void bnxt_sched_reset_rxr(struct bnxt *bp, struct bnxt_rx_ring_info *rxr) in bnxt_sched_reset_rxr() argument 401 if (!rxr->bnapi->in_reset) { in bnxt_sched_reset_rxr() 402 rxr->bnapi->in_reset = true; in bnxt_sched_reset_rxr() 409 rxr->rx_next_cons = 0xffff; in bnxt_sched_reset_rxr() 906 static bool bnxt_separate_head_pool(struct bnxt_rx_ring_info *rxr) in bnxt_separate_head_pool() argument 908 return rxr->need_head_pool || PAGE_SIZE > BNXT_RX_PAGE_SIZE; in bnxt_separate_head_pool() 912 struct bnxt_rx_ring_info *rxr, in __bnxt_alloc_rx_page() argument 919 page = page_pool_dev_alloc_frag(rxr->page_pool, offset, in __bnxt_alloc_rx_page() 922 page = page_pool_dev_alloc_pages(rxr->page_pool); in __bnxt_alloc_rx_page() 933 struct bnxt_rx_ring_info *rxr, in __bnxt_alloc_rx_netmem() argument [all …]
|
| H A D | bnxt_xdp.h | 20 bool bnxt_rx_xdp(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, u16 cons, 27 bool bnxt_xdp_attached(struct bnxt *bp, struct bnxt_rx_ring_info *rxr); 29 void bnxt_xdp_buff_init(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, 32 void bnxt_xdp_buff_frags_free(struct bnxt_rx_ring_info *rxr,
|
| H A D | bnxt.h | 2894 int bnxt_alloc_rx_data(struct bnxt *bp, struct bnxt_rx_ring_info *rxr, 2896 void bnxt_reuse_rx_data(struct bnxt_rx_ring_info *rxr, u16 cons, void *data);
|
| /linux/drivers/net/ethernet/sgi/ |
| H A D | ioc3-eth.c | 84 unsigned long *rxr; /* pointer to receiver ring */ member 371 unsigned long *rxr; in ioc3_rx() local 375 rxr = ip->rxr; /* Ring base */ in ioc3_rx() 396 d = rxr[rx_entry]; in ioc3_rx() 405 dma_unmap_single(ip->dma_dev, rxr[rx_entry], in ioc3_rx() 420 d = rxr[rx_entry]; in ioc3_rx() 430 rxr[n_entry] = cpu_to_be64(ioc3_map(d, PCI64_ATTR_BAR)); in ioc3_rx() 660 be64_to_cpu(ip->rxr[n_entry]), in ioc3_free_rx_bufs() 684 ip->rxr[i] = cpu_to_be64(ioc3_map(d, PCI64_ATTR_BAR)); in ioc3_alloc_rx_bufs() 885 ip->rxr = dma_alloc_coherent(ip->dma_dev, RX_RING_SIZE, &ip->rxr_dma, in ioc3eth_probe() [all …]
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_txrx.h | 312 bool iavf_alloc_rx_buffers(struct iavf_ring *rxr, u16 cleaned_count);
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| H A D | ena_netdev.c | 187 struct ena_ring *txr, *rxr; in ena_init_io_rings() local 194 rxr = &adapter->rx_ring[i]; in ena_init_io_rings() 212 ena_init_io_rings_common(adapter, rxr, i); in ena_init_io_rings() 215 rxr->ring_size = adapter->requested_rx_ring_size; in ena_init_io_rings() 216 rxr->rx_copybreak = adapter->rx_copybreak; in ena_init_io_rings() 217 rxr->sgl_size = adapter->max_rx_sgl_size; in ena_init_io_rings() 218 rxr->smoothed_interval = in ena_init_io_rings() 220 rxr->empty_rx_queue = 0; in ena_init_io_rings() 221 rxr->rx_headroom = NET_SKB_PAD; in ena_init_io_rings() 223 rxr->xdp_ring = &adapter->tx_ring[i + adapter->num_io_queues]; in ena_init_io_rings()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_txrx.h | 461 bool i40e_alloc_rx_buffers(struct i40e_ring *rxr, u16 cleaned_count);
|
| H A D | i40e_main.c | 13372 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_toggle_napi() local 13373 struct i40e_q_vector *q_vector = rxr->q_vector; in i40e_queue_pair_toggle_napi() 13448 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_enable_irq() local 13454 i40e_irq_dynamic_enable(vsi, rxr->q_vector->v_idx); in i40e_queue_pair_enable_irq() 13468 struct i40e_ring *rxr = vsi->rx_rings[queue_pair]; in i40e_queue_pair_disable_irq() local 13479 u32 intpf = vsi->base_vector + rxr->q_vector->v_idx; in i40e_queue_pair_disable_irq()
|