Home
last modified time | relevance | path

Searched refs:next_to_use (Results 1 – 25 of 85) sorted by relevance

1234

/linux/drivers/net/ethernet/mscc/
H A Docelot_fdma.c60 if (rx_ring->next_to_use >= rx_ring->next_to_clean) in ocelot_fdma_rx_ring_free()
62 (rx_ring->next_to_use - rx_ring->next_to_clean) - 1; in ocelot_fdma_rx_ring_free()
64 return rx_ring->next_to_clean - rx_ring->next_to_use - 1; in ocelot_fdma_rx_ring_free()
71 if (tx_ring->next_to_use >= tx_ring->next_to_clean) in ocelot_fdma_tx_ring_free()
73 (tx_ring->next_to_use - tx_ring->next_to_clean) - 1; in ocelot_fdma_tx_ring_free()
75 return tx_ring->next_to_clean - tx_ring->next_to_use - 1; in ocelot_fdma_tx_ring_free()
82 return tx_ring->next_to_clean == tx_ring->next_to_use; in ocelot_fdma_tx_ring_empty()
157 idx = rx_ring->next_to_use; in ocelot_fdma_alloc_rx_buffs()
180 rx_ring->next_to_use = idx; in ocelot_fdma_alloc_rx_buffs()
228 idx = ocelot_fdma_idx_prev(rx_ring->next_to_use, in ocelot_fdma_rx_set_llp()
[all …]
/linux/drivers/net/ethernet/intel/ice/
H A Dice_xsk.c297 u16 ntu = rx_ring->next_to_use; in __ice_alloc_rx_bufs_zc()
325 if (rx_ring->next_to_use != ntu) in __ice_alloc_rx_bufs_zc()
374 last_rs = xdp_ring->next_to_use ? xdp_ring->next_to_use - 1 : cnt - 1; in ice_clean_xdp_irq_zc()
438 u32 ntu = xdp_ring->next_to_use; in ice_xmit_xdp_tx_zc()
491 xdp_ring->next_to_use = ntu; in ice_xmit_xdp_tx_zc()
574 u32 ntu = rx_ring->next_to_use; in ice_clean_rx_irq_zc()
691 if (failure || ntc == rx_ring->next_to_use) in ice_clean_rx_irq_zc()
719 tx_desc = ICE_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in ice_xmit_pkt()
739 u16 ntu = xdp_ring->next_to_use; in ice_xmit_pkt_batch()
758 xdp_ring->next_to_use = ntu; in ice_xmit_pkt_batch()
[all …]
H A Dice_controlq.h18 ((u16)((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \
19 (R)->next_to_clean - (R)->next_to_use - 1))
62 u16 next_to_use; member
H A Dice_txrx.c68 i = tx_ring->next_to_use; in ice_prgm_fdir_fltr()
79 tx_ring->next_to_use = (i < tx_ring->count) ? i : 0; in ice_prgm_fdir_fltr()
103 writel(tx_ring->next_to_use, tx_ring->tail); in ice_prgm_fdir_fltr()
163 tstamp_ring->next_to_use = 0; in ice_clean_tstamp_ring()
228 tx_ring->next_to_use = 0; in ice_clean_tx_ring()
435 tstamp_ring->next_to_use = 0; in ice_setup_tstamp_ring()
500 tx_ring->next_to_use = 0; in ice_setup_tx_ring()
553 for (u32 i = rx_ring->next_to_clean; i != rx_ring->next_to_use; ) { in ice_clean_rx_ring()
580 rx_ring->next_to_use = 0; in ice_clean_rx_ring()
631 rx_ring->next_to_use = 0; in ice_setup_rx_ring()
[all …]
H A Dice_controlq.c364 cq->sq.next_to_use = 0; in ice_init_sq()
424 cq->rq.next_to_use = 0; in ice_init_rq()
977 head, head == cq->sq.next_to_use, in ice_sq_done()
1057 desc_on_ring = ICE_CTL_Q_DESC(cq->sq, cq->sq.next_to_use); in ice_sq_send_cmd()
1064 dma_buf = &cq->sq.r.sq_bi[cq->sq.next_to_use]; in ice_sq_send_cmd()
1083 (cq->sq.next_to_use)++; in ice_sq_send_cmd()
1084 if (cq->sq.next_to_use == cq->sq.count) in ice_sq_send_cmd()
1085 cq->sq.next_to_use = 0; in ice_sq_send_cmd()
1086 wr32(hw, cq->sq.tail, cq->sq.next_to_use); in ice_sq_send_cmd()
1248 cq->rq.next_to_use = ntu; in ice_clean_rq_elem()
/linux/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_xsk.c155 u16 i = rx_ring->next_to_use; in ixgbe_alloc_rx_buffers_zc()
198 if (rx_ring->next_to_use != i) { in ixgbe_alloc_rx_buffers_zc()
199 rx_ring->next_to_use = i; in ixgbe_alloc_rx_buffers_zc()
365 if (failure || rx_ring->next_to_clean == rx_ring->next_to_use) in ixgbe_clean_rx_irq_zc()
416 tx_bi = &xdp_ring->tx_buffer_info[xdp_ring->next_to_use]; in ixgbe_xmit_zc()
421 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc()
433 xdp_ring->next_to_use++; in ixgbe_xmit_zc()
434 if (xdp_ring->next_to_use == xdp_ring->count) in ixgbe_xmit_zc()
435 xdp_ring->next_to_use = 0; in ixgbe_xmit_zc()
461 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_clean_xdp_tx_irq()
[all …]
/linux/drivers/net/ethernet/intel/igb/
H A Digb_xsk.c205 u16 ntu = rx_ring->next_to_use; in igb_alloc_rx_buffers_zc()
235 if (rx_ring->next_to_use != ntu) { in igb_alloc_rx_buffers_zc()
236 rx_ring->next_to_use = ntu; in igb_alloc_rx_buffers_zc()
253 u16 ntu = rx_ring->next_to_use; in igb_clean_rx_ring_zc()
455 if (failure || rx_ring->next_to_clean == rx_ring->next_to_use) in igb_clean_rx_irq_zc()
489 tx_buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_use]; in igb_xmit_zc()
496 tx_desc = IGB_TX_DESC(tx_ring, tx_ring->next_to_use); in igb_xmit_zc()
515 tx_ring->next_to_use++; in igb_xmit_zc()
517 if (tx_ring->next_to_use == tx_ring->count) in igb_xmit_zc()
518 tx_ring->next_to_use = 0; in igb_xmit_zc()
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_xsk.c243 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers_zc()
498 if (failure || next_to_clean == rx_ring->next_to_use) in i40e_clean_rx_irq_zc()
518 tx_desc = I40E_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in i40e_xmit_pkt()
528 u16 ntu = xdp_ring->next_to_use; in i40e_xmit_pkt_batch()
547 xdp_ring->next_to_use = ntu; in i40e_xmit_pkt_batch()
565 u16 ntu = xdp_ring->next_to_use ? xdp_ring->next_to_use - 1 : xdp_ring->count - 1; in i40e_set_rs_bit()
589 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in i40e_xmit_zc()
590 nb_processed = xdp_ring->count - xdp_ring->next_to_use; in i40e_xmit_zc()
592 xdp_ring->next_to_use = 0; in i40e_xmit_zc()
730 u16 ntu = rx_ring->next_to_use; in i40e_xsk_clean_rx_ring()
[all …]
H A Di40e_adminq.c315 hw->aq.asq.next_to_use = 0; in i40e_init_asq()
374 hw->aq.arq.next_to_use = 0; in i40e_init_arq()
735 return rd32(hw, I40E_PF_ATQH) == hw->aq.asq.next_to_use; in i40e_asq_done()
784 details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec()
839 desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec()
846 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); in i40e_asq_send_command_atomic_exec()
864 (hw->aq.asq.next_to_use)++; in i40e_asq_send_command_atomic_exec()
865 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in i40e_asq_send_command_atomic_exec()
866 hw->aq.asq.next_to_use = 0; in i40e_asq_send_command_atomic_exec()
868 wr32(hw, I40E_PF_ATQT, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec()
[all …]
H A Di40e_txrx.c32 i = tx_ring->next_to_use; in i40e_fdir()
36 tx_ring->next_to_use = (i < tx_ring->count) ? i : 0; in i40e_fdir()
116 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
121 i = tx_ring->next_to_use; in i40e_program_fdir_filter()
125 tx_ring->next_to_use = ((i + 1) < tx_ring->count) ? i + 1 : 0; in i40e_program_fdir_filter()
150 writel(tx_ring->next_to_use, tx_ring->tail); in i40e_program_fdir_filter()
807 tx_ring->next_to_use = 0; in i40e_clean_tx_ring()
853 tail = ring->next_to_use; in i40e_get_tx_pending()
1445 tx_ring->next_to_use = 0; in i40e_setup_tx_descriptors()
1521 rx_ring->next_to_use = 0; in i40e_clean_rx_ring()
[all …]
/linux/drivers/net/ethernet/hisilicon/hns3/hns3_common/
H A Dhclge_comm_cmd.c289 int ntu = ring->next_to_use; in hclge_comm_ring_space()
302 desc_to_use = &hw->cmq.csq.desc[hw->cmq.csq.next_to_use]; in hclge_comm_cmd_copy_desc()
304 (hw->cmq.csq.next_to_use)++; in hclge_comm_cmd_copy_desc()
305 if (hw->cmq.csq.next_to_use >= hw->cmq.csq.desc_num) in hclge_comm_cmd_copy_desc()
306 hw->cmq.csq.next_to_use = 0; in hclge_comm_cmd_copy_desc()
315 int ntu = ring->next_to_use; in hclge_comm_is_valid_csq_clean_head()
334 head, csq->next_to_use, csq->next_to_clean); in hclge_comm_cmd_csq_clean()
351 return head == (u32)hw->cmq.csq.next_to_use; in hclge_comm_cmd_csq_done()
505 ntc = hw->cmq.csq.next_to_use; in hclge_comm_cmd_send()
511 hw->cmq.csq.next_to_use); in hclge_comm_cmd_send()
[all …]
/linux/drivers/net/ethernet/atheros/atlx/
H A Datl1.c1112 atomic_set(&tpd_ring->next_to_use, 0); in atl1_init_ring_ptrs()
1116 atomic_set(&rfd_ring->next_to_use, 0); in atl1_init_ring_ptrs()
1118 rrd_ring->next_to_use = 0; in atl1_init_ring_ptrs()
1156 atomic_set(&rfd_ring->next_to_use, 0); in atl1_clean_rx_ring()
1158 rrd_ring->next_to_use = 0; in atl1_clean_rx_ring()
1198 atomic_set(&tpd_ring->next_to_use, 0); in atl1_clean_tx_ring()
1497 value = ((atomic_read(&adapter->tpd_ring.next_to_use) in atl1_configure()
1501 ((atomic_read(&adapter->rfd_ring.next_to_use) in atl1_configure()
1743 tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); in atl1_update_mailbox()
1744 rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use); in atl1_update_mailbox()
[all …]
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_adminq.c321 hw->aq.asq.next_to_use = 0; in iavf_init_asq()
386 hw->aq.arq.next_to_use = 0; in iavf_init_arq()
603 return rd32(hw, IAVF_VF_ATQH1) == hw->aq.asq.next_to_use; in iavf_asq_done()
650 details = IAVF_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command()
705 desc_on_ring = IAVF_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command()
712 dma_buff = &hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]; in iavf_asq_send_command()
730 (hw->aq.asq.next_to_use)++; in iavf_asq_send_command()
731 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in iavf_asq_send_command()
732 hw->aq.asq.next_to_use = 0; in iavf_asq_send_command()
734 wr32(hw, IAVF_VF_ATQT1, hw->aq.asq.next_to_use); in iavf_asq_send_command()
[all …]
H A Diavf_txrx.c98 tx_ring->next_to_use = 0; in iavf_clean_tx_ring()
143 tail = ring->next_to_use; in iavf_get_tx_pending()
696 tx_ring->next_to_use = 0; in iavf_setup_tx_descriptors()
723 for (u32 i = rx_ring->next_to_clean; i != rx_ring->next_to_use; ) { in iavf_clean_rx_ring()
733 rx_ring->next_to_use = 0; in iavf_clean_rx_ring()
801 rx_ring->next_to_use = 0; in iavf_setup_rx_descriptors()
820 rx_ring->next_to_use = val; in iavf_release_rx_desc()
867 u16 ntu = rx_ring->next_to_use; in iavf_alloc_rx_buffers()
901 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers()
907 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers()
[all …]
/linux/drivers/net/ethernet/hisilicon/hns3/hns3vf/
H A Dhclgevf_mbx.c162 return tail == (u32)hw->hw.cmq.crq.next_to_use; in hclgevf_cmd_crq_empty()
241 desc = &crq->desc[crq->next_to_use]; in hclgevf_mbx_handler()
244 flag = le16_to_cpu(crq->desc[crq->next_to_use].flag); in hclgevf_mbx_handler()
252 crq->desc[crq->next_to_use].flag = 0; in hclgevf_mbx_handler()
282 crq->desc[crq->next_to_use].flag = 0; in hclgevf_mbx_handler()
288 crq->next_to_use); in hclgevf_mbx_handler()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.h123 ((((txq)->next_to_clean > (txq)->next_to_use) ? 0 : (txq)->desc_count) + \
124 (txq)->next_to_clean - (txq)->next_to_use - 1)
556 u32 next_to_use; member
677 u32 next_to_use; member
779 u32 next_to_use; member
842 u32 next_to_use; member
879 u32 next_to_use; member
1072 return (refillq->next_to_use > refillq->next_to_clean ? in idpf_tx_splitq_get_free_bufs()
1074 refillq->next_to_use - refillq->next_to_clean - 1; in idpf_tx_splitq_get_free_bufs()
H A Didpf_controlq.c139 cq->next_to_use = 0; in idpf_ctlq_add()
284 desc = IDPF_CTLQ_DESC(cq, cq->next_to_use); in idpf_ctlq_send()
317 cq->bi.tx_msg[cq->next_to_use] = msg; in idpf_ctlq_send()
319 (cq->next_to_use)++; in idpf_ctlq_send()
320 if (cq->next_to_use == cq->ring_size) in idpf_ctlq_send()
321 cq->next_to_use = 0; in idpf_ctlq_send()
329 idpf_mbx_wr32(hw, cq->reg.tail, cq->next_to_use); in idpf_ctlq_send()
/linux/drivers/net/ethernet/intel/igc/
H A Digc_dump.c142 n, tx_ring->next_to_use, tx_ring->next_to_clean, in igc_rings_dump()
181 if (i == tx_ring->next_to_use && in igc_rings_dump()
184 else if (i == tx_ring->next_to_use) in igc_rings_dump()
215 netdev_info(netdev, "%5d %5X %5X\n", n, rx_ring->next_to_use, in igc_rings_dump()
264 if (i == rx_ring->next_to_use) in igc_rings_dump()
/linux/drivers/net/ethernet/hisilicon/hns/
H A Dhns_enet.c39 struct hnae_desc *desc = &ring->desc[ring->next_to_use]; in fill_v2_desc_hw()
40 struct hnae_desc_cb *desc_cb = &ring->desc_cb[ring->next_to_use]; in fill_v2_desc_hw()
125 ring_ptr_move_fw(ring, next_to_use); in fill_v2_desc_hw()
148 struct hnae_desc *desc = &ring->desc[ring->next_to_use]; in fill_desc()
149 struct hnae_desc_cb *desc_cb = &ring->desc_cb[ring->next_to_use]; in fill_desc()
202 ring_ptr_move_fw(ring, next_to_use); in fill_desc()
207 ring_ptr_move_bw(ring, next_to_use); in unfill_desc()
338 int size, next_to_use; in hns_nic_net_xmit_hw() local
356 next_to_use = ring->next_to_use; in hns_nic_net_xmit_hw()
401 while (ring->next_to_use != next_to_use) { in hns_nic_net_xmit_hw()
[all …]
/linux/drivers/net/ethernet/freescale/enetc/
H A Denetc_cbdr.c24 cbdr->next_to_use = 0; in enetc_setup_cbdr()
41 enetc_wr_reg(cbdr->cir, cbdr->next_to_use); in enetc_setup_cbdr()
121 return (r->next_to_clean - r->next_to_use - 1 + r->bd_count) % in enetc_cbd_unused()
138 i = ring->next_to_use; in enetc_send_cmd()
145 ring->next_to_use = i; in enetc_send_cmd()
/linux/drivers/net/ethernet/atheros/atl1e/
H A Datl1e_main.c751 tx_ring->next_to_use = 0; in atl1e_init_ring_ptrs()
1562 u16 next_to_use = 0; in atl1e_tpd_avail() local
1566 next_to_use = tx_ring->next_to_use; in atl1e_tpd_avail()
1568 return (u16)(next_to_clean > next_to_use) ? in atl1e_tpd_avail()
1569 (next_to_clean - next_to_use - 1) : in atl1e_tpd_avail()
1570 (tx_ring->count + next_to_clean - next_to_use - 1); in atl1e_tpd_avail()
1581 u16 next_to_use = 0; in atl1e_get_tpd() local
1583 next_to_use = tx_ring->next_to_use; in atl1e_get_tpd()
1584 if (++tx_ring->next_to_use == tx_ring->count) in atl1e_get_tpd()
1585 tx_ring->next_to_use = 0; in atl1e_get_tpd()
[all …]
/linux/drivers/net/ethernet/microchip/lan966x/
H A Dlan966x_fdma.c607 int next_to_use; in lan966x_fdma_xmit_xdpf() local
614 next_to_use = lan966x_fdma_get_next_dcb(tx); in lan966x_fdma_xmit_xdpf()
615 if (next_to_use < 0) { in lan966x_fdma_xmit_xdpf()
622 next_dcb_buf = &tx->dcbs_buf[next_to_use]; in lan966x_fdma_xmit_xdpf()
676 next_to_use, in lan966x_fdma_xmit_xdpf()
704 int next_to_use; in lan966x_fdma_xmit() local
708 next_to_use = lan966x_fdma_get_next_dcb(tx); in lan966x_fdma_xmit()
709 if (next_to_use < 0) { in lan966x_fdma_xmit()
746 next_dcb_buf = &tx->dcbs_buf[next_to_use]; in lan966x_fdma_xmit()
757 next_to_use, in lan966x_fdma_xmit()
/linux/drivers/net/ethernet/amazon/ena/
H A Dena_xdp.c82 u16 next_to_use, req_id; in ena_xdp_xmit_frame() local
85 next_to_use = tx_ring->next_to_use; in ena_xdp_xmit_frame()
86 req_id = tx_ring->free_ids[next_to_use]; in ena_xdp_xmit_frame()
100 next_to_use, in ena_xdp_xmit_frame()
/linux/drivers/net/ethernet/atheros/atl1c/
H A Datl1c_main.c895 tpd_ring->next_to_use = 0; in atl1c_clean_tx_ring()
918 rfd_ring->next_to_use = 0; in atl1c_clean_rx_ring()
919 rrd_ring->next_to_use = 0; in atl1c_clean_rx_ring()
935 tpd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs()
943 rfd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs()
945 rrd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs()
1760 next_next = rfd_next_to_use = rfd_ring->next_to_use; in atl1c_alloc_rx_buffer()
1824 rfd_ring->next_to_use = rfd_next_to_use; in atl1c_alloc_rx_buffer()
1826 rfd_ring->next_to_use & MB_RFDX_PROD_IDX_MASK); in atl1c_alloc_rx_buffer()
1982 u16 next_to_use = 0; in atl1c_tpd_avail() local
[all …]
/linux/drivers/net/ethernet/hisilicon/hns3/
H A Dhns3_enet.c980 ntu = tx_spare->next_to_use; in hns3_tx_spare_space()
1113 u32 ntu = tx_spare->next_to_use; in hns3_tx_spare_alloc()
1126 tx_spare->next_to_use = ntu + size; in hns3_tx_spare_alloc()
1127 if (tx_spare->next_to_use == tx_spare->len) in hns3_tx_spare_alloc()
1128 tx_spare->next_to_use = 0; in hns3_tx_spare_alloc()
1139 if (len > tx_spare->next_to_use) { in hns3_tx_spare_rollback()
1140 len -= tx_spare->next_to_use; in hns3_tx_spare_rollback()
1141 tx_spare->next_to_use = tx_spare->len - len; in hns3_tx_spare_rollback()
1143 tx_spare->next_to_use -= len; in hns3_tx_spare_rollback()
1704 struct hns3_desc *desc = &ring->desc[ring->next_to_use]; in hns3_fill_desc()
[all …]

1234