| /linux/drivers/net/ethernet/mscc/ |
| H A D | ocelot_fdma.c | 60 if (rx_ring->next_to_use >= rx_ring->next_to_clean) in ocelot_fdma_rx_ring_free() 62 (rx_ring->next_to_use - rx_ring->next_to_clean) - 1; in ocelot_fdma_rx_ring_free() 64 return rx_ring->next_to_clean - rx_ring->next_to_use - 1; in ocelot_fdma_rx_ring_free() 71 if (tx_ring->next_to_use >= tx_ring->next_to_clean) in ocelot_fdma_tx_ring_free() 73 (tx_ring->next_to_use - tx_ring->next_to_clean) - 1; in ocelot_fdma_tx_ring_free() 75 return tx_ring->next_to_clean - tx_ring->next_to_use - 1; in ocelot_fdma_tx_ring_free() 82 return tx_ring->next_to_clean == tx_ring->next_to_use; in ocelot_fdma_tx_ring_empty() 157 idx = rx_ring->next_to_use; in ocelot_fdma_alloc_rx_buffs() 180 rx_ring->next_to_use = idx; in ocelot_fdma_alloc_rx_buffs() 228 idx = ocelot_fdma_idx_prev(rx_ring->next_to_use, in ocelot_fdma_rx_set_llp() [all …]
|
| H A D | ocelot_fdma.h | 98 u16 next_to_use; member 130 u16 next_to_use; member
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_xsk.c | 155 u16 i = rx_ring->next_to_use; in ixgbe_alloc_rx_buffers_zc() 198 if (rx_ring->next_to_use != i) { in ixgbe_alloc_rx_buffers_zc() 199 rx_ring->next_to_use = i; in ixgbe_alloc_rx_buffers_zc() 365 if (failure || rx_ring->next_to_clean == rx_ring->next_to_use) in ixgbe_clean_rx_irq_zc() 416 tx_bi = &xdp_ring->tx_buffer_info[xdp_ring->next_to_use]; in ixgbe_xmit_zc() 421 tx_desc = IXGBE_TX_DESC(xdp_ring, xdp_ring->next_to_use); in ixgbe_xmit_zc() 433 xdp_ring->next_to_use++; in ixgbe_xmit_zc() 434 if (xdp_ring->next_to_use == xdp_ring->count) in ixgbe_xmit_zc() 435 xdp_ring->next_to_use = 0; in ixgbe_xmit_zc() 461 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_clean_xdp_tx_irq() [all …]
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 205 u16 ntu = rx_ring->next_to_use; in igb_alloc_rx_buffers_zc() 235 if (rx_ring->next_to_use != ntu) { in igb_alloc_rx_buffers_zc() 236 rx_ring->next_to_use = ntu; in igb_alloc_rx_buffers_zc() 253 u16 ntu = rx_ring->next_to_use; in igb_clean_rx_ring_zc() 455 if (failure || rx_ring->next_to_clean == rx_ring->next_to_use) in igb_clean_rx_irq_zc() 489 tx_buffer_info = &tx_ring->tx_buffer_info[tx_ring->next_to_use]; in igb_xmit_zc() 496 tx_desc = IGB_TX_DESC(tx_ring, tx_ring->next_to_use); in igb_xmit_zc() 515 tx_ring->next_to_use++; in igb_xmit_zc() 517 if (tx_ring->next_to_use == tx_ring->count) in igb_xmit_zc() 518 tx_ring->next_to_use = 0; in igb_xmit_zc()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 243 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers_zc() 498 if (failure || next_to_clean == rx_ring->next_to_use) in i40e_clean_rx_irq_zc() 518 tx_desc = I40E_TX_DESC(xdp_ring, xdp_ring->next_to_use++); in i40e_xmit_pkt() 528 u16 ntu = xdp_ring->next_to_use; in i40e_xmit_pkt_batch() 547 xdp_ring->next_to_use = ntu; in i40e_xmit_pkt_batch() 565 u16 ntu = xdp_ring->next_to_use ? xdp_ring->next_to_use - 1 : xdp_ring->count - 1; in i40e_set_rs_bit() 589 if (xdp_ring->next_to_use + nb_pkts >= xdp_ring->count) { in i40e_xmit_zc() 590 nb_processed = xdp_ring->count - xdp_ring->next_to_use; in i40e_xmit_zc() 592 xdp_ring->next_to_use = 0; in i40e_xmit_zc() 730 u16 ntu = rx_ring->next_to_use; in i40e_xsk_clean_rx_ring() [all …]
|
| H A D | i40e_adminq.c | 315 hw->aq.asq.next_to_use = 0; in i40e_init_asq() 374 hw->aq.arq.next_to_use = 0; in i40e_init_arq() 735 return rd32(hw, I40E_PF_ATQH) == hw->aq.asq.next_to_use; in i40e_asq_done() 784 details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec() 839 desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec() 846 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]); in i40e_asq_send_command_atomic_exec() 864 (hw->aq.asq.next_to_use)++; in i40e_asq_send_command_atomic_exec() 865 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in i40e_asq_send_command_atomic_exec() 866 hw->aq.asq.next_to_use = 0; in i40e_asq_send_command_atomic_exec() 868 wr32(hw, I40E_PF_ATQT, hw->aq.asq.next_to_use); in i40e_asq_send_command_atomic_exec() [all …]
|
| H A D | i40e_txrx.c | 32 i = tx_ring->next_to_use; in i40e_fdir() 36 tx_ring->next_to_use = (i < tx_ring->count) ? i : 0; in i40e_fdir() 116 i = tx_ring->next_to_use; in i40e_program_fdir_filter() 121 i = tx_ring->next_to_use; in i40e_program_fdir_filter() 125 tx_ring->next_to_use = ((i + 1) < tx_ring->count) ? i + 1 : 0; in i40e_program_fdir_filter() 150 writel(tx_ring->next_to_use, tx_ring->tail); in i40e_program_fdir_filter() 807 tx_ring->next_to_use = 0; in i40e_clean_tx_ring() 853 tail = ring->next_to_use; in i40e_get_tx_pending() 1445 tx_ring->next_to_use = 0; in i40e_setup_tx_descriptors() 1518 rx_ring->next_to_use = 0; in i40e_clean_rx_ring() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hns3/hns3_common/ |
| H A D | hclge_comm_cmd.c | 289 int ntu = ring->next_to_use; in hclge_comm_ring_space() 302 desc_to_use = &hw->cmq.csq.desc[hw->cmq.csq.next_to_use]; in hclge_comm_cmd_copy_desc() 304 (hw->cmq.csq.next_to_use)++; in hclge_comm_cmd_copy_desc() 305 if (hw->cmq.csq.next_to_use >= hw->cmq.csq.desc_num) in hclge_comm_cmd_copy_desc() 306 hw->cmq.csq.next_to_use = 0; in hclge_comm_cmd_copy_desc() 315 int ntu = ring->next_to_use; in hclge_comm_is_valid_csq_clean_head() 334 head, csq->next_to_use, csq->next_to_clean); in hclge_comm_cmd_csq_clean() 351 return head == (u32)hw->cmq.csq.next_to_use; in hclge_comm_cmd_csq_done() 505 ntc = hw->cmq.csq.next_to_use; in hclge_comm_cmd_send() 511 hw->cmq.csq.next_to_use); in hclge_comm_cmd_send() [all …]
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_controlq.h | 18 ((u16)((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->count) + \ 19 (R)->next_to_clean - (R)->next_to_use - 1)) 62 u16 next_to_use; member
|
| H A D | ice_controlq.c | 364 cq->sq.next_to_use = 0; in ice_init_sq() 424 cq->rq.next_to_use = 0; in ice_init_rq() 977 head, head == cq->sq.next_to_use, in ice_sq_done() 1057 desc_on_ring = ICE_CTL_Q_DESC(cq->sq, cq->sq.next_to_use); in ice_sq_send_cmd() 1064 dma_buf = &cq->sq.r.sq_bi[cq->sq.next_to_use]; in ice_sq_send_cmd() 1083 (cq->sq.next_to_use)++; in ice_sq_send_cmd() 1084 if (cq->sq.next_to_use == cq->sq.count) in ice_sq_send_cmd() 1085 cq->sq.next_to_use = 0; in ice_sq_send_cmd() 1086 wr32(hw, cq->sq.tail, cq->sq.next_to_use); in ice_sq_send_cmd() 1248 cq->rq.next_to_use = ntu; in ice_clean_rq_elem()
|
| /linux/drivers/net/ethernet/atheros/atlx/ |
| H A D | atl1.c | 1112 atomic_set(&tpd_ring->next_to_use, 0); in atl1_init_ring_ptrs() 1116 atomic_set(&rfd_ring->next_to_use, 0); in atl1_init_ring_ptrs() 1118 rrd_ring->next_to_use = 0; in atl1_init_ring_ptrs() 1156 atomic_set(&rfd_ring->next_to_use, 0); in atl1_clean_rx_ring() 1158 rrd_ring->next_to_use = 0; in atl1_clean_rx_ring() 1198 atomic_set(&tpd_ring->next_to_use, 0); in atl1_clean_tx_ring() 1497 value = ((atomic_read(&adapter->tpd_ring.next_to_use) in atl1_configure() 1501 ((atomic_read(&adapter->rfd_ring.next_to_use) in atl1_configure() 1743 tpd_next_to_use = atomic_read(&adapter->tpd_ring.next_to_use); in atl1_update_mailbox() 1744 rfd_next_to_use = atomic_read(&adapter->rfd_ring.next_to_use); in atl1_update_mailbox() [all …]
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_adminq.c | 321 hw->aq.asq.next_to_use = 0; in iavf_init_asq() 386 hw->aq.arq.next_to_use = 0; in iavf_init_arq() 603 return rd32(hw, IAVF_VF_ATQH1) == hw->aq.asq.next_to_use; in iavf_asq_done() 650 details = IAVF_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command() 705 desc_on_ring = IAVF_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use); in iavf_asq_send_command() 712 dma_buff = &hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]; in iavf_asq_send_command() 730 (hw->aq.asq.next_to_use)++; in iavf_asq_send_command() 731 if (hw->aq.asq.next_to_use == hw->aq.asq.count) in iavf_asq_send_command() 732 hw->aq.asq.next_to_use = 0; in iavf_asq_send_command() 734 wr32(hw, IAVF_VF_ATQT1, hw->aq.asq.next_to_use); in iavf_asq_send_command() [all …]
|
| H A D | iavf_txrx.c | 98 tx_ring->next_to_use = 0; in iavf_clean_tx_ring() 143 tail = ring->next_to_use; in iavf_get_tx_pending() 696 tx_ring->next_to_use = 0; in iavf_setup_tx_descriptors() 723 for (u32 i = rx_ring->next_to_clean; i != rx_ring->next_to_use; ) { in iavf_clean_rx_ring() 733 rx_ring->next_to_use = 0; in iavf_clean_rx_ring() 801 rx_ring->next_to_use = 0; in iavf_setup_rx_descriptors() 820 rx_ring->next_to_use = val; in iavf_release_rx_desc() 867 u16 ntu = rx_ring->next_to_use; in iavf_alloc_rx_buffers() 901 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers() 907 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hns3/hns3vf/ |
| H A D | hclgevf_mbx.c | 162 return tail == (u32)hw->hw.cmq.crq.next_to_use; in hclgevf_cmd_crq_empty() 241 desc = &crq->desc[crq->next_to_use]; in hclgevf_mbx_handler() 244 flag = le16_to_cpu(crq->desc[crq->next_to_use].flag); in hclgevf_mbx_handler() 252 crq->desc[crq->next_to_use].flag = 0; in hclgevf_mbx_handler() 282 crq->desc[crq->next_to_use].flag = 0; in hclgevf_mbx_handler() 288 crq->next_to_use); in hclgevf_mbx_handler()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.h | 122 ((((txq)->next_to_clean > (txq)->next_to_use) ? 0 : (txq)->desc_count) + \ 123 (txq)->next_to_clean - (txq)->next_to_use - 1) 553 u32 next_to_use; member 674 u32 next_to_use; member 776 u32 next_to_use; member 839 u32 next_to_use; member 876 u32 next_to_use; member 1067 return (refillq->next_to_use > refillq->next_to_clean ? in idpf_tx_splitq_get_free_bufs() 1069 refillq->next_to_use - refillq->next_to_clean - 1; in idpf_tx_splitq_get_free_bufs()
|
| H A D | idpf_controlq.c | 139 cq->next_to_use = 0; in idpf_ctlq_add() 286 desc = IDPF_CTLQ_DESC(cq, cq->next_to_use); in idpf_ctlq_send() 319 cq->bi.tx_msg[cq->next_to_use] = msg; in idpf_ctlq_send() 321 (cq->next_to_use)++; in idpf_ctlq_send() 322 if (cq->next_to_use == cq->ring_size) in idpf_ctlq_send() 323 cq->next_to_use = 0; in idpf_ctlq_send() 331 idpf_mbx_wr32(hw, cq->reg.tail, cq->next_to_use); in idpf_ctlq_send()
|
| H A D | idpf_controlq.h | 18 ((u16)((((R)->next_to_clean > (R)->next_to_use) ? 0 : (R)->ring_size) + \ 19 (R)->next_to_clean - (R)->next_to_use - 1))
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc_dump.c | 142 n, tx_ring->next_to_use, tx_ring->next_to_clean, in igc_rings_dump() 181 if (i == tx_ring->next_to_use && in igc_rings_dump() 184 else if (i == tx_ring->next_to_use) in igc_rings_dump() 215 netdev_info(netdev, "%5d %5X %5X\n", n, rx_ring->next_to_use, in igc_rings_dump() 264 if (i == rx_ring->next_to_use) in igc_rings_dump()
|
| /linux/drivers/net/ethernet/hisilicon/hns/ |
| H A D | hns_enet.c | 39 struct hnae_desc *desc = &ring->desc[ring->next_to_use]; in fill_v2_desc_hw() 40 struct hnae_desc_cb *desc_cb = &ring->desc_cb[ring->next_to_use]; in fill_v2_desc_hw() 125 ring_ptr_move_fw(ring, next_to_use); in fill_v2_desc_hw() 148 struct hnae_desc *desc = &ring->desc[ring->next_to_use]; in fill_desc() 149 struct hnae_desc_cb *desc_cb = &ring->desc_cb[ring->next_to_use]; in fill_desc() 202 ring_ptr_move_fw(ring, next_to_use); in fill_desc() 207 ring_ptr_move_bw(ring, next_to_use); in unfill_desc() 338 int size, next_to_use; in hns_nic_net_xmit_hw() local 356 next_to_use = ring->next_to_use; in hns_nic_net_xmit_hw() 401 while (ring->next_to_use != next_to_use) { in hns_nic_net_xmit_hw() [all …]
|
| /linux/drivers/net/ethernet/atheros/atl1e/ |
| H A D | atl1e_main.c | 751 tx_ring->next_to_use = 0; in atl1e_init_ring_ptrs() 1562 u16 next_to_use = 0; in atl1e_tpd_avail() local 1566 next_to_use = tx_ring->next_to_use; in atl1e_tpd_avail() 1568 return (u16)(next_to_clean > next_to_use) ? in atl1e_tpd_avail() 1569 (next_to_clean - next_to_use - 1) : in atl1e_tpd_avail() 1570 (tx_ring->count + next_to_clean - next_to_use - 1); in atl1e_tpd_avail() 1581 u16 next_to_use = 0; in atl1e_get_tpd() local 1583 next_to_use = tx_ring->next_to_use; in atl1e_get_tpd() 1584 if (++tx_ring->next_to_use == tx_ring->count) in atl1e_get_tpd() 1585 tx_ring->next_to_use = 0; in atl1e_get_tpd() [all …]
|
| /linux/drivers/net/ethernet/microchip/lan966x/ |
| H A D | lan966x_fdma.c | 608 int next_to_use; in lan966x_fdma_xmit_xdpf() local 615 next_to_use = lan966x_fdma_get_next_dcb(tx); in lan966x_fdma_xmit_xdpf() 616 if (next_to_use < 0) { in lan966x_fdma_xmit_xdpf() 623 next_dcb_buf = &tx->dcbs_buf[next_to_use]; in lan966x_fdma_xmit_xdpf() 677 next_to_use, in lan966x_fdma_xmit_xdpf() 705 int next_to_use; in lan966x_fdma_xmit() local 709 next_to_use = lan966x_fdma_get_next_dcb(tx); in lan966x_fdma_xmit() 710 if (next_to_use < 0) { in lan966x_fdma_xmit() 747 next_dcb_buf = &tx->dcbs_buf[next_to_use]; in lan966x_fdma_xmit() 758 next_to_use, in lan966x_fdma_xmit()
|
| /linux/drivers/net/ethernet/amazon/ena/ |
| H A D | ena_xdp.c | 82 u16 next_to_use, req_id; in ena_xdp_xmit_frame() local 85 next_to_use = tx_ring->next_to_use; in ena_xdp_xmit_frame() 86 req_id = tx_ring->free_ids[next_to_use]; in ena_xdp_xmit_frame() 100 next_to_use, in ena_xdp_xmit_frame()
|
| /linux/drivers/net/ethernet/atheros/atl1c/ |
| H A D | atl1c_main.c | 895 tpd_ring->next_to_use = 0; in atl1c_clean_tx_ring() 918 rfd_ring->next_to_use = 0; in atl1c_clean_rx_ring() 919 rrd_ring->next_to_use = 0; in atl1c_clean_rx_ring() 935 tpd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs() 943 rfd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs() 945 rrd_ring[i].next_to_use = 0; in atl1c_init_ring_ptrs() 1760 next_next = rfd_next_to_use = rfd_ring->next_to_use; in atl1c_alloc_rx_buffer() 1824 rfd_ring->next_to_use = rfd_next_to_use; in atl1c_alloc_rx_buffer() 1826 rfd_ring->next_to_use & MB_RFDX_PROD_IDX_MASK); in atl1c_alloc_rx_buffer() 1982 u16 next_to_use = 0; in atl1c_tpd_avail() local [all …]
|
| /linux/drivers/net/ethernet/freescale/enetc/ |
| H A D | enetc.h | 159 int next_to_use; member 193 if (bdr->next_to_clean > bdr->next_to_use) in enetc_bd_unused() 194 return bdr->next_to_clean - bdr->next_to_use - 1; in enetc_bd_unused() 196 return bdr->bd_count + bdr->next_to_clean - bdr->next_to_use - 1; in enetc_bd_unused() 216 int next_to_use; member
|
| /linux/drivers/net/ethernet/hisilicon/hns3/ |
| H A D | hclge_mbx.h | 248 (crq->next_to_use = (crq->next_to_use + 1) % crq->desc_num)
|