Home
last modified time | relevance | path

Searched refs:ntu (Results 1 – 25 of 26) sorted by relevance

12

/linux/drivers/net/ethernet/intel/ice/
H A Dice_xsk.c297 u16 ntu = rx_ring->next_to_use; in __ice_alloc_rx_bufs_zc() local
301 rx_desc = ICE_RX_DESC(rx_ring, ntu); in __ice_alloc_rx_bufs_zc()
302 xdp = ice_xdp_buf(rx_ring, ntu); in __ice_alloc_rx_bufs_zc()
304 if (ntu + count >= rx_ring->count) { in __ice_alloc_rx_bufs_zc()
306 rx_ring->count - ntu); in __ice_alloc_rx_bufs_zc()
307 if (nb_buffs_extra != rx_ring->count - ntu) { in __ice_alloc_rx_bufs_zc()
308 ntu += nb_buffs_extra; in __ice_alloc_rx_bufs_zc()
313 ntu = 0; in __ice_alloc_rx_bufs_zc()
320 ntu += nb_buffs; in __ice_alloc_rx_bufs_zc()
321 if (ntu == rx_ring->count) in __ice_alloc_rx_bufs_zc()
[all …]
H A Dice_txrx.c788 u32 ntu = rx_ring->next_to_use; in ice_init_ctrl_rx_descs() local
793 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_init_ctrl_rx_descs()
797 ntu++; in ice_init_ctrl_rx_descs()
798 if (unlikely(ntu == rx_ring->count)) { in ice_init_ctrl_rx_descs()
800 ntu = 0; in ice_init_ctrl_rx_descs()
807 if (rx_ring->next_to_use != ntu) in ice_init_ctrl_rx_descs()
808 ice_release_rx_desc(rx_ring, ntu); in ice_init_ctrl_rx_descs()
839 u16 ntu = rx_ring->next_to_use; in ice_alloc_rx_bufs() local
846 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_alloc_rx_bufs()
851 addr = libeth_rx_alloc(&fq, ntu); in ice_alloc_rx_bufs()
[all …]
H A Dice_controlq.c1184 u16 ntu; in ice_clean_rq_elem() local
1199 ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask); in ice_clean_rq_elem()
1201 if (ntu == ntc) { in ice_clean_rq_elem()
1248 cq->rq.next_to_use = ntu; in ice_clean_rq_elem()
1254 ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask); in ice_clean_rq_elem()
1255 *pending = (u16)((ntc > ntu ? cq->rq.count : 0) + (ntu - ntc)); in ice_clean_rq_elem()
/linux/drivers/net/ethernet/intel/igb/
H A Digb_xsk.c205 u16 ntu = rx_ring->next_to_use; in igb_alloc_rx_buffers_zc() local
209 rx_desc = IGB_RX_DESC(rx_ring, ntu); in igb_alloc_rx_buffers_zc()
210 xdp = &rx_ring->rx_buffer_info_zc[ntu]; in igb_alloc_rx_buffers_zc()
212 if (ntu + count >= rx_ring->count) { in igb_alloc_rx_buffers_zc()
214 rx_ring->count - ntu); in igb_alloc_rx_buffers_zc()
215 if (nb_buffs_extra != rx_ring->count - ntu) { in igb_alloc_rx_buffers_zc()
216 ntu += nb_buffs_extra; in igb_alloc_rx_buffers_zc()
221 ntu = 0; in igb_alloc_rx_buffers_zc()
226 ntu += nb_buffs; in igb_alloc_rx_buffers_zc()
227 if (ntu == rx_ring->count) in igb_alloc_rx_buffers_zc()
[all …]
/linux/drivers/net/ethernet/intel/i40e/
H A Di40e_xsk.c243 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers_zc() local
249 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers_zc()
250 xdp = i40e_rx_bi(rx_ring, ntu); in i40e_alloc_rx_buffers_zc()
252 nb_buffs = min_t(u16, count, rx_ring->count - ntu); in i40e_alloc_rx_buffers_zc()
267 ntu += nb_buffs; in i40e_alloc_rx_buffers_zc()
268 if (ntu == rx_ring->count) { in i40e_alloc_rx_buffers_zc()
270 ntu = 0; in i40e_alloc_rx_buffers_zc()
275 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers_zc()
528 u16 ntu = xdp_ring->next_to_use; in i40e_xmit_pkt_batch() local
540 tx_desc = I40E_TX_DESC(xdp_ring, ntu++); in i40e_xmit_pkt_batch()
[all …]
H A Di40e_adminq.c1056 u16 ntu; in i40e_clean_arq_element() local
1072 ntu = rd32(hw, I40E_PF_ARQH) & I40E_PF_ARQH_ARQH_MASK; in i40e_clean_arq_element()
1073 if (ntu == ntc) { in i40e_clean_arq_element()
1126 hw->aq.arq.next_to_use = ntu; in i40e_clean_arq_element()
1132 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); in i40e_clean_arq_element()
H A Di40e_txrx.c1680 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers() local
1688 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers()
1689 bi = i40e_rx_bi(rx_ring, ntu); in i40e_alloc_rx_buffers()
1708 ntu++; in i40e_alloc_rx_buffers()
1709 if (unlikely(ntu == rx_ring->count)) { in i40e_alloc_rx_buffers()
1712 ntu = 0; in i40e_alloc_rx_buffers()
1721 if (rx_ring->next_to_use != ntu) in i40e_alloc_rx_buffers()
1722 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers()
1727 if (rx_ring->next_to_use != ntu) in i40e_alloc_rx_buffers()
1728 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers()
/linux/include/net/libeth/
H A Dxsk.h609 u32 ntu;
651 this = fq->count - fq->ntu; in libeth_xskfqe_alloc()
656 xskb = (typeof(xskb))&fq->fqes[fq->ntu]; in libeth_xskfqe_alloc()
659 for (u32 i = 0, ntu = fq->ntu; likely(i < ret); i++) in libeth_xskfqe_alloc() local
660 fill(fq, ntu + i); in libeth_xskfqe_alloc()
663 fq->ntu += ret; in libeth_xskfqe_alloc()
665 if (likely(fq->ntu < fq->count) || unlikely(ret < this)) in libeth_xskfqe_alloc()
668 fq->ntu = 0; in libeth_xskfqe_alloc()
H A Dxdp.h425 u32 *ntu; member
528 u32 ntu, i = 0; in libeth_xdp_tx_xmit_bulk() local
534 ntu = *sq.ntu; in libeth_xdp_tx_xmit_bulk()
536 this = sq.count - ntu; in libeth_xdp_tx_xmit_bulk()
547 u32 base = ntu + i - off; in libeth_xdp_tx_xmit_bulk()
558 xmit(fill(bulk[i], ntu + i - off, &sq, priv), in libeth_xdp_tx_xmit_bulk()
559 ntu + i - off, &sq, priv); in libeth_xdp_tx_xmit_bulk()
562 ntu += this; in libeth_xdp_tx_xmit_bulk()
563 if (likely(ntu < sq.count)) in libeth_xdp_tx_xmit_bulk()
566 ntu = 0; in libeth_xdp_tx_xmit_bulk()
[all …]
/linux/drivers/net/ethernet/hisilicon/hns3/hns3_common/
H A Dhclge_comm_cmd.c289 int ntu = ring->next_to_use; in hclge_comm_ring_space() local
290 int used = (ntu - ntc + ring->desc_num) % ring->desc_num; in hclge_comm_ring_space()
315 int ntu = ring->next_to_use; in hclge_comm_is_valid_csq_clean_head() local
317 if (ntu > ntc) in hclge_comm_is_valid_csq_clean_head()
318 return head >= ntc && head <= ntu; in hclge_comm_is_valid_csq_clean_head()
320 return head >= ntc || head <= ntu; in hclge_comm_is_valid_csq_clean_head()
/linux/drivers/net/ethernet/intel/iavf/
H A Diavf_adminq.c841 u16 ntu; in iavf_clean_arq_element() local
857 ntu = rd32(hw, IAVF_VF_ARQH1) & IAVF_VF_ARQH1_ARQH_MASK; in iavf_clean_arq_element()
858 if (ntu == ntc) { in iavf_clean_arq_element()
911 hw->aq.arq.next_to_use = ntu; in iavf_clean_arq_element()
916 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); in iavf_clean_arq_element()
H A Diavf_txrx.c867 u16 ntu = rx_ring->next_to_use; in iavf_alloc_rx_buffers() local
874 rx_desc = IAVF_RX_DESC(rx_ring, ntu); in iavf_alloc_rx_buffers()
879 addr = libeth_rx_alloc(&fq, ntu); in iavf_alloc_rx_buffers()
889 ntu++; in iavf_alloc_rx_buffers()
890 if (unlikely(ntu == rx_ring->count)) { in iavf_alloc_rx_buffers()
892 ntu = 0; in iavf_alloc_rx_buffers()
901 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers()
902 iavf_release_rx_desc(rx_ring, ntu); in iavf_alloc_rx_buffers()
907 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers()
908 iavf_release_rx_desc(rx_ring, ntu); in iavf_alloc_rx_buffers()
/linux/drivers/net/ethernet/intel/idpf/
H A Dxsk.c287 .ntu = &xdpsq->next_to_use, in idpf_xsk_tx_prep()
305 .ntu = &xdpsq->next_to_use, in idpf_xsk_xmit_prep()
361 .ntu = bufq->next_to_use, in idpf_xskfq_refill_thresh()
367 writel(fq.ntu, bufq->tail); in idpf_xskfq_refill_thresh()
369 bufq->next_to_use = fq.ntu; in idpf_xskfq_refill_thresh()
H A Dxdp.c349 .ntu = &xdpsq->next_to_use, in idpf_xdp_tx_prep()
H A Didpf_txrx.c2551 static unsigned int idpf_tx_splitq_bump_ntu(struct idpf_tx_queue *txq, u16 ntu) in idpf_tx_splitq_bump_ntu() argument
2553 ntu++; in idpf_tx_splitq_bump_ntu()
2555 if (ntu == txq->desc_count) in idpf_tx_splitq_bump_ntu()
2556 ntu = 0; in idpf_tx_splitq_bump_ntu()
2558 return ntu; in idpf_tx_splitq_bump_ntu()
/linux/drivers/net/ethernet/intel/ixgbe/
H A Dixgbe_xsk.c461 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_clean_xdp_tx_irq() local
471 while (ntc != ntu) { in ixgbe_clean_xdp_tx_irq()
544 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_xsk_clean_tx_ring() local
549 while (ntc != ntu) { in ixgbe_xsk_clean_tx_ring()
H A Dixgbe.h566 u16 ntu = ring->next_to_use; in ixgbe_desc_unused() local
568 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbe_desc_unused()
/linux/drivers/net/ethernet/intel/ixgbevf/
H A Dixgbevf.h273 u16 ntu = ring->next_to_use; in ixgbevf_desc_unused() local
275 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbevf_desc_unused()
/linux/drivers/net/ethernet/hisilicon/hibmcge/
H A Dhbg_txrx.h37 return (READ_ONCE(ring->ntu) + len - READ_ONCE(ring->ntc)) % len; in hbg_get_queue_used_num()
H A Dhbg_common.h80 u32 ntu; member
/linux/drivers/net/ethernet/intel/igc/
H A Digc.h687 u16 ntu = ring->next_to_use; in igc_desc_unused() local
689 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in igc_desc_unused()
H A Digc_main.c3044 u16 budget, ntu; in igc_xdp_xmit_zc() local
3054 ntu = ring->next_to_use; in igc_xdp_xmit_zc()
3079 bi = &ring->tx_buffer_info[ntu]; in igc_xdp_xmit_zc()
3089 ntu = ring->next_to_use; in igc_xdp_xmit_zc()
3097 tx_desc = IGC_TX_DESC(ring, ntu); in igc_xdp_xmit_zc()
3111 ntu++; in igc_xdp_xmit_zc()
3112 if (ntu == ring->count) in igc_xdp_xmit_zc()
3113 ntu = 0; in igc_xdp_xmit_zc()
3115 ring->next_to_use = ntu; in igc_xdp_xmit_zc()
/linux/drivers/net/ethernet/hisilicon/hns3/
H A Dhns3_enet.c974 u32 ntc, ntu; in hns3_tx_spare_space() local
980 ntu = tx_spare->next_to_use; in hns3_tx_spare_space()
982 if (ntc > ntu) in hns3_tx_spare_space()
983 return ntc - ntu - 1; in hns3_tx_spare_space()
988 return max(ntc, tx_spare->len - ntu) - 1; in hns3_tx_spare_space()
1113 u32 ntu = tx_spare->next_to_use; in hns3_tx_spare_alloc() local
1121 if (ntu + size > tx_spare->len) { in hns3_tx_spare_alloc()
1122 *cb_len += (tx_spare->len - ntu); in hns3_tx_spare_alloc()
1123 ntu = 0; in hns3_tx_spare_alloc()
1126 tx_spare->next_to_use = ntu + size; in hns3_tx_spare_alloc()
[all …]
/linux/drivers/net/ethernet/hisilicon/hns/
H A Dhns_enet.c729 int ntu = ring->next_to_use; in hns_desc_unused() local
731 return ((ntc >= ntu) ? 0 : ring->desc_num) + ntc - ntu; in hns_desc_unused()
/linux/drivers/net/ethernet/wangxun/libwx/
H A Dwx_lib.c379 u16 ntu = ring->next_to_use; in wx_desc_unused() local
381 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in wx_desc_unused()

12