| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_xsk.c | 297 u16 ntu = rx_ring->next_to_use; in __ice_alloc_rx_bufs_zc() local 301 rx_desc = ICE_RX_DESC(rx_ring, ntu); in __ice_alloc_rx_bufs_zc() 302 xdp = ice_xdp_buf(rx_ring, ntu); in __ice_alloc_rx_bufs_zc() 304 if (ntu + count >= rx_ring->count) { in __ice_alloc_rx_bufs_zc() 306 rx_ring->count - ntu); in __ice_alloc_rx_bufs_zc() 307 if (nb_buffs_extra != rx_ring->count - ntu) { in __ice_alloc_rx_bufs_zc() 308 ntu += nb_buffs_extra; in __ice_alloc_rx_bufs_zc() 313 ntu = 0; in __ice_alloc_rx_bufs_zc() 320 ntu += nb_buffs; in __ice_alloc_rx_bufs_zc() 321 if (ntu == rx_ring->count) in __ice_alloc_rx_bufs_zc() [all …]
|
| H A D | ice_txrx.c | 788 u32 ntu = rx_ring->next_to_use; in ice_init_ctrl_rx_descs() local 793 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_init_ctrl_rx_descs() 797 ntu++; in ice_init_ctrl_rx_descs() 798 if (unlikely(ntu == rx_ring->count)) { in ice_init_ctrl_rx_descs() 800 ntu = 0; in ice_init_ctrl_rx_descs() 807 if (rx_ring->next_to_use != ntu) in ice_init_ctrl_rx_descs() 808 ice_release_rx_desc(rx_ring, ntu); in ice_init_ctrl_rx_descs() 839 u16 ntu = rx_ring->next_to_use; in ice_alloc_rx_bufs() local 846 rx_desc = ICE_RX_DESC(rx_ring, ntu); in ice_alloc_rx_bufs() 851 addr = libeth_rx_alloc(&fq, ntu); in ice_alloc_rx_bufs() [all …]
|
| H A D | ice_controlq.c | 1184 u16 ntu; in ice_clean_rq_elem() local 1199 ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask); in ice_clean_rq_elem() 1201 if (ntu == ntc) { in ice_clean_rq_elem() 1248 cq->rq.next_to_use = ntu; in ice_clean_rq_elem() 1254 ntu = (u16)(rd32(hw, cq->rq.head) & cq->rq.head_mask); in ice_clean_rq_elem() 1255 *pending = (u16)((ntc > ntu ? cq->rq.count : 0) + (ntu - ntc)); in ice_clean_rq_elem()
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 205 u16 ntu = rx_ring->next_to_use; in igb_alloc_rx_buffers_zc() local 209 rx_desc = IGB_RX_DESC(rx_ring, ntu); in igb_alloc_rx_buffers_zc() 210 xdp = &rx_ring->rx_buffer_info_zc[ntu]; in igb_alloc_rx_buffers_zc() 212 if (ntu + count >= rx_ring->count) { in igb_alloc_rx_buffers_zc() 214 rx_ring->count - ntu); in igb_alloc_rx_buffers_zc() 215 if (nb_buffs_extra != rx_ring->count - ntu) { in igb_alloc_rx_buffers_zc() 216 ntu += nb_buffs_extra; in igb_alloc_rx_buffers_zc() 221 ntu = 0; in igb_alloc_rx_buffers_zc() 226 ntu += nb_buffs; in igb_alloc_rx_buffers_zc() 227 if (ntu == rx_ring->count) in igb_alloc_rx_buffers_zc() [all …]
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 243 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers_zc() local 249 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers_zc() 250 xdp = i40e_rx_bi(rx_ring, ntu); in i40e_alloc_rx_buffers_zc() 252 nb_buffs = min_t(u16, count, rx_ring->count - ntu); in i40e_alloc_rx_buffers_zc() 267 ntu += nb_buffs; in i40e_alloc_rx_buffers_zc() 268 if (ntu == rx_ring->count) { in i40e_alloc_rx_buffers_zc() 270 ntu = 0; in i40e_alloc_rx_buffers_zc() 275 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers_zc() 528 u16 ntu = xdp_ring->next_to_use; in i40e_xmit_pkt_batch() local 540 tx_desc = I40E_TX_DESC(xdp_ring, ntu++); in i40e_xmit_pkt_batch() [all …]
|
| H A D | i40e_adminq.c | 1056 u16 ntu; in i40e_clean_arq_element() local 1072 ntu = rd32(hw, I40E_PF_ARQH) & I40E_PF_ARQH_ARQH_MASK; in i40e_clean_arq_element() 1073 if (ntu == ntc) { in i40e_clean_arq_element() 1126 hw->aq.arq.next_to_use = ntu; in i40e_clean_arq_element() 1132 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); in i40e_clean_arq_element()
|
| H A D | i40e_txrx.c | 1680 u16 ntu = rx_ring->next_to_use; in i40e_alloc_rx_buffers() local 1688 rx_desc = I40E_RX_DESC(rx_ring, ntu); in i40e_alloc_rx_buffers() 1689 bi = i40e_rx_bi(rx_ring, ntu); in i40e_alloc_rx_buffers() 1708 ntu++; in i40e_alloc_rx_buffers() 1709 if (unlikely(ntu == rx_ring->count)) { in i40e_alloc_rx_buffers() 1712 ntu = 0; in i40e_alloc_rx_buffers() 1721 if (rx_ring->next_to_use != ntu) in i40e_alloc_rx_buffers() 1722 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers() 1727 if (rx_ring->next_to_use != ntu) in i40e_alloc_rx_buffers() 1728 i40e_release_rx_desc(rx_ring, ntu); in i40e_alloc_rx_buffers()
|
| /linux/include/net/libeth/ |
| H A D | xsk.h | 609 u32 ntu; 651 this = fq->count - fq->ntu; in libeth_xskfqe_alloc() 656 xskb = (typeof(xskb))&fq->fqes[fq->ntu]; in libeth_xskfqe_alloc() 659 for (u32 i = 0, ntu = fq->ntu; likely(i < ret); i++) in libeth_xskfqe_alloc() local 660 fill(fq, ntu + i); in libeth_xskfqe_alloc() 663 fq->ntu += ret; in libeth_xskfqe_alloc() 665 if (likely(fq->ntu < fq->count) || unlikely(ret < this)) in libeth_xskfqe_alloc() 668 fq->ntu = 0; in libeth_xskfqe_alloc()
|
| H A D | xdp.h | 425 u32 *ntu; member 528 u32 ntu, i = 0; in libeth_xdp_tx_xmit_bulk() local 534 ntu = *sq.ntu; in libeth_xdp_tx_xmit_bulk() 536 this = sq.count - ntu; in libeth_xdp_tx_xmit_bulk() 547 u32 base = ntu + i - off; in libeth_xdp_tx_xmit_bulk() 558 xmit(fill(bulk[i], ntu + i - off, &sq, priv), in libeth_xdp_tx_xmit_bulk() 559 ntu + i - off, &sq, priv); in libeth_xdp_tx_xmit_bulk() 562 ntu += this; in libeth_xdp_tx_xmit_bulk() 563 if (likely(ntu < sq.count)) in libeth_xdp_tx_xmit_bulk() 566 ntu = 0; in libeth_xdp_tx_xmit_bulk() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hns3/hns3_common/ |
| H A D | hclge_comm_cmd.c | 289 int ntu = ring->next_to_use; in hclge_comm_ring_space() local 290 int used = (ntu - ntc + ring->desc_num) % ring->desc_num; in hclge_comm_ring_space() 315 int ntu = ring->next_to_use; in hclge_comm_is_valid_csq_clean_head() local 317 if (ntu > ntc) in hclge_comm_is_valid_csq_clean_head() 318 return head >= ntc && head <= ntu; in hclge_comm_is_valid_csq_clean_head() 320 return head >= ntc || head <= ntu; in hclge_comm_is_valid_csq_clean_head()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_adminq.c | 841 u16 ntu; in iavf_clean_arq_element() local 857 ntu = rd32(hw, IAVF_VF_ARQH1) & IAVF_VF_ARQH1_ARQH_MASK; in iavf_clean_arq_element() 858 if (ntu == ntc) { in iavf_clean_arq_element() 911 hw->aq.arq.next_to_use = ntu; in iavf_clean_arq_element() 916 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc); in iavf_clean_arq_element()
|
| H A D | iavf_txrx.c | 867 u16 ntu = rx_ring->next_to_use; in iavf_alloc_rx_buffers() local 874 rx_desc = IAVF_RX_DESC(rx_ring, ntu); in iavf_alloc_rx_buffers() 879 addr = libeth_rx_alloc(&fq, ntu); in iavf_alloc_rx_buffers() 889 ntu++; in iavf_alloc_rx_buffers() 890 if (unlikely(ntu == rx_ring->count)) { in iavf_alloc_rx_buffers() 892 ntu = 0; in iavf_alloc_rx_buffers() 901 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers() 902 iavf_release_rx_desc(rx_ring, ntu); in iavf_alloc_rx_buffers() 907 if (rx_ring->next_to_use != ntu) in iavf_alloc_rx_buffers() 908 iavf_release_rx_desc(rx_ring, ntu); in iavf_alloc_rx_buffers()
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | xsk.c | 287 .ntu = &xdpsq->next_to_use, in idpf_xsk_tx_prep() 305 .ntu = &xdpsq->next_to_use, in idpf_xsk_xmit_prep() 361 .ntu = bufq->next_to_use, in idpf_xskfq_refill_thresh() 367 writel(fq.ntu, bufq->tail); in idpf_xskfq_refill_thresh() 369 bufq->next_to_use = fq.ntu; in idpf_xskfq_refill_thresh()
|
| H A D | xdp.c | 349 .ntu = &xdpsq->next_to_use, in idpf_xdp_tx_prep()
|
| H A D | idpf_txrx.c | 2551 static unsigned int idpf_tx_splitq_bump_ntu(struct idpf_tx_queue *txq, u16 ntu) in idpf_tx_splitq_bump_ntu() argument 2553 ntu++; in idpf_tx_splitq_bump_ntu() 2555 if (ntu == txq->desc_count) in idpf_tx_splitq_bump_ntu() 2556 ntu = 0; in idpf_tx_splitq_bump_ntu() 2558 return ntu; in idpf_tx_splitq_bump_ntu()
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_xsk.c | 461 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_clean_xdp_tx_irq() local 471 while (ntc != ntu) { in ixgbe_clean_xdp_tx_irq() 544 u16 ntc = tx_ring->next_to_clean, ntu = tx_ring->next_to_use; in ixgbe_xsk_clean_tx_ring() local 549 while (ntc != ntu) { in ixgbe_xsk_clean_tx_ring()
|
| H A D | ixgbe.h | 566 u16 ntu = ring->next_to_use; in ixgbe_desc_unused() local 568 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbe_desc_unused()
|
| /linux/drivers/net/ethernet/intel/ixgbevf/ |
| H A D | ixgbevf.h | 273 u16 ntu = ring->next_to_use; in ixgbevf_desc_unused() local 275 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbevf_desc_unused()
|
| /linux/drivers/net/ethernet/hisilicon/hibmcge/ |
| H A D | hbg_txrx.h | 37 return (READ_ONCE(ring->ntu) + len - READ_ONCE(ring->ntc)) % len; in hbg_get_queue_used_num()
|
| H A D | hbg_common.h | 80 u32 ntu; member
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc.h | 687 u16 ntu = ring->next_to_use; in igc_desc_unused() local 689 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in igc_desc_unused()
|
| H A D | igc_main.c | 3044 u16 budget, ntu; in igc_xdp_xmit_zc() local 3054 ntu = ring->next_to_use; in igc_xdp_xmit_zc() 3079 bi = &ring->tx_buffer_info[ntu]; in igc_xdp_xmit_zc() 3089 ntu = ring->next_to_use; in igc_xdp_xmit_zc() 3097 tx_desc = IGC_TX_DESC(ring, ntu); in igc_xdp_xmit_zc() 3111 ntu++; in igc_xdp_xmit_zc() 3112 if (ntu == ring->count) in igc_xdp_xmit_zc() 3113 ntu = 0; in igc_xdp_xmit_zc() 3115 ring->next_to_use = ntu; in igc_xdp_xmit_zc()
|
| /linux/drivers/net/ethernet/hisilicon/hns3/ |
| H A D | hns3_enet.c | 974 u32 ntc, ntu; in hns3_tx_spare_space() local 980 ntu = tx_spare->next_to_use; in hns3_tx_spare_space() 982 if (ntc > ntu) in hns3_tx_spare_space() 983 return ntc - ntu - 1; in hns3_tx_spare_space() 988 return max(ntc, tx_spare->len - ntu) - 1; in hns3_tx_spare_space() 1113 u32 ntu = tx_spare->next_to_use; in hns3_tx_spare_alloc() local 1121 if (ntu + size > tx_spare->len) { in hns3_tx_spare_alloc() 1122 *cb_len += (tx_spare->len - ntu); in hns3_tx_spare_alloc() 1123 ntu = 0; in hns3_tx_spare_alloc() 1126 tx_spare->next_to_use = ntu + size; in hns3_tx_spare_alloc() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hns/ |
| H A D | hns_enet.c | 729 int ntu = ring->next_to_use; in hns_desc_unused() local 731 return ((ntc >= ntu) ? 0 : ring->desc_num) + ntc - ntu; in hns_desc_unused()
|
| /linux/drivers/net/ethernet/wangxun/libwx/ |
| H A D | wx_lib.c | 379 u16 ntu = ring->next_to_use; in wx_desc_unused() local 381 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in wx_desc_unused()
|