| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_xsk.c | 365 u16 ntc = xdp_ring->next_to_clean; in ice_clean_xdp_irq_zc() local 378 if (last_rs >= ntc) in ice_clean_xdp_irq_zc() 379 completed_frames = last_rs - ntc + 1; in ice_clean_xdp_irq_zc() 381 completed_frames = last_rs + cnt - ntc + 1; in ice_clean_xdp_irq_zc() 392 ntc = xdp_ring->next_to_clean; in ice_clean_xdp_irq_zc() 394 tx_buf = &xdp_ring->tx_buf[ntc]; in ice_clean_xdp_irq_zc() 404 ntc++; in ice_clean_xdp_irq_zc() 405 if (ntc >= xdp_ring->count) in ice_clean_xdp_irq_zc() 406 ntc = 0; in ice_clean_xdp_irq_zc() 573 u32 ntc = rx_ring->next_to_clean; in ice_clean_rx_irq_zc() local [all …]
|
| H A D | ice_controlq.c | 860 u16 ntc = sq->next_to_clean; in ice_clean_sq() local 863 desc = ICE_CTL_Q_DESC(*sq, ntc); in ice_clean_sq() 865 while (rd32(hw, cq->sq.head) != ntc) { in ice_clean_sq() 866 ice_debug(hw, ICE_DBG_AQ_MSG, "ntc %d head %d.\n", ntc, rd32(hw, cq->sq.head)); in ice_clean_sq() 868 ntc++; in ice_clean_sq() 869 if (ntc == sq->count) in ice_clean_sq() 870 ntc = 0; in ice_clean_sq() 871 desc = ICE_CTL_Q_DESC(*sq, ntc); in ice_clean_sq() 874 sq->next_to_clean = ntc; in ice_clean_sq() 1177 u16 ntc = cq->rq.next_to_clean; in ice_clean_rq_elem() local [all …]
|
| H A D | ice_txrx.c | 905 u32 ntc = rx_ring->next_to_clean; in ice_clean_ctrl_rx_irq() local 914 rx_desc = ICE_RX_DESC(rx_ring, ntc); in ice_clean_ctrl_rx_irq() 925 if (++ntc == cnt) in ice_clean_ctrl_rx_irq() 926 ntc = 0; in ice_clean_ctrl_rx_irq() 930 rx_ring->next_to_clean = ntc; in ice_clean_ctrl_rx_irq() 951 u32 ntc = rx_ring->next_to_clean; in ice_clean_rx_irq() local 977 rx_desc = ICE_RX_DESC(rx_ring, ntc); in ice_clean_rx_irq() 1010 rx_buf = &rx_ring->hdr_fqes[ntc]; in ice_clean_rx_irq() 1021 rx_buf = &rx_ring->rx_fqes[ntc]; in ice_clean_rx_irq() 1024 if (++ntc == cnt) in ice_clean_rx_irq() [all …]
|
| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_controlq.c | 357 u16 ntc, desc_err; in idpf_ctlq_clean_sq() local 366 ntc = cq->next_to_clean; in idpf_ctlq_clean_sq() 372 desc = IDPF_CTLQ_DESC(cq, ntc); in idpf_ctlq_clean_sq() 382 msg_status[i] = cq->bi.tx_msg[ntc]; in idpf_ctlq_clean_sq() 385 cq->bi.tx_msg[ntc] = NULL; in idpf_ctlq_clean_sq() 390 ntc++; in idpf_ctlq_clean_sq() 391 if (ntc == cq->ring_size) in idpf_ctlq_clean_sq() 392 ntc = 0; in idpf_ctlq_clean_sq() 395 cq->next_to_clean = ntc; in idpf_ctlq_clean_sq() 546 u16 num_to_clean, ntc, flags; in idpf_ctlq_recv() local [all …]
|
| H A D | xsk.c | 171 u32 ntc = xdpsq->next_to_clean; in idpf_xsksq_clean() local 182 while (ntc != xdpsq->next_to_use) { in idpf_xsksq_clean() 183 struct libeth_sqe *sqe = &xdpsq->tx_buf[ntc]; in idpf_xsksq_clean() 190 if (unlikely(++ntc == xdpsq->desc_count)) in idpf_xsksq_clean() 191 ntc = 0; in idpf_xsksq_clean() 204 u32 ntc = xdpsq->next_to_clean; in idpf_xsksq_complete_slow() local 218 struct libeth_sqe *sqe = &xdpsq->tx_buf[ntc]; in idpf_xsksq_complete_slow() 225 if (unlikely(++ntc == cnt)) in idpf_xsksq_complete_slow() 226 ntc = 0; in idpf_xsksq_complete_slow() 231 xdpsq->next_to_clean = ntc; in idpf_xsksq_complete_slow() [all …]
|
| H A D | idpf_txrx.c | 2032 #define idpf_tx_splitq_clean_bump_ntc(txq, ntc, desc, buf) \ argument 2034 if (unlikely(++(ntc) == (txq)->desc_count)) { \ 2035 ntc = 0; \ 2067 u32 ntc = tx_q->next_to_clean; in idpf_tx_splitq_clean() local 2081 tx_desc = &tx_q->flex_tx[ntc]; in idpf_tx_splitq_clean() 2083 tx_buf = &tx_q->tx_buf[ntc]; in idpf_tx_splitq_clean() 2102 while (ntc != eop_idx) { in idpf_tx_splitq_clean() 2103 idpf_tx_splitq_clean_bump_ntc(tx_q, ntc, in idpf_tx_splitq_clean() 2111 idpf_tx_splitq_clean_bump_ntc(tx_q, ntc, tx_desc, tx_buf); in idpf_tx_splitq_clean() 2114 tx_q->next_to_clean = ntc; in idpf_tx_splitq_clean() [all …]
|
| H A D | xdp.c | 258 u32 ntc = cq->next_to_clean; in idpf_xdpsq_poll() local 268 ret = idpf_xdp_parse_cqe(&cq->comp_4b[ntc], gen); in idpf_xdpsq_poll() 283 if (unlikely(++ntc == cnt)) { in idpf_xdpsq_poll() 284 ntc = 0; in idpf_xdpsq_poll() 291 cq->next_to_clean = ntc; in idpf_xdpsq_poll()
|
| H A D | idpf_txrx.h | 90 #define IDPF_RX_BUMP_NTC(rxq, ntc) \ argument 92 if (unlikely(++(ntc) == (rxq)->desc_count)) { \ 93 ntc = 0; \
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_adminq.c | 560 u16 ntc = asq->next_to_clean; in iavf_clean_asq() local 564 desc = IAVF_ADMINQ_DESC(*asq, ntc); in iavf_clean_asq() 565 details = IAVF_ADMINQ_DETAILS(*asq, ntc); in iavf_clean_asq() 566 while (rd32(hw, IAVF_VF_ATQH1) != ntc) { in iavf_clean_asq() 568 "ntc %d head %d.\n", ntc, rd32(hw, IAVF_VF_ATQH1)); in iavf_clean_asq() 579 ntc++; in iavf_clean_asq() 580 if (ntc == asq->count) in iavf_clean_asq() 581 ntc = 0; in iavf_clean_asq() 582 desc = IAVF_ADMINQ_DESC(*asq, ntc); in iavf_clean_asq() 583 details = IAVF_ADMINQ_DETAILS(*asq, ntc); in iavf_clean_asq() [all …]
|
| H A D | iavf_txrx.c | 1254 u32 ntc = rx_ring->next_to_clean + 1; in iavf_is_non_eop() local 1257 ntc = (ntc < rx_ring->count) ? ntc : 0; in iavf_is_non_eop() 1258 rx_ring->next_to_clean = ntc; in iavf_is_non_eop() 1260 prefetch(IAVF_RX_DESC(rx_ring, ntc)); in iavf_is_non_eop()
|
| /linux/drivers/net/ethernet/intel/igb/ |
| H A D | igb_xsk.c | 252 u16 ntc = rx_ring->next_to_clean; in igb_clean_rx_ring_zc() local 255 while (ntc != ntu) { in igb_clean_rx_ring_zc() 256 struct xdp_buff *xdp = rx_ring->rx_buffer_info_zc[ntc]; in igb_clean_rx_ring_zc() 259 ntc++; in igb_clean_rx_ring_zc() 260 if (ntc >= rx_ring->count) in igb_clean_rx_ring_zc() 261 ntc = 0; in igb_clean_rx_ring_zc() 347 u32 ntc = rx_ring->next_to_clean; in igb_clean_rx_irq_zc() local 364 rx_desc = IGB_RX_DESC(rx_ring, ntc); in igb_clean_rx_irq_zc() 375 xdp = rx_ring->rx_buffer_info_zc[ntc]; in igb_clean_rx_irq_zc() 407 ntc++; in igb_clean_rx_irq_zc() [all …]
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_xsk.c | 636 unsigned int ntc; in i40e_clean_xdp_tx_irq() local 650 ntc = tx_ring->next_to_clean; in i40e_clean_xdp_tx_irq() 653 tx_bi = &tx_ring->tx_bi[ntc]; in i40e_clean_xdp_tx_irq() 662 if (++ntc >= tx_ring->count) in i40e_clean_xdp_tx_irq() 663 ntc = 0; in i40e_clean_xdp_tx_irq() 729 u16 ntc = rx_ring->next_to_clean; in i40e_xsk_clean_rx_ring() local 732 while (ntc != ntu) { in i40e_xsk_clean_rx_ring() 733 struct xdp_buff *rx_bi = *i40e_rx_bi(rx_ring, ntc); in i40e_xsk_clean_rx_ring() 736 ntc++; in i40e_xsk_clean_rx_ring() 737 if (ntc >= rx_ring->count) in i40e_xsk_clean_rx_ring() [all …]
|
| H A D | i40e_adminq.c | 693 u16 ntc = asq->next_to_clean; in i40e_clean_asq() local 697 desc = I40E_ADMINQ_DESC(*asq, ntc); in i40e_clean_asq() 698 details = I40E_ADMINQ_DETAILS(*asq, ntc); in i40e_clean_asq() 699 while (rd32(hw, I40E_PF_ATQH) != ntc) { in i40e_clean_asq() 701 "ntc %d head %d.\n", ntc, rd32(hw, I40E_PF_ATQH)); in i40e_clean_asq() 711 ntc++; in i40e_clean_asq() 712 if (ntc == asq->count) in i40e_clean_asq() 713 ntc = 0; in i40e_clean_asq() 714 desc = I40E_ADMINQ_DESC(*asq, ntc); in i40e_clean_asq() 715 details = I40E_ADMINQ_DETAILS(*asq, ntc); in i40e_clean_asq() [all …]
|
| /linux/drivers/net/ethernet/hisilicon/hns3/hns3_common/ |
| H A D | hclge_comm_cmd.c | 288 int ntc = ring->next_to_clean; in hclge_comm_ring_space() local 290 int used = (ntu - ntc + ring->desc_num) % ring->desc_num; in hclge_comm_ring_space() 314 int ntc = ring->next_to_clean; in hclge_comm_is_valid_csq_clean_head() local 317 if (ntu > ntc) in hclge_comm_is_valid_csq_clean_head() 318 return head >= ntc && head <= ntu; in hclge_comm_is_valid_csq_clean_head() 320 return head >= ntc || head <= ntu; in hclge_comm_is_valid_csq_clean_head() 413 int ntc) in hclge_comm_cmd_check_retval() argument 420 desc[handle] = hw->cmq.csq.desc[ntc]; in hclge_comm_cmd_check_retval() 421 ntc++; in hclge_comm_cmd_check_retval() 422 if (ntc >= hw->cmq.csq.desc_num) in hclge_comm_cmd_check_retval() [all …]
|
| /linux/arch/mips/kernel/ |
| H A D | smp-mt.c | 186 unsigned int mvpconf0, ntc, tc, ncpu = 0; in vsmp_smp_setup() local 205 ntc = (mvpconf0 & MVPCONF0_PTC) >> MVPCONF0_PTC_SHIFT; in vsmp_smp_setup() 212 for (tc = 0; tc <= ntc; tc++) { in vsmp_smp_setup()
|
| /linux/drivers/net/ethernet/intel/ixgbevf/ |
| H A D | ixgbevf.h | 272 u16 ntc = ring->next_to_clean; in ixgbevf_desc_unused() local 275 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbevf_desc_unused()
|
| /linux/drivers/net/ethernet/hisilicon/hibmcge/ |
| H A D | hbg_txrx.h | 37 return (READ_ONCE(ring->ntu) + len - READ_ONCE(ring->ntc)) % len; in hbg_get_queue_used_num()
|
| H A D | hbg_common.h | 76 u32 ntc; member
|
| /linux/drivers/net/ethernet/mscc/ |
| H A D | ocelot_fdma.c | 492 u16 ntc; in ocelot_fdma_tx_cleanup() local 501 ntc = tx_ring->next_to_clean; in ocelot_fdma_tx_cleanup() 502 dcb = &tx_ring->dcbs[ntc]; in ocelot_fdma_tx_cleanup() 506 buf = &tx_ring->bufs[ntc]; in ocelot_fdma_tx_cleanup() 514 tx_ring->next_to_clean = ocelot_fdma_idx_next(ntc, in ocelot_fdma_tx_cleanup()
|
| /linux/drivers/net/ethernet/intel/igc/ |
| H A D | igc.h | 686 u16 ntc = ring->next_to_clean; in igc_desc_unused() local 689 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in igc_desc_unused()
|
| H A D | igc_main.c | 2125 u32 ntc = rx_ring->next_to_clean + 1; in igc_is_non_eop() local 2128 ntc = (ntc < rx_ring->count) ? ntc : 0; in igc_is_non_eop() 2129 rx_ring->next_to_clean = ntc; in igc_is_non_eop() 2131 prefetch(IGC_RX_DESC(rx_ring, ntc)); in igc_is_non_eop() 2807 u16 ntc = ring->next_to_clean; in igc_clean_rx_irq_zc() local 2823 desc = IGC_RX_DESC(ring, ntc); in igc_clean_rx_irq_zc() 2834 bi = &ring->rx_buffer_info[ntc]; in igc_clean_rx_irq_zc() 2874 ntc++; in igc_clean_rx_irq_zc() 2875 if (ntc == ring->count) in igc_clean_rx_irq_zc() 2876 ntc = 0; in igc_clean_rx_irq_zc() [all …]
|
| /linux/drivers/net/ethernet/intel/fm10k/ |
| H A D | fm10k_main.c | 492 u32 ntc = rx_ring->next_to_clean + 1; in fm10k_is_non_eop() local 495 ntc = (ntc < rx_ring->count) ? ntc : 0; in fm10k_is_non_eop() 496 rx_ring->next_to_clean = ntc; in fm10k_is_non_eop() 498 prefetch(FM10K_RX_DESC(rx_ring, ntc)); in fm10k_is_non_eop()
|
| /linux/drivers/net/ethernet/wangxun/libwx/ |
| H A D | wx_lib.c | 378 u16 ntc = ring->next_to_clean; in wx_desc_unused() local 381 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in wx_desc_unused() 400 u32 ntc = rx_ring->next_to_clean + 1; in wx_is_non_eop() local 403 ntc = (ntc < rx_ring->count) ? ntc : 0; in wx_is_non_eop() 404 rx_ring->next_to_clean = ntc; in wx_is_non_eop() 406 prefetch(WX_RX_DESC(rx_ring, ntc)); in wx_is_non_eop() 420 ntc = le32_to_cpu(rx_desc->wb.upper.status_error); in wx_is_non_eop() 421 ntc &= WX_RXD_NEXTP_MASK; in wx_is_non_eop() 422 ntc >>= WX_RXD_NEXTP_SHIFT; in wx_is_non_eop() 430 rx_ring->rx_buffer_info[ntc].skb = skb; in wx_is_non_eop()
|
| /linux/drivers/net/ethernet/hisilicon/hns3/ |
| H A D | hns3_enet.c | 974 u32 ntc, ntu; in hns3_tx_spare_space() local 979 ntc = smp_load_acquire(&tx_spare->last_to_clean); in hns3_tx_spare_space() 982 if (ntc > ntu) in hns3_tx_spare_space() 983 return ntc - ntu - 1; in hns3_tx_spare_space() 988 return max(ntc, tx_spare->len - ntu) - 1; in hns3_tx_spare_space() 1151 u32 ntc = tx_spare->next_to_clean; in hns3_tx_spare_reclaim_cb() local 1160 ntc = 0; in hns3_tx_spare_reclaim_cb() 1171 dma_addr_t dma = tx_spare->dma + ntc; in hns3_tx_spare_reclaim_cb() 1176 struct sg_table *sgt = tx_spare->buf + ntc; in hns3_tx_spare_reclaim_cb() 3633 int ntc = ring->next_to_clean; in hns3_nic_reclaim_desc() local [all …]
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe.h | 566 u16 ntc = ring->next_to_clean; in ixgbe_desc_unused() local 569 return ((ntc > ntu) ? 0 : ring->count) + ntc - ntu - 1; in ixgbe_desc_unused()
|