Searched refs:r_idx (Results 1 – 11 of 11) sorted by relevance
716 u16 r_idx; in qtnf_pcie_pearl_rx_poll() local724 r_idx = priv->rx_bd_r_index; in qtnf_pcie_pearl_rx_poll()725 rxbd = &ps->rx_bd_vbase[r_idx]; in qtnf_pcie_pearl_rx_poll()728 skb = priv->rx_skb[r_idx]; in qtnf_pcie_pearl_rx_poll()733 pr_warn("skip invalid rxbd[%d]\n", r_idx); in qtnf_pcie_pearl_rx_poll()738 pr_warn("skip missing rx_skb[%d]\n", r_idx); in qtnf_pcie_pearl_rx_poll()774 priv->rx_skb[r_idx] = NULL; in qtnf_pcie_pearl_rx_poll()775 if (++r_idx >= priv->rx_bd_num) in qtnf_pcie_pearl_rx_poll()776 r_idx = 0; in qtnf_pcie_pearl_rx_poll()778 priv->rx_bd_r_index = r_idx; in qtnf_pcie_pearl_rx_poll()
621 u16 r_idx; in qtnf_topaz_rx_poll() local629 r_idx = priv->rx_bd_r_index; in qtnf_topaz_rx_poll()630 rxbd = &ts->rx_bd_vbase[r_idx]; in qtnf_topaz_rx_poll()633 skb = priv->rx_skb[r_idx]; in qtnf_topaz_rx_poll()639 pr_warn("skip invalid rxbd[%d]\n", r_idx); in qtnf_topaz_rx_poll()644 pr_warn("skip missing rx_skb[%d]\n", r_idx); in qtnf_topaz_rx_poll()685 priv->rx_skb[r_idx] = NULL; in qtnf_topaz_rx_poll()686 if (++r_idx >= priv->rx_bd_num) in qtnf_topaz_rx_poll()687 r_idx = 0; in qtnf_topaz_rx_poll()689 priv->rx_bd_r_index = r_idx; in qtnf_topaz_rx_poll()
197 int r_idx) in __bench_mem_function() 199 const struct function *r = &info->functions[r_idx]; in __bench_mem_function() 192 __bench_mem_function(struct bench_mem_info * info,struct bench_params * p,int r_idx) __bench_mem_function() argument
514 unsigned int r_idx = skb->queue_mapping; in fm10k_xmit_frame() local572 if (r_idx >= num_tx_queues) in fm10k_xmit_frame()573 r_idx %= num_tx_queues; in fm10k_xmit_frame()575 err = fm10k_xmit_frame_ring(skb, interface->tx_ring[r_idx]); in fm10k_xmit_frame()
144 unsigned int r_idx = skb->queue_mapping; in alx_tx_queue_mapping() local146 if (r_idx >= alx->num_txq) in alx_tx_queue_mapping()147 r_idx = r_idx % alx->num_txq; in alx_tx_queue_mapping()149 return alx->qnapi[r_idx]->txq; in alx_tx_queue_mapping()
1661 unsigned int r_idx = skb->queue_mapping; in wx_xmit_frame() local1676 if (r_idx >= wx->num_tx_queues) in wx_xmit_frame()1677 r_idx = r_idx % wx->num_tx_queues; in wx_xmit_frame()1678 tx_ring = wx->tx_ring[r_idx]; in wx_xmit_frame()
1724 unsigned int r_idx = skb->queue_mapping; in igc_tx_queue_mapping() local1726 if (r_idx >= adapter->num_tx_queues) in igc_tx_queue_mapping()1727 r_idx = r_idx % adapter->num_tx_queues; in igc_tx_queue_mapping()1729 return adapter->tx_ring[r_idx]; in igc_tx_queue_mapping()
1290 qe += q->u.r_idx * q->size; in sli_mq_read()1293 if (q->index == q->u.r_idx) { in sli_mq_read()1299 q->u.r_idx = (q->u.r_idx + 1) & (q->length - 1); in sli_mq_read()
3543 u32 r_idx; /* "read" index (MQ only) */ member
429 iavf_map_vector_to_rxq(struct iavf_adapter *adapter, int v_idx, int r_idx) in iavf_map_vector_to_rxq() argument432 struct iavf_ring *rx_ring = &adapter->rx_rings[r_idx]; in iavf_map_vector_to_rxq()442 q_vector->ring_mask |= BIT(r_idx); in iavf_map_vector_to_rxq()
6620 unsigned int r_idx = skb->queue_mapping; in igb_tx_queue_mapping() local6622 if (r_idx >= adapter->num_tx_queues) in igb_tx_queue_mapping()6623 r_idx = r_idx % adapter->num_tx_queues; in igb_tx_queue_mapping()6625 return adapter->tx_ring[r_idx]; in igb_tx_queue_mapping()