| /linux/net/core/ |
| H A D | netdev_rx_queue.c | 39 bool netif_rxq_is_leased(struct net_device *dev, unsigned int rxq_idx) in netif_rxq_is_leased() argument 41 if (rxq_idx < dev->real_num_rx_queues) in netif_rxq_is_leased() 42 return READ_ONCE(__netif_get_rx_queue(dev, rxq_idx)->lease); in netif_rxq_is_leased() 66 __netif_get_rx_queue_lease(struct net_device **dev, unsigned int *rxq_idx, in __netif_get_rx_queue_lease() argument 70 struct netdev_rx_queue *rxq = __netif_get_rx_queue(orig_dev, *rxq_idx); in __netif_get_rx_queue_lease() 76 *rxq_idx = get_netdev_rx_queue_index(rxq); in __netif_get_rx_queue_lease() 83 bool netif_rxq_has_unreadable_mp(struct net_device *dev, unsigned int rxq_idx) in netif_rxq_has_unreadable_mp() argument 85 if (rxq_idx < dev->real_num_rx_queues) in netif_rxq_has_unreadable_mp() 86 return __netif_get_rx_queue(dev, rxq_idx)->mp_params.mp_ops; in netif_rxq_has_unreadable_mp() 91 bool netif_rxq_has_mp(struct net_device *dev, unsigned int rxq_idx) in netif_rxq_has_mp() argument [all …]
|
| H A D | devmem.c | 128 unsigned int rxq_idx; in net_devmem_unbind_dmabuf() local 146 rxq_idx = get_netdev_rx_queue_index(rxq); in net_devmem_unbind_dmabuf() 148 netif_mp_close_rxq(binding->dev, rxq_idx, &mp_params); in net_devmem_unbind_dmabuf() 154 int net_devmem_bind_dmabuf_to_queue(struct net_device *dev, u32 rxq_idx, in net_devmem_bind_dmabuf_to_queue() argument 166 err = netif_mp_open_rxq(dev, rxq_idx, &mp_params, extack); in net_devmem_bind_dmabuf_to_queue() 170 rxq = __netif_get_rx_queue(dev, rxq_idx); in net_devmem_bind_dmabuf_to_queue() 179 netif_mp_close_rxq(dev, rxq_idx, &mp_params); in net_devmem_bind_dmabuf_to_queue()
|
| H A D | dev.h | 105 int netdev_queue_config_validate(struct net_device *dev, int rxq_idx, 109 bool netif_rxq_has_mp(struct net_device *dev, unsigned int rxq_idx); 110 bool netif_rxq_is_leased(struct net_device *dev, unsigned int rxq_idx);
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_arfs.h | 46 u16 rxq_idx, u32 flow_id); 71 u16 __always_unused rxq_idx, u32 __always_unused flow_id) in ice_rx_flow_steer() argument
|
| H A D | ice_arfs.c | 302 u16 rxq_idx, u32 flow_id) in ice_arfs_build_entry() argument 315 fltr_info->q_index = rxq_idx; in ice_arfs_build_entry() 442 u16 rxq_idx, u32 flow_id) in ice_rx_flow_steer() argument 501 if (fltr_info->q_index == rxq_idx || in ice_rx_flow_steer() 506 fltr_info->q_index = rxq_idx; in ice_rx_flow_steer() 512 arfs_entry = ice_arfs_build_entry(vsi, &fk, rxq_idx, flow_id); in ice_rx_flow_steer()
|
| H A D | ice_base.c | 840 ice_vsi_ctrl_one_rx_ring(struct ice_vsi *vsi, bool ena, u16 rxq_idx, bool wait) in ice_vsi_ctrl_one_rx_ring() argument 842 int pf_q = vsi->rxq_map[rxq_idx]; in ice_vsi_ctrl_one_rx_ring() 878 int ice_vsi_wait_one_rx_ring(struct ice_vsi *vsi, bool ena, u16 rxq_idx) in ice_vsi_wait_one_rx_ring() argument 880 int pf_q = vsi->rxq_map[rxq_idx]; in ice_vsi_wait_one_rx_ring()
|
| H A D | ice_sriov.c | 1119 u16 rxq_idx; in ice_get_vf_from_pfq() local 1125 ice_for_each_rxq(vsi, rxq_idx) in ice_get_vf_from_pfq() 1126 if (vsi->rxq_map[rxq_idx] == pfq) { in ice_get_vf_from_pfq()
|
| /linux/drivers/net/ethernet/microsoft/mana/ |
| H A D | mana_en.c | 1701 rxq->rxq_idx, err); in mana_fence_rq() 1708 rxq->rxq_idx, err, resp.hdr.status); in mana_fence_rq() 1717 rxq->rxq_idx); in mana_fence_rq() 1726 unsigned int rxq_idx; in mana_fence_rqs() local 1730 for (rxq_idx = 0; rxq_idx < apc->num_queues; rxq_idx++) { in mana_fence_rqs() 1731 rxq = apc->rxqs[rxq_idx]; in mana_fence_rqs() 1945 u16 rxq_idx = rxq->rxq_idx; in mana_rx_skb() local 1980 skb_record_rx_queue(skb, rxq_idx); in mana_rx_skb() 2011 skb_set_queue_mapping(skb, rxq_idx); in mana_rx_skb() 2660 pprm.queue_idx = rxq->rxq_idx; in mana_create_page_pool() [all …]
|
| /linux/drivers/net/ethernet/meta/fbnic/ |
| H A D | fbnic_txrx.c | 1552 unsigned int rxq_idx) in fbnic_alloc_qt_page_pools() argument 1565 .queue_idx = rxq_idx, in fbnic_alloc_qt_page_pools() 1586 if (netif_rxq_has_unreadable_mp(fbn->netdev, rxq_idx)) { in fbnic_alloc_qt_page_pools() 1618 unsigned int rxq_count, unsigned int rxq_idx) in fbnic_alloc_napi_vector() argument 1705 unsigned int xdp_idx = FBNIC_MAX_TXQS + rxq_idx; in fbnic_alloc_napi_vector() 1729 db = &uc_addr[FBNIC_QUEUE(rxq_idx) + FBNIC_QUEUE_BDQ_HPQ_TAIL]; in fbnic_alloc_napi_vector() 1734 db = &uc_addr[FBNIC_QUEUE(rxq_idx) + FBNIC_QUEUE_BDQ_PPQ_TAIL]; in fbnic_alloc_napi_vector() 1739 db = &uc_addr[FBNIC_QUEUE(rxq_idx) + FBNIC_QUEUE_RCQ_HEAD]; in fbnic_alloc_napi_vector() 1740 fbnic_ring_init(&qt->cmpl, db, rxq_idx, FBNIC_RING_F_STATS); in fbnic_alloc_napi_vector() 1741 fbn->rx[rxq_idx] = &qt->cmpl; in fbnic_alloc_napi_vector() [all …]
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/chtls/ |
| H A D | chtls_cm.c | 1128 int rxq_idx; in chtls_recv_sock() local 1232 rxq_idx = port_id * step; in chtls_recv_sock() 1233 rxq_idx += cdev->round_robin_cnt++ % step; in chtls_recv_sock() 1234 csk->rss_qid = cdev->lldi->rxq_ids[rxq_idx]; in chtls_recv_sock() 1235 csk->txq_idx = (rxq_idx < cdev->lldi->ntxq) ? rxq_idx : in chtls_recv_sock()
|
| /linux/drivers/target/iscsi/cxgbit/ |
| H A D | cxgbit_cm.c | 906 u16 rxq_idx, port_id; in cxgbit_offload_init() local 984 rxq_idx = (port_id * step) + in cxgbit_offload_init() 986 csk->rss_qid = cdev->lldi.rxq_ids[rxq_idx]; in cxgbit_offload_init()
|
| /linux/drivers/scsi/cxgbi/cxgb4i/ |
| H A D | cxgb4i.c | 1775 unsigned int rxq_idx; in init_act_open() local 1866 rxq_idx = (cxgb4_port_idx(ndev) * step) + (cdev->rxq_idx_cntr % step); in init_act_open() 1868 csk->rss_qid = lldi->rxq_ids[rxq_idx]; in init_act_open()
|
| /linux/drivers/net/ethernet/broadcom/ |
| H A D | tg3.c | 7397 int rxq_idx = tp->rxq_cnt - 1; in tg3_napi_disable() local 7409 netif_queue_set_napi(tp->dev, rxq_idx, in tg3_napi_disable() 7411 rxq_idx--; in tg3_napi_disable() 7419 int txq_idx = 0, rxq_idx = 0; in tg3_napi_enable() local 7433 netif_queue_set_napi(tp->dev, rxq_idx, in tg3_napi_enable() 7436 rxq_idx++; in tg3_napi_enable()
|