Home
last modified time | relevance | path

Searched refs:rxqs (Results 1 – 25 of 31) sorted by relevance

12

/linux/drivers/net/ethernet/fungible/funeth/
H A Dfuneth_main.c182 struct funeth_rxq **rxqs = rtnl_dereference(fp->rxqs); in fun_config_rss() local
214 *indir_tab++ = cpu_to_be32(rxqs[*qtable++]->hw_cqid); in fun_config_rss()
378 static void free_rxqs(struct funeth_rxq **rxqs, unsigned int nqs, in free_rxqs() argument
383 for (i = start; i < nqs && rxqs[i]; i++) in free_rxqs()
384 rxqs[i] = funeth_rxq_free(rxqs[i], state); in free_rxqs()
387 static int alloc_rxqs(struct net_device *dev, struct funeth_rxq **rxqs, in alloc_rxqs() argument
398 state, &rxqs[i]); in alloc_rxqs()
400 free_rxqs(rxqs, nqs, start, FUN_QSTATE_DESTROYED); in alloc_rxqs()
445 struct funeth_rxq **rxqs = qset->rxqs; in fun_free_rings() local
450 if (!rxqs) { in fun_free_rings()
[all …]
/linux/drivers/net/ethernet/huawei/hinic/
H A Dhinic_main.c117 gather_rx_stats(nic_rx_stats, &nic_dev->rxqs[i]); in gather_nic_stats()
231 if (nic_dev->rxqs) in create_rxqs()
234 nic_dev->rxqs = devm_kcalloc(&netdev->dev, num_rxqs, in create_rxqs()
235 sizeof(*nic_dev->rxqs), GFP_KERNEL); in create_rxqs()
236 if (!nic_dev->rxqs) in create_rxqs()
244 err = hinic_init_rxq(&nic_dev->rxqs[i], rq, netdev); in create_rxqs()
262 hinic_clean_rxq(&nic_dev->rxqs[i]); in create_rxqs()
265 hinic_rq_debug_rem(nic_dev->rxqs[j].rq); in create_rxqs()
266 hinic_clean_rxq(&nic_dev->rxqs[j]); in create_rxqs()
271 devm_kfree(&netdev->dev, nic_dev->rxqs); in create_rxqs()
[all …]
H A Dhinic_dev.h98 struct hinic_rxq *rxqs; member
H A Dhinic_debugfs.c240 rq = dev->rxqs[rq_id].rq; in hinic_rq_debug_add()
H A Dhinic_ethtool.c700 msix_idx = set_rx_coal ? nic_dev->rxqs[q_id].rq->msix_entry : in set_queue_coalesce()
1361 if (!nic_dev->rxqs) in get_drv_queue_stats()
1364 hinic_rxq_get_stats(&nic_dev->rxqs[qid], &rxq_stats); in get_drv_queue_stats()
/linux/net/
H A Ddevres.c22 unsigned int txqs, unsigned int rxqs) in devm_alloc_etherdev_mqs() argument
30 dr->ndev = alloc_etherdev_mqs(sizeof_priv, txqs, rxqs); in devm_alloc_etherdev_mqs()
/linux/drivers/net/ethernet/huawei/hinic3/
H A Dhinic3_irq.c155 nic_dev->rxqs[q_id].last_coalesc_timer_cfg = coalesc_timer_cfg; in hinic3_set_interrupt_moder()
156 nic_dev->rxqs[q_id].last_pending_limit = pending_limit; in hinic3_set_interrupt_moder()
216 irq_cfg->rxq = &nic_dev->rxqs[q_id]; in hinic3_qps_irq_init()
217 nic_dev->rxqs[q_id].irq_cfg = irq_cfg; in hinic3_qps_irq_init()
H A Dhinic3_rx.c69 nic_dev->rxqs = kzalloc_objs(*nic_dev->rxqs, num_rxqs); in hinic3_alloc_rxqs()
70 if (!nic_dev->rxqs) in hinic3_alloc_rxqs()
74 rxq = &nic_dev->rxqs[q_id]; in hinic3_alloc_rxqs()
93 kfree(nic_dev->rxqs); in hinic3_free_rxqs()
506 rxq = &nic_dev->rxqs[q_id]; in hinic3_configure_rxqs()
H A Dhinic3_netdev_ops.c815 if (!nic_dev->rxqs) in hinic3_get_stats64()
818 rxq = &nic_dev->rxqs[i]; in hinic3_get_stats64()
/linux/drivers/net/ethernet/netronome/nfp/
H A Dnfp_net_repr.h102 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs);
H A Dnfp_net_repr.c424 nfp_repr_alloc_mqs(struct nfp_app *app, unsigned int txqs, unsigned int rxqs) in nfp_repr_alloc_mqs() argument
429 netdev = alloc_etherdev_mqs(sizeof(*repr), txqs, rxqs); in nfp_repr_alloc_mqs()
/linux/drivers/net/can/dev/
H A Ddev.c299 unsigned int txqs, unsigned int rxqs) in alloc_candev_mqs() argument
326 txqs, rxqs); in alloc_candev_mqs()
/linux/include/linux/can/
H A Ddev.h106 unsigned int txqs, unsigned int rxqs);
/linux/drivers/net/ethernet/marvell/
H A Dmvneta.c510 struct mvneta_rx_queue *rxqs; member
1269 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_port_up()
3324 &pp->rxqs[rx_queue]); in mvneta_poll()
3327 &pp->rxqs[rx_queue]); in mvneta_poll()
3665 mvneta_rxq_deinit(pp, &pp->rxqs[queue]); in mvneta_cleanup_rxqs()
3675 int err = mvneta_rxq_init(pp, &pp->rxqs[queue]); in mvneta_setup_rxqs()
4736 struct mvneta_rx_queue *rxq = &pp->rxqs[queue]; in mvneta_ethtool_set_coalesce()
4761 c->rx_coalesce_usecs = pp->rxqs[0].time_coal; in mvneta_ethtool_get_coalesce()
4762 c->rx_max_coalesced_frames = pp->rxqs[0].pkts_coal; in mvneta_ethtool_get_coalesce()
4973 if (pp->rxqs[i].page_pool) in mvneta_ethtool_pp_stats()
[all …]
/linux/drivers/net/ethernet/microsoft/mana/
H A Dmana_en.c604 rx_stats = &apc->rxqs[q]->stats; in mana_get_stats64()
985 kfree(apc->rxqs); in mana_cleanup_port_context()
986 apc->rxqs = NULL; in mana_cleanup_port_context()
998 apc->rxqs = kzalloc_objs(struct mana_rxq *, apc->num_queues); in mana_init_port_context()
1000 return !apc->rxqs ? -ENOMEM : 0; in mana_init_port_context()
1710 rxq = apc->rxqs[rxq_idx]; in mana_fence_rqs()
2753 rxq = apc->rxqs[idx]; in mana_create_rxq_debugfs()
2788 apc->rxqs[i] = rxq; in mana_add_rx_queues()
2793 apc->default_rxobj = apc->rxqs[0]->rxobj; in mana_add_rx_queues()
2805 rxq = apc->rxqs[rxq_idx]; in mana_destroy_vport()
[all …]
/linux/drivers/infiniband/hw/hfi1/
H A Dipoib_main.c245 params->rxqs = dd->num_netdev_contexts; in hfi1_ipoib_rn_get_params()
/linux/drivers/net/ethernet/intel/idpf/
H A Didpf_txrx.c509 idpf_rx_desc_rel(rx_qgrp->singleq.rxqs[j], dev, in idpf_rx_desc_rel_all()
806 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_bufs_init_all()
926 q = rx_qgrp->singleq.rxqs[j]; in idpf_rx_desc_alloc_all()
1380 kfree(rx_qgrp->singleq.rxqs[j]); in idpf_rxq_group_rel()
1381 rx_qgrp->singleq.rxqs[j] = NULL; in idpf_rxq_group_rel()
1817 rx_qgrp->singleq.rxqs[j] = kzalloc_obj(*rx_qgrp->singleq.rxqs[j]); in idpf_rxq_group_alloc()
1818 if (!rx_qgrp->singleq.rxqs[j]) { in idpf_rxq_group_alloc()
1886 q = rx_qgrp->singleq.rxqs[j]; in idpf_rxq_group_alloc()
4394 q = rx_qgrp->singleq.rxqs[j]; in idpf_vport_intr_map_vector_to_qs()
H A Didpf_ptp.c345 rxq = grp->singleq.rxqs[i]; in idpf_ptp_update_phctime_rxq_grp()
712 rx_queue = grp->singleq.rxqs[j]; in idpf_ptp_set_rx_tstamp()
H A Didpf_ethtool.c1143 rxq = rxq_grp->singleq.rxqs[j]; in idpf_collect_queue_stats()
1278 rxq = rxq_grp->singleq.rxqs[j]; in idpf_get_ethtool_stats()
1308 return rsrc->rxq_grps->singleq.rxqs[q_num]->q_vector; in idpf_find_rxq_vec()
H A Dxdp.c35 q = rx_qgrp->singleq.rxqs[j]; in idpf_rxq_for_each()
H A Didpf_txrx.h944 struct idpf_rx_queue *rxqs[IDPF_LARGE_MAX_Q]; member
H A Didpf_virtchnl.c1435 q = rx_qgrp->singleq.rxqs[j]; in __idpf_queue_reg_init()
2099 qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_config_rx_queues_msg()
2261 qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_ena_dis_queues_msg()
2443 qs->qs[k++].rxq = rx_qgrp->singleq.rxqs[j]; in idpf_send_map_unmap_queue_vector_msg()
3942 q = rx_qgrp->singleq.rxqs[j]; in __idpf_vport_queue_ids_init()
/linux/drivers/net/ethernet/marvell/mvpp2/
H A Dmvpp2_main.c727 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_long_pool_set()
748 prxq = port->rxqs[lrxq]->id; in mvpp2_rxq_short_pool_set()
2305 queue = port->rxqs[lrxq]->id; in mvpp2_defaults_set()
2323 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_enable()
2336 queue = port->rxqs[lrxq]->id; in mvpp2_ingress_disable()
2846 return port->rxqs[queue]; in mvpp2_get_rx_queue()
3300 mvpp2_rxq_deinit(port, port->rxqs[queue]); in mvpp2_cleanup_rxqs()
3312 err = mvpp2_rxq_init(port, port->rxqs[queue]); in mvpp2_setup_rxqs()
5408 struct mvpp2_rx_queue *rxq = port->rxqs[queue]; in mvpp2_ethtool_set_coalesce()
5442 c->rx_coalesce_usecs = port->rxqs[0]->time_coal; in mvpp2_ethtool_get_coalesce()
[all …]
H A Dmvpp2.h1212 struct mvpp2_rx_queue **rxqs; member
/linux/drivers/net/ethernet/ibm/
H A Dibmvnic.c195 struct ibmvnic_sub_crq_queue **rxqs; in ibmvnic_clean_affinity() local
200 rxqs = adapter->rx_scrq; in ibmvnic_clean_affinity()
210 if (rxqs) { in ibmvnic_clean_affinity()
212 ibmvnic_clean_queue_affinity(adapter, rxqs[i]); in ibmvnic_clean_affinity()
259 struct ibmvnic_sub_crq_queue **rxqs = adapter->rx_scrq; in ibmvnic_set_affinity() local
290 queue = rxqs[i_rxqs++]; in ibmvnic_set_affinity()

12