| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_xsk.c | 122 u16 tx_queues = pf->hw.tx_queues; in otx2_xsk_pool_enable() local 125 if (qidx >= rx_queues || qidx >= tx_queues) in otx2_xsk_pool_enable() 150 sq = &pf->qset.sq[qidx + pf->hw.tx_queues]; in otx2_xsk_pool_disable() 178 if (queue_id >= pf->hw.rx_queues || queue_id >= pf->hw.tx_queues) in otx2_xsk_wakeup()
|
| H A D | qos.c | 29 int tx_queues, qos_txqs, err; in otx2_qos_update_tx_netdev_queues() local 34 tx_queues = hw->tx_queues + qos_txqs; in otx2_qos_update_tx_netdev_queues() 36 err = netif_set_real_num_tx_queues(pfvf->netdev, tx_queues); in otx2_qos_update_tx_netdev_queues() 39 "Failed to set no of Tx queues: %d\n", tx_queues); in otx2_qos_update_tx_netdev_queues() 608 res = pfvf->hw.tx_queues + qid; in otx2_get_txq_by_classid() 1344 return pfvf->hw.tx_queues + qid; in otx2_qos_leaf_alloc_queue() 1593 otx2_reset_qdisc(pfvf->netdev, pfvf->hw.tx_queues + moved_qid); in otx2_qos_leaf_del()
|
| H A D | otx2_pf.c | 1320 int tx_queues, int rx_queues) in otx2_set_real_num_queues() argument 1324 err = netif_set_real_num_tx_queues(netdev, tx_queues); in otx2_set_real_num_queues() 1327 "Failed to set no of Tx queues: %d\n", tx_queues); in otx2_set_real_num_queues() 1933 pf->hw.non_qos_queues = pf->hw.tx_queues + pf->hw.xdp_queues; in otx2_alloc_queue_mem() 1934 pf->hw.cint_cnt = max3(pf->hw.rx_queues, pf->hw.tx_queues, in otx2_alloc_queue_mem() 1998 cq_poll->cq_ids[CQ_TX] = (qidx < pf->hw.tx_queues) ? in otx2_open() 2003 pf->hw.tx_queues) : in otx2_open() 2229 sq_idx = (qidx >= pf->hw.tx_queues) ? (qidx + pf->hw.xdp_queues) : qidx; in otx2_xmit() 2284 qos_enabled = netdev->real_num_tx_queues > pf->hw.tx_queues; in otx2_select_queue() 2305 if ((vlan_prio > pf->hw.tx_queues - 1) || in otx2_select_queue() [all …]
|
| H A D | otx2_txrx.c | 527 if (qidx >= pfvf->hw.tx_queues) in otx2_tx_napi_handler() 813 sqe_hdr->sq = (qidx >= pfvf->hw.tx_queues) ? in otx2_sqe_add_hdr() 1354 if (qidx >= pfvf->hw.tx_queues) in otx2_cleanup_tx_cqes() 1394 for (sq_idx = 0; sq_idx < pfvf->hw.tx_queues; sq_idx++) { in otx2_free_pending_sqe() 1534 qidx += pfvf->hw.tx_queues; in otx2_xdp_rcv_pkt_handler()
|
| H A D | otx2_ethtool.c | 270 channel->tx_count = pfvf->hw.tx_queues; in otx2_get_channels() 302 pfvf->hw.tx_queues = channel->tx_count; in otx2_set_channels() 310 pfvf->hw.tx_queues, pfvf->hw.rx_queues); in otx2_set_channels()
|
| H A D | rep.c | 517 cq_poll->cq_ids[CQ_TX] = (qidx < hw->tx_queues) ? in rvu_rep_napi_init() 756 priv->hw.tx_queues = rsp->rep_cnt; in rvu_get_rep_cnt()
|
| H A D | otx2_common.h | 201 u16 tx_queues; member 1112 int tx_queues, int rx_queues);
|
| H A D | otx2_common.c | 1000 if (qidx < pfvf->hw.tx_queues) { in otx2_sq_init() 1012 if (pfvf->ptp && qidx < pfvf->hw.tx_queues) { in otx2_sq_init() 1061 non_xdp_queues = pfvf->hw.rx_queues + pfvf->hw.tx_queues; in otx2_cq_init()
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| H A D | nicvf_main.c | 406 int rx_queues = 0, tx_queues = 0; in nicvf_request_sqs() local 428 tx_queues = nic->tx_queues + nic->xdp_tx_queues; in nicvf_request_sqs() 429 if (tx_queues > MAX_SND_QUEUES_PER_QS) in nicvf_request_sqs() 430 tx_queues = tx_queues - MAX_SND_QUEUES_PER_QS; in nicvf_request_sqs() 448 if (tx_queues > MAX_SND_QUEUES_PER_QS) { in nicvf_request_sqs() 450 tx_queues -= MAX_SND_QUEUES_PER_QS; in nicvf_request_sqs() 452 nic->snicvf[sqs]->qs->sq_cnt = tx_queues; in nicvf_request_sqs() 453 tx_queues = 0; in nicvf_request_sqs() 466 nic->tx_queues, nic->rx_queues); in nicvf_request_sqs() 492 int tx_queues, int rx_queues) in nicvf_set_real_num_queues() argument [all …]
|
| H A D | nic.h | 308 u8 tx_queues; member 628 int tx_queues, int rx_queues);
|
| /linux/drivers/net/ethernet/pensando/ionic/ |
| H A D | ionic_stats.c | 237 u64 total = 0, tx_queues = MAX_Q(lif), rx_queues = MAX_Q(lif); in ionic_sw_stats_get_count() local 240 tx_queues += 1; in ionic_sw_stats_get_count() 248 total += tx_queues * IONIC_NUM_TX_STATS; in ionic_sw_stats_get_count()
|
| /linux/drivers/net/ethernet/broadcom/genet/ |
| H A D | bcmgenet.c | 50 (TOTAL_DESC - priv->hw_params->tx_queues * priv->hw_params->tx_bds_per_q) 916 for (i = 0; i <= priv->hw_params->tx_queues; i++) in bcmgenet_set_coalesce() 2042 } while (i <= priv->hw_params->tx_queues && netif_is_multiqueue(dev)); in bcmgenet_tx_reclaim_all() 2806 for (i = 0; i <= priv->hw_params->tx_queues; ++i) { in bcmgenet_enable_tx_napi() 2818 for (i = 0; i <= priv->hw_params->tx_queues; ++i) { in bcmgenet_disable_tx_napi() 2829 for (i = 0; i <= priv->hw_params->tx_queues; ++i) { in bcmgenet_fini_tx_napi() 2841 mask = (1 << (priv->hw_params->tx_queues + 1)) - 1; in bcmgenet_tdma_disable() 2906 for (i = 0; i <= priv->hw_params->tx_queues; i++) { in bcmgenet_init_tx_queues() 2921 ring_mask = (1 << (priv->hw_params->tx_queues + 1)) - 1; in bcmgenet_init_tx_queues() 3029 for (i = 0; i <= priv->hw_params->tx_queues; i++) { in bcmgenet_fini_dma() [all …]
|
| H A D | bcmgenet.h | 514 u8 tx_queues; member
|
| /linux/drivers/net/wireless/ralink/rt2x00/ |
| H A D | rt2x00queue.h | 505 &(__dev)->tx[(__dev)->ops->tx_queues]
|
| H A D | rt2x00queue.c | 1245 rt2x00dev->data_queues = 2 + rt2x00dev->ops->tx_queues + req_atim; in rt2x00queue_allocate() 1256 rt2x00dev->bcn = &queue[1 + rt2x00dev->ops->tx_queues]; in rt2x00queue_allocate() 1257 rt2x00dev->atim = req_atim ? &queue[2 + rt2x00dev->ops->tx_queues] : NULL; in rt2x00queue_allocate()
|
| H A D | rt2x00dev.c | 1135 rt2x00dev->hw->queues = rt2x00dev->ops->tx_queues; in rt2x00lib_probe_hw() 1169 roundup_pow_of_two(rt2x00dev->ops->tx_queues * in rt2x00lib_probe_hw()
|
| H A D | rt2800usb.c | 766 .tx_queues = NUM_TX_QUEUES,
|
| H A D | rt2500usb.c | 1896 .tx_queues = NUM_TX_QUEUES,
|
| H A D | rt2400pci.c | 1809 .tx_queues = NUM_TX_QUEUES,
|
| /linux/drivers/net/bonding/ |
| H A D | bond_main.c | 103 static int tx_queues = BOND_DEFAULT_TX_QUEUES; variable 129 module_param(tx_queues, int, 0); 130 MODULE_PARM_DESC(tx_queues, "Max number of transmit queues (default = 16)"); 6197 if (tx_queues < 1 || tx_queues > 255) { in bond_check_params() 6199 tx_queues, BOND_DEFAULT_TX_QUEUES); in bond_check_params() 6200 tx_queues = BOND_DEFAULT_TX_QUEUES; in bond_check_params() 6431 params->tx_queues = tx_queues; in bond_check_params() 6499 return tx_queues; in bond_get_num_tx_queues() 6517 bond_setup, tx_queues); in bond_create()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_virtchnl.c | 451 vqs.tx_queues = BIT(adapter->num_active_queues) - 1; in iavf_enable_queues() 452 vqs.rx_queues = vqs.tx_queues; in iavf_enable_queues() 476 vqs.tx_queues = BIT(adapter->num_active_queues) - 1; in iavf_disable_queues() 477 vqs.rx_queues = vqs.tx_queues; in iavf_disable_queues()
|
| /linux/include/net/ |
| H A D | bonding.h | 140 int tx_queues; member
|
| /linux/drivers/net/ethernet/qlogic/qlcnic/ |
| H A D | qlcnic_main.c | 2229 u8 tx_queues, u8 rx_queues) in qlcnic_set_real_num_queues() argument 2234 if (tx_queues) { in qlcnic_set_real_num_queues() 2235 err = netif_set_real_num_tx_queues(netdev, tx_queues); in qlcnic_set_real_num_queues() 2238 tx_queues); in qlcnic_set_real_num_queues()
|
| /linux/drivers/net/ethernet/ibm/ |
| H A D | ibmvnic.c | 3633 adapter->fallback.tx_queues = adapter->req_tx_queues; in wait_for_reset() 3656 adapter->desired.tx_queues = adapter->fallback.tx_queues; in wait_for_reset() 3837 adapter->desired.tx_queues = channels->tx_count; in ibmvnic_set_channels() 4582 if (adapter->desired.tx_queues) in send_request_cap() 4584 adapter->desired.tx_queues; in send_request_cap()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_virtchnl_pf.c | 2599 if ((!vqs->rx_queues && !vqs->tx_queues) || in i40e_vc_validate_vqs_bitmaps() 2601 vqs->tx_queues >= BIT(I40E_MAX_VF_QUEUES)) in i40e_vc_validate_vqs_bitmaps() 2651 if (i40e_ctrl_vf_tx_rings(pf->vsi[vf->lan_vsi_idx], vqs->tx_queues, in i40e_vc_enable_queues_msg() 2703 if (i40e_ctrl_vf_tx_rings(pf->vsi[vf->lan_vsi_idx], vqs->tx_queues, in i40e_vc_disable_queues_msg()
|