| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 163 for (unsigned int j = 0; j < txq_grp->num_txq; j++) in idpf_tx_desc_rel_all() 311 for (unsigned int j = 0; j < rsrc->txq_grps[i].num_txq; j++) { in idpf_tx_desc_alloc_all() 1076 if (qv->num_txq) { in idpf_qvec_ena_irq() 1098 if (qv->num_txq || qv->num_rxq) in idpf_qvec_ena_irq() 1123 num = qv->num_rxq + qv->num_bufq + qv->num_txq + qv->num_complq; in idpf_vector_to_queue_set() 1145 for (u32 i = 0; i < qv->num_txq; i++) { in idpf_vector_to_queue_set() 1316 for (unsigned int j = 0; j < txq_grp->num_txq; j++) { in idpf_txq_group_rel() 1446 vport->txqs = kzalloc_objs(*vport->txqs, rsrc->num_txq); in idpf_vport_init_fast_path_txqs() 1450 vport->num_txq = rsrc->num_txq; in idpf_vport_init_fast_path_txqs() 1454 for (unsigned int j = 0; j < tx_grp->num_txq; j++, k++) { in idpf_vport_init_fast_path_txqs() [all …]
|
| H A D | idpf_ethtool.c | 524 u16 num_txq, num_rxq; in idpf_get_channels() local 529 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_get_channels() 532 combined = min(num_txq, num_rxq); in idpf_get_channels() 545 ch->tx_count = num_txq - combined; in idpf_get_channels() 563 u16 num_txq, num_rxq; in idpf_set_channels() local 579 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_set_channels() 603 if (num_req_tx_q == num_txq && num_req_rx_q == num_rxq) in idpf_set_channels() 612 vport_config->user_config.num_req_tx_qs = num_txq; in idpf_set_channels() 1167 for (unsigned int j = 0; j < txq_grp->num_txq; j++) { in idpf_collect_queue_stats() 1236 for (unsigned int j = 0; j < txq_grp->num_txq; j++, total++) { in idpf_get_ethtool_stats() [all …]
|
| H A D | idpf_virtchnl.c | 861 vport->vport_id, vport->num_txq); in idpf_wait_for_marker_event() 1422 for (j = 0; j < tx_qgrp->num_txq && k < num_regs; j++, k++) in __idpf_queue_reg_init() 1486 if (num_regs < rsrc->num_txq) { in idpf_queue_reg_init() 1493 if (num_regs < rsrc->num_txq) { in idpf_queue_reg_init() 1878 u32 totqs = rsrc->num_txq + rsrc->num_complq; in idpf_send_config_tx_queues_msg() 1889 for (u32 j = 0; j < tx_qgrp->num_txq; j++) { in idpf_send_config_tx_queues_msg() 2213 u32 num_txq, num_q, k = 0; in idpf_send_ena_dis_queues_msg() local 2216 num_txq = rsrc->num_txq + rsrc->num_complq; in idpf_send_ena_dis_queues_msg() 2217 num_q = num_txq + rsrc->num_rxq + rsrc->num_bufq; in idpf_send_ena_dis_queues_msg() 2228 for (u32 j = 0; j < tx_qgrp->num_txq; j++) { in idpf_send_ena_dis_queues_msg() [all …]
|
| H A D | xdp.c | 179 for (u32 i = sqs; i < vport->num_txq; i++) { in idpf_xdpsqs_get() 216 for (u32 i = sqs; i < vport->num_txq; i++) { in idpf_xdpsqs_put() 476 if (!vport->num_xdp_txq && vport->num_txq == cfg->max_q.max_txq) { in idpf_xdp_setup_prog()
|
| H A D | idpf.h | 336 u16 num_txq; member 385 u16 num_txq; member
|
| H A D | idpf_txrx.h | 399 u16 num_txq; member 972 u16 num_txq; member
|
| H A D | idpf_lib.c | 1434 int err, txq = vport->dflt_qv_rsrc.num_txq - vport->num_xdp_txq; in idpf_set_real_num_queues()
|
| /linux/drivers/net/ethernet/atheros/alx/ |
| H A D | main.c | 146 if (r_idx >= alx->num_txq) in alx_tx_queue_mapping() 147 r_idx = r_idx % alx->num_txq; in alx_tx_queue_mapping() 537 for (i = 0; i < alx->num_txq; i++) in alx_free_buffers() 659 alx->num_txq + in alx_alloc_rings() 672 for (i = 0; i < alx->num_txq; i++) { in alx_alloc_rings() 695 for (i = 0; i < alx->num_txq; i++) in alx_free_rings() 759 for (i = 0; i < alx->num_txq; i++) { in alx_alloc_napis() 814 for (i = 0, vector = 1; i < alx->num_txq; i++, vector++) { in alx_config_vector_mapping() 831 int err, num_vec, num_txq, num_rxq; in alx_enable_msix() local 833 num_txq = min_t(int, num_online_cpus(), ALX_MAX_TX_QUEUES); in alx_enable_msix() [all …]
|
| H A D | alx.h | 121 int num_txq; member
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_sriov.c | 373 u16 num_msix_per_vf, num_txq, num_rxq, avail_qs; in ice_set_per_vf_res() local 400 num_txq = min_t(u16, num_msix_per_vf - ICE_NONQ_VECS_VF, in ice_set_per_vf_res() 404 num_txq = 0; in ice_set_per_vf_res() 405 else if (num_txq > avail_qs) in ice_set_per_vf_res() 406 num_txq = rounddown_pow_of_two(avail_qs); in ice_set_per_vf_res() 416 if (num_txq < ICE_MIN_QS_PER_VF || num_rxq < ICE_MIN_QS_PER_VF) { in ice_set_per_vf_res() 423 pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq); in ice_set_per_vf_res()
|
| H A D | ice_lib.c | 187 vsi->num_txq = vsi->req_txq; in ice_vsi_set_num_qs() 592 vsi->num_txq = ch->num_txq; in ice_vsi_alloc_def() 1084 vsi->num_txq = tx_count; in ice_vsi_setup_q_map() 1087 if (vsi->type == ICE_VSI_VF && vsi->num_txq != vsi->num_rxq) { in ice_vsi_setup_q_map() 1092 vsi->num_txq = vsi->num_rxq; in ice_vsi_setup_q_map() 1989 if (vsi->num_txq > ICE_LAN_TXQ_MAX_QDIS) in ice_vsi_stop_tx_rings() 2020 return ice_vsi_stop_tx_rings(vsi, rst_src, rel_vmvf_num, vsi->tx_rings, vsi->num_txq); in ice_vsi_stop_lan_tx_rings() 2282 if (!vsi->alloc_txq && vsi->num_txq) in ice_vsi_cfg_tc_lan() 2283 max_txqs[i] = vsi->num_txq; in ice_vsi_cfg_tc_lan() 2920 if (i < vsi->num_txq) in ice_vsi_rebuild_get_coalesce() [all …]
|
| H A D | ice.h | 166 for ((i) = 0; (i) < (vsi)->num_txq; (i)++) 224 u16 num_txq; member 405 u16 num_txq; /* Used Tx queues */ member
|
| H A D | ice_base.c | 935 tx_rings_rem = vsi->num_txq; in ice_vsi_map_rings_to_vectors() 949 q_base = vsi->num_txq - tx_rings_rem; in ice_vsi_map_rings_to_vectors() 1181 return ice_vsi_cfg_txqs(vsi, vsi->tx_rings, vsi->num_txq); in ice_vsi_cfg_lan_txqs() 1445 if (q_idx >= vsi->num_rxq || q_idx >= vsi->num_txq) in ice_qp_dis()
|
| H A D | ice_xsk.c | 203 if (qid >= vsi->num_rxq || qid >= vsi->num_txq) { in ice_xsk_pool_setup() 854 if (queue_id >= vsi->num_txq || queue_id >= vsi->num_rxq) in ice_xsk_wakeup()
|
| H A D | ice_main.c | 2802 max_txqs[i] = vsi->num_txq + vsi->num_xdp_txq; in ice_prepare_xdp_rings() 2905 max_txqs[i] = vsi->num_txq; in ice_destroy_xdp_rings() 4026 pf->max_pf_txqs = func_caps->common_cap.num_txq; in ice_set_pf_caps() 6976 vsi->num_txq); in ice_update_vsi_ring_stats() 7223 if (!vsi->num_txq || !vsi->num_rxq) in ice_get_stats64() 7410 if (!vsi->num_txq) { in ice_vsi_setup_tx_rings() 7558 err = netif_set_real_num_tx_queues(vsi->netdev, vsi->num_txq); in ice_vsi_open() 8557 if (vsi->num_txq < in ice_validate_mqprio_qopt() 8685 for (i = 0; i < ch->num_txq; i++) { in ice_chnl_cfg_res() 8726 if (ch->num_txq || ch->num_rxq) in ice_chnl_cfg_res() [all …]
|
| H A D | ice_ethtool.c | 3272 tx_rings = kzalloc_objs(*tx_rings, vsi->num_txq); in ice_set_ringparam() 3782 return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_txq); in ice_get_max_txq() 3835 ch->tx_count = vsi->num_txq - ch->combined_count; in ice_get_channels() 4087 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_get_q_coalesce() 4098 } else if (q_num < vsi->num_txq) { in ice_get_q_coalesce() 4255 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_set_q_coalesce() 4270 } else if (q_num < vsi->num_txq) { in ice_set_q_coalesce() 4339 if (v_idx >= vsi->num_rxq && v_idx >= vsi->num_txq) in __ice_set_coalesce()
|
| H A D | ice_common.c | 2414 caps->num_txq = number; in ice_parse_common_caps() 2417 caps->num_txq); in ice_parse_common_caps() 3212 func_caps->common_cap.num_txq = 1; in ice_set_safe_mode_caps() 3242 dev_caps->common_cap.num_txq = num_funcs; in ice_set_safe_mode_caps()
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_enet.c | 167 ioceth->attr.num_txq = ntohl(rsp->max_cfg); in bna_bfi_attr_get_rsp() 1679 ioceth->attr.num_txq = BFI_ENET_DEF_TXQ; in bna_attr_init() 1906 attr->num_txq * sizeof(struct bna_tx); in bna_mod_res_req() 1914 attr->num_txq * sizeof(struct bna_txq); in bna_mod_res_req() 2026 bna_num_txq_set(struct bna *bna, int num_txq) in bna_num_txq_set() argument 2029 (num_txq <= bna->ioceth.attr.num_txq)) { in bna_num_txq_set() 2030 bna->ioceth.attr.num_txq = num_txq; in bna_num_txq_set()
|
| H A D | bna_types.h | 306 int num_txq; member 476 int num_txq; member 505 int num_txq; member
|
| H A D | bna_tx_rx.c | 3097 cfg_req->num_queues = tx->num_txq; in bna_bfi_tx_enet_start() 3098 for (i = 0; i < tx->num_txq; i++) { in bna_bfi_tx_enet_start() 3282 i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) { in bna_bfi_tx_enet_start_rsp() 3316 bna_tx_res_req(int num_txq, int txq_depth, struct bna_res_info *res_info) in bna_tx_res_req() argument 3326 mem_info->num = num_txq; in bna_tx_res_req() 3336 mem_info->num = num_txq; in bna_tx_res_req() 3342 mem_info->num = num_txq; in bna_tx_res_req() 3348 mem_info->num = num_txq; in bna_tx_res_req() 3354 mem_info->num = num_txq; in bna_tx_res_req() 3359 res_info[BNA_TX_RES_INTR_T_TXCMPL].res_u.intr_info.num = num_txq; in bna_tx_res_req() [all …]
|
| H A D | bna.h | 265 int bna_num_txq_set(struct bna *bna, int num_txq); 306 void bna_tx_res_req(int num_txq, int txq_depth,
|
| H A D | bnad.c | 1965 tx_config->num_txq = bnad->num_txq_per_tx; in bnad_setup_tx() 3679 bnad_q_num_adjust(bnad, bna_attr(bna)->num_txq - 1, in bnad_pci_probe()
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_type_e610.h | 850 u32 num_txq; /* Number/Total Tx queues */ member
|
| /linux/drivers/net/ethernet/intel/ice/virt/ |
| H A D | virtchnl.c | 334 vfres->num_queue_pairs = vsi->num_txq; in ice_vc_get_vf_res_msg() 342 vfres->vsi_res[0].num_queue_pairs = vsi->num_txq; in ice_vc_get_vf_res_msg()
|
| /linux/net/mac80211/ |
| H A D | tx.c | 4192 unsigned int num_txq = 0; in ieee80211_txq_schedule_airtime_check() local 4200 num_txq++; in ieee80211_txq_schedule_airtime_check() 4202 aql_limit = (num_txq - 1) * local->aql_txq_limit_low[ac] / 2 + in ieee80211_txq_schedule_airtime_check()
|