| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_txrx.c | 163 for (j = 0; j < txq_grp->num_txq; j++) in idpf_tx_desc_rel_all() 311 for (j = 0; j < vport->txq_grps[i].num_txq; j++) { in idpf_tx_desc_alloc_all() 1073 if (qv->num_txq) { in idpf_qvec_ena_irq() 1095 if (qv->num_txq || qv->num_rxq) in idpf_qvec_ena_irq() 1119 num = qv->num_rxq + qv->num_bufq + qv->num_txq + qv->num_complq; in idpf_vector_to_queue_set() 1140 for (u32 i = 0; i < qv->num_txq; i++) { in idpf_vector_to_queue_set() 1313 for (j = 0; j < txq_grp->num_txq; j++) { in idpf_txq_group_rel() 1437 vport->txqs = kcalloc(vport->num_txq, sizeof(*vport->txqs), in idpf_vport_init_fast_path_txqs() 1446 for (j = 0; j < tx_grp->num_txq; j++, k++) { in idpf_vport_init_fast_path_txqs() 1473 vport->num_txq in idpf_vport_init_num_qs() 1661 idpf_vport_calc_numq_per_grp(struct idpf_vport * vport,u16 * num_txq,u16 * num_rxq) idpf_vport_calc_numq_per_grp() argument 1699 idpf_txq_group_alloc(struct idpf_vport * vport,u16 num_txq) idpf_txq_group_alloc() argument 1922 u16 num_txq, num_rxq; idpf_vport_queue_grp_alloc_all() local 4390 u16 num_txq; idpf_vport_intr_map_vector_to_qs() local [all...] |
| H A D | idpf_ethtool.c | 527 u16 num_txq, num_rxq; in idpf_get_channels() local 532 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_get_channels() 535 combined = min(num_txq, num_rxq); in idpf_get_channels() 548 ch->tx_count = num_txq - combined; in idpf_get_channels() 566 u16 num_txq, num_rxq; in idpf_set_channels() local 582 num_txq = vport_config->user_config.num_req_tx_qs; in idpf_set_channels() 606 if (num_req_tx_q == num_txq && num_req_rx_q == num_rxq) in idpf_set_channels() 615 vport_config->user_config.num_req_tx_qs = num_txq; in idpf_set_channels() 1168 for (j = 0; j < txq_grp->num_txq; j++) { in idpf_collect_queue_stats() 1236 for (j = 0; j < txq_grp->num_txq; j++, total++) { in idpf_get_ethtool_stats() [all …]
|
| H A D | idpf_virtchnl.c | 848 qs = idpf_alloc_queue_set(vport, vport->num_txq); in idpf_wait_for_marker_event() 1370 for (j = 0; j < tx_qgrp->num_txq && k < num_regs; j++, k++) in __idpf_queue_reg_init() 1444 if (num_regs < vport->num_txq) { in idpf_queue_reg_init() 1451 if (num_regs < vport->num_txq) { in idpf_queue_reg_init() 1831 u32 totqs = vport->num_txq + vport->num_complq; in idpf_send_config_tx_queues_msg() 1842 for (u32 j = 0; j < tx_qgrp->num_txq; j++) { in idpf_send_config_tx_queues_msg() 2158 u32 num_txq, num_q, k = 0; in idpf_send_ena_dis_queues_msg() local 2161 num_txq = vport->num_txq + vport->num_complq; in idpf_send_ena_dis_queues_msg() 2162 num_q = num_txq + vport->num_rxq + vport->num_bufq; in idpf_send_ena_dis_queues_msg() 2173 for (u32 j = 0; j < tx_qgrp->num_txq; j++) { in idpf_send_ena_dis_queues_msg() [all …]
|
| H A D | idpf_txrx.h | 397 u16 num_txq; member 968 u16 num_txq; member
|
| H A D | idpf.h | 346 u16 num_txq; member
|
| H A D | idpf_lib.c | 1420 int err, txq = vport->num_txq - vport->num_xdp_txq; in idpf_set_real_num_queues() 2062 err = idpf_send_add_queues_msg(vport, new_vport->num_txq, in idpf_initiate_soft_reset() 2091 idpf_send_add_queues_msg(vport, vport->num_txq, vport->num_complq, in idpf_initiate_soft_reset()
|
| /linux/drivers/net/ethernet/atheros/alx/ |
| H A D | alx.h | 121 int num_txq; member
|
| /linux/drivers/net/ethernet/brocade/bna/ |
| H A D | bna_enet.c | 167 ioceth->attr.num_txq = ntohl(rsp->max_cfg); in bna_bfi_attr_get_rsp() 1679 ioceth->attr.num_txq = BFI_ENET_DEF_TXQ; in bna_attr_init() 1906 attr->num_txq * sizeof(struct bna_tx); in bna_mod_res_req() 1914 attr->num_txq * sizeof(struct bna_txq); in bna_mod_res_req() 2026 bna_num_txq_set(struct bna *bna, int num_txq) in bna_num_txq_set() argument 2029 (num_txq <= bna->ioceth.attr.num_txq)) { in bna_num_txq_set() 2030 bna->ioceth.attr.num_txq = num_txq; in bna_num_txq_set()
|
| H A D | bna_types.h | 306 int num_txq; member 476 int num_txq; member 505 int num_txq; member
|
| H A D | bna_tx_rx.c | 3097 cfg_req->num_queues = tx->num_txq; in bna_bfi_tx_enet_start() 3098 for (i = 0; i < tx->num_txq; i++) { in bna_bfi_tx_enet_start() 3282 i < tx->num_txq; i++, txq = list_next_entry(txq, qe)) { in bna_bfi_tx_enet_start_rsp() 3316 bna_tx_res_req(int num_txq, int txq_depth, struct bna_res_info *res_info) in bna_tx_res_req() argument 3326 mem_info->num = num_txq; in bna_tx_res_req() 3336 mem_info->num = num_txq; in bna_tx_res_req() 3342 mem_info->num = num_txq; in bna_tx_res_req() 3348 mem_info->num = num_txq; in bna_tx_res_req() 3354 mem_info->num = num_txq; in bna_tx_res_req() 3359 res_info[BNA_TX_RES_INTR_T_TXCMPL].res_u.intr_info.num = num_txq; in bna_tx_res_req() [all …]
|
| H A D | bna.h | 265 int bna_num_txq_set(struct bna *bna, int num_txq); 306 void bna_tx_res_req(int num_txq, int txq_depth,
|
| H A D | bnad.c | 1968 tx_config->num_txq = bnad->num_txq_per_tx; in bnad_setup_tx() 3682 bnad_q_num_adjust(bnad, bna_attr(bna)->num_txq - 1, in bnad_pci_probe()
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_type_e610.h | 850 u32 num_txq; /* Number/Total Tx queues */ member
|
| H A D | ixgbe_e610.c | 582 caps->num_txq = number; in ixgbe_parse_e610_caps()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_common.c | 2445 caps->num_txq = number; in ice_parse_common_caps() 2447 ice_debug(hw, ICE_DBG_INIT, "%s: num_txq = %d\n", prefix, in ice_parse_common_caps() 2448 caps->num_txq); in ice_parse_common_caps() 3243 func_caps->common_cap.num_txq = 1; in ice_set_safe_mode_caps() 3273 dev_caps->common_cap.num_txq = num_funcs; in ice_set_safe_mode_caps()
|
| /linux/net/mac80211/ |
| H A D | tx.c | 4189 unsigned int num_txq = 0; in ieee80211_txq_schedule_airtime_check() local 4197 num_txq++; in ieee80211_txq_schedule_airtime_check() 4199 aql_limit = (num_txq - 1) * local->aql_txq_limit_low[ac] / 2 + in ieee80211_txq_schedule_airtime_check()
|