| /linux/drivers/net/ethernet/intel/idpf/ |
| H A D | idpf_ethtool.c | 17 u32 num_rxq; in idpf_get_rx_ring_count() local 21 num_rxq = vport->dflt_qv_rsrc.num_rxq; in idpf_get_rx_ring_count() 24 return num_rxq; in idpf_get_rx_ring_count() 180 u16 num_rxq; in idpf_add_flow_steer() local 186 num_rxq = user_config->num_req_rx_qs; in idpf_add_flow_steer() 200 if (q_index >= num_rxq) in idpf_add_flow_steer() 524 u16 num_txq, num_rxq; in idpf_get_channels() local 530 num_rxq = vport_config->user_config.num_req_rx_qs; in idpf_get_channels() 532 combined = min(num_txq, num_rxq); in idpf_get_channels() 544 ch->rx_count = num_rxq - combined; in idpf_get_channels() [all …]
|
| H A D | idpf_txrx.c | 499 u16 num_rxq; in idpf_rx_desc_rel_all() local 508 for (unsigned int j = 0; j < rx_qgrp->singleq.num_rxq; j++) in idpf_rx_desc_rel_all() 514 num_rxq = rx_qgrp->splitq.num_rxq_sets; in idpf_rx_desc_rel_all() 515 for (unsigned int j = 0; j < num_rxq; j++) in idpf_rx_desc_rel_all() 801 int num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rx_bufs_init_all() local 803 for (unsigned int j = 0; j < num_rxq; j++) { in idpf_rx_bufs_init_all() 910 u16 num_rxq; in idpf_rx_desc_alloc_all() local 916 num_rxq = rx_qgrp->splitq.num_rxq_sets; in idpf_rx_desc_alloc_all() 918 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rx_desc_alloc_all() 920 for (unsigned int j = 0; j < num_rxq; j++) { in idpf_rx_desc_alloc_all() [all …]
|
| H A D | idpf_virtchnl.c | 1430 u16 num_rxq = rx_qgrp->singleq.num_rxq; in __idpf_queue_reg_init() local 1432 for (j = 0; j < num_rxq && k < num_regs; j++, k++) { in __idpf_queue_reg_init() 1520 if (num_regs < rsrc->num_rxq) { in idpf_queue_reg_init() 1527 if (num_regs < rsrc->num_rxq) { in idpf_queue_reg_init() 2067 u32 totqs = rsrc->num_rxq + rsrc->num_bufq; in idpf_send_config_rx_queues_msg() 2077 u32 num_rxq; in idpf_send_config_rx_queues_msg() local 2080 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_send_config_rx_queues_msg() 2089 num_rxq = rx_qgrp->splitq.num_rxq_sets; in idpf_send_config_rx_queues_msg() 2092 for (u32 j = 0; j < num_rxq; j++) { in idpf_send_config_rx_queues_msg() 2217 num_q = num_txq + rsrc->num_rxq + rsrc->num_bufq; in idpf_send_ena_dis_queues_msg() [all …]
|
| H A D | xdp.c | 21 u32 num_rxq; in idpf_rxq_for_each() local 24 num_rxq = rx_qgrp->splitq.num_rxq_sets; in idpf_rxq_for_each() 26 num_rxq = rx_qgrp->singleq.num_rxq; in idpf_rxq_for_each() 28 for (u32 j = 0; j < num_rxq; j++) { in idpf_rxq_for_each()
|
| H A D | idpf_ptp.c | 344 for (i = 0; i < grp->singleq.num_rxq; i++) { in idpf_ptp_update_phctime_rxq_grp() 701 u16 j, num_rxq; in idpf_ptp_set_rx_tstamp() local 704 num_rxq = grp->splitq.num_rxq_sets; in idpf_ptp_set_rx_tstamp() 706 num_rxq = grp->singleq.num_rxq; in idpf_ptp_set_rx_tstamp() 708 for (j = 0; j < num_rxq; j++) { in idpf_ptp_set_rx_tstamp()
|
| H A D | idpf_txrx.h | 398 u16 num_rxq; member 943 u16 num_rxq; member
|
| H A D | idpf.h | 344 u16 num_rxq; member
|
| H A D | idpf_lib.c | 1437 vport->dflt_qv_rsrc.num_rxq); in idpf_set_real_num_queues() 1477 for (unsigned int j = 0; j < grp->singleq.num_rxq; j++) { in idpf_rx_init_buf_tail()
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_lib.c | 200 vsi->num_rxq = vsi->req_rxq; in ice_vsi_set_num_qs() 591 vsi->num_rxq = ch->num_rxq; in ice_vsi_alloc_def() 910 vsi->rss_size = min_t(u16, vsi->num_rxq, max_rss_size); in ice_vsi_set_rss_params() 1085 vsi->num_rxq = rx_count; in ice_vsi_setup_q_map() 1087 if (vsi->type == ICE_VSI_VF && vsi->num_txq != vsi->num_rxq) { in ice_vsi_setup_q_map() 1092 vsi->num_txq = vsi->num_rxq; in ice_vsi_setup_q_map() 1102 ctxt->info.q_mapping[1] = cpu_to_le16(vsi->num_rxq); in ice_vsi_setup_q_map() 1195 qcount = vsi->num_rxq; in ice_chnl_vsi_setup_q_map() 1511 vsi->rss_size = min_t(u16, vsi->rss_size, vsi->num_rxq); in ice_vsi_cfg_rss_lut_key() 1520 vsi->orig_rss_size <= vsi->num_rxq) { in ice_vsi_cfg_rss_lut_key() [all …]
|
| H A D | ice_sriov.c | 373 u16 num_msix_per_vf, num_txq, num_rxq, avail_qs; in ice_set_per_vf_res() local 408 num_rxq = min_t(u16, num_msix_per_vf - ICE_NONQ_VECS_VF, in ice_set_per_vf_res() 412 num_rxq = 0; in ice_set_per_vf_res() 413 else if (num_rxq > avail_qs) in ice_set_per_vf_res() 414 num_rxq = rounddown_pow_of_two(avail_qs); in ice_set_per_vf_res() 416 if (num_txq < ICE_MIN_QS_PER_VF || num_rxq < ICE_MIN_QS_PER_VF) { in ice_set_per_vf_res() 423 pf->vfs.num_qps_per = min_t(int, num_txq, num_rxq); in ice_set_per_vf_res()
|
| H A D | ice.h | 172 for ((i) = 0; (i) < (vsi)->num_rxq; (i)++) 223 u16 num_rxq; member 407 u16 num_rxq; /* Used Rx queues */ member
|
| H A D | ice_base.c | 756 if (q_idx >= vsi->num_rxq) in ice_vsi_cfg_single_rxq() 936 rx_rings_rem = vsi->num_rxq; in ice_vsi_map_rings_to_vectors() 966 q_base = vsi->num_rxq - rx_rings_rem; in ice_vsi_map_rings_to_vectors() 1445 if (q_idx >= vsi->num_rxq || q_idx >= vsi->num_txq) in ice_qp_dis()
|
| H A D | ice_xsk.c | 203 if (qid >= vsi->num_rxq || qid >= vsi->num_txq) { in ice_xsk_pool_setup() 854 if (queue_id >= vsi->num_txq || queue_id >= vsi->num_rxq) in ice_xsk_wakeup()
|
| H A D | ice_ethtool.c | 3332 rx_rings = kzalloc_objs(*rx_rings, vsi->num_rxq); in ice_set_ringparam() 3791 return min(num_online_cpus(), pf->hw.func_caps.common_cap.num_rxq); in ice_get_max_rxq() 3834 ch->rx_count = vsi->num_rxq - ch->combined_count; in ice_get_channels() 4087 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_get_q_coalesce() 4094 } else if (q_num < vsi->num_rxq) { in ice_get_q_coalesce() 4255 if (q_num < vsi->num_rxq && q_num < vsi->num_txq) { in ice_set_q_coalesce() 4265 } else if (q_num < vsi->num_rxq) { in ice_set_q_coalesce() 4339 if (v_idx >= vsi->num_rxq && v_idx >= vsi->num_txq) in __ice_set_coalesce()
|
| H A D | ice_main.c | 4027 pf->max_pf_rxqs = func_caps->common_cap.num_rxq; in ice_set_pf_caps() 6980 vsi->num_rxq); in ice_update_vsi_ring_stats() 7223 if (!vsi->num_txq || !vsi->num_rxq) in ice_get_stats64() 7442 if (!vsi->num_rxq) { in ice_vsi_setup_rx_rings() 7562 err = netif_set_real_num_rx_queues(vsi->netdev, vsi->num_rxq); in ice_vsi_open() 8554 if (vsi->num_rxq < in ice_validate_mqprio_qopt() 8726 if (ch->num_txq || ch->num_rxq) in ice_chnl_cfg_res() 8780 vsi->next_base_q = vsi->next_base_q + ch->num_rxq; in ice_setup_hw_channel() 8782 ch->num_rxq); in ice_setup_hw_channel() 8857 if (!ch->num_txq || !ch->num_rxq) { in ice_create_q_channel() [all …]
|
| H A D | ice_tc_lib.c | 1066 return queue < vsi->num_rxq ? vsi->rx_rings[queue] : NULL; in ice_locate_rx_ring_using_queue() 2043 if (queue >= vsi->num_rxq) { in ice_tc_forward_to_queue()
|
| H A D | ice_common.c | 2406 caps->num_rxq = number; in ice_parse_common_caps() 2409 caps->num_rxq); in ice_parse_common_caps() 3211 func_caps->common_cap.num_rxq = 1; in ice_set_safe_mode_caps() 3241 dev_caps->common_cap.num_rxq = num_funcs; in ice_set_safe_mode_caps()
|
| /linux/drivers/net/ethernet/atheros/alx/ |
| H A D | alx.h | 122 int num_rxq; member
|
| H A D | main.c | 831 int err, num_vec, num_txq, num_rxq; in alx_enable_msix() local 834 num_rxq = 1; in alx_enable_msix() 835 num_vec = max_t(int, num_txq, num_rxq) + 1; in alx_enable_msix() 847 alx->num_rxq = num_rxq; in alx_enable_msix() 911 alx->num_rxq = 1; in alx_init_intr() 1236 netif_set_real_num_rx_queues(alx->dev, alx->num_rxq); in __alx_open()
|
| /linux/drivers/net/ethernet/intel/ixgbe/ |
| H A D | ixgbe_type_e610.h | 848 u32 num_rxq; /* Number/Total Rx queues */ member
|
| H A D | ixgbe_e610.c | 578 caps->num_rxq = number; in ixgbe_parse_e610_caps()
|