| /linux/include/net/ |
| H A D | red.h | 146 int qcount; /* Number of packets since last random member 167 v->qcount = -1; in red_set_vars() 286 v->qcount = -1; in red_restart() 389 return !(((qavg - p->qth_min) >> p->Wlog) * v->qcount < v->qR); in red_mark_probability() 420 v->qcount = -1; in red_action() 424 if (++v->qcount) { in red_action() 426 v->qcount = 0; in red_action() 436 v->qcount = -1; in red_action()
|
| /linux/drivers/net/ethernet/chelsio/cxgb4/ |
| H A D | cxgb4_tc_mqprio.c | 14 u32 speed, qcount = 0, qoffset = 0; in cxgb4_mqprio_validate() local 49 qcount += mqprio->qopt.count[i]; in cxgb4_mqprio_validate() 81 if (qoffset >= adap->tids.neotids || qcount > adap->tids.neotids) in cxgb4_mqprio_validate() 428 u32 qoffset, qcount, tot_qcount, qid, hwqid; in cxgb4_mqprio_enable_offload() local 443 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 444 for (j = 0; j < qcount; j++) { in cxgb4_mqprio_enable_offload() 483 qcount = mqprio->qopt.count[i]; in cxgb4_mqprio_enable_offload() 484 if (qcount) { in cxgb4_mqprio_enable_offload() 487 qcount = pi->nqsets; in cxgb4_mqprio_enable_offload() 491 ret = netdev_set_tc_queue(dev, i, qcount, qoffset); in cxgb4_mqprio_enable_offload() [all …]
|
| /linux/drivers/net/ethernet/intel/ice/ |
| H A D | ice_dcb_lib.c | 220 u16 qoffset, qcount; in ice_vsi_cfg_dcb_rings() local 241 qcount = vsi->tc_cfg.tc_info[n].qcount_tx; in ice_vsi_cfg_dcb_rings() 242 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 245 qcount = vsi->tc_cfg.tc_info[n].qcount_rx; in ice_vsi_cfg_dcb_rings() 246 for (i = qoffset; i < (qoffset + qcount); i++) in ice_vsi_cfg_dcb_rings() 263 qcount = vsi->mqprio_qopt.qopt.count[n]; in ice_vsi_cfg_dcb_rings() 264 for (i = qoffset; i < (qoffset + qcount); i++) { in ice_vsi_cfg_dcb_rings()
|
| H A D | ice_main.c | 8463 int qcount = mqprio_qopt->qopt.count[i]; in ice_validate_mqprio_qopt() local 8466 if (!qcount) in ice_validate_mqprio_qopt() 8469 if (is_power_of_2(qcount)) { in ice_validate_mqprio_qopt() 8471 qcount > non_power_of_2_qcount) { in ice_validate_mqprio_qopt() 8473 qcount, non_power_of_2_qcount); in ice_validate_mqprio_qopt() 8476 if (qcount > max_rss_q_cnt) in ice_validate_mqprio_qopt() 8477 max_rss_q_cnt = qcount; in ice_validate_mqprio_qopt() 8480 qcount != non_power_of_2_qcount) { in ice_validate_mqprio_qopt() 8482 qcount, non_power_of_2_qcount); in ice_validate_mqprio_qopt() 8485 if (qcount < max_rss_q_cnt) { in ice_validate_mqprio_qopt() [all …]
|
| H A D | ice_lib.c | 1193 u16 qcount, qmap; in ice_chnl_vsi_setup_q_map() local 1197 qcount = vsi->num_rxq; in ice_chnl_vsi_setup_q_map() 1199 pow = order_base_2(qcount); in ice_chnl_vsi_setup_q_map() 1206 ctxt->info.q_mapping[1] = cpu_to_le16(qcount); in ice_chnl_vsi_setup_q_map()
|
| H A D | ice_tc_lib.c | 1046 int qcount = vsi->mqprio_qopt.qopt.count[tc]; in ice_locate_vsi_using_queue() local 1049 if (queue >= offset && queue < offset + qcount) { in ice_locate_vsi_using_queue()
|
| H A D | ice_ethtool.c | 3615 u16 qcount, offset; in ice_get_rxfh() local 3624 qcount = vsi->mqprio_qopt.qopt.count[0]; in ice_get_rxfh() 3644 rxfh->indir[i] = offset + lut[i] % qcount; in ice_get_rxfh()
|
| /linux/net/sched/ |
| H A D | sch_choke.c | 226 q->vars.qcount = -1; in choke_enqueue() 239 q->vars.qcount = -1; in choke_enqueue() 249 } else if (++q->vars.qcount) { in choke_enqueue() 251 q->vars.qcount = 0; in choke_enqueue()
|
| /linux/drivers/net/ethernet/intel/i40e/ |
| H A D | i40e_main.c | 1943 u16 qcount = 0, max_qcount, qmap, sections = 0; in i40e_vsi_setup_queue_map_mqprio() local 1968 qcount = vsi->mqprio_qopt.qopt.count[i]; in i40e_vsi_setup_queue_map_mqprio() 1969 if (qcount > max_qcount) in i40e_vsi_setup_queue_map_mqprio() 1970 max_qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1972 vsi->tc_config.tc_info[i].qcount = qcount; in i40e_vsi_setup_queue_map_mqprio() 1980 vsi->tc_config.tc_info[i].qcount = 1; in i40e_vsi_setup_queue_map_mqprio() 1986 vsi->num_queue_pairs = offset + qcount; in i40e_vsi_setup_queue_map_mqprio() 2037 u16 qcount; in i40e_vsi_setup_queue_map() local 2108 qcount = min_t(int, pf->alloc_rss_size, in i40e_vsi_setup_queue_map() 2117 qcount = num_tc_qps; in i40e_vsi_setup_queue_map() [all …]
|
| H A D | i40e_txrx.c | 3745 u16 qcount; in i40e_lan_select_queue() local 3763 qcount = vsi->tc_config.tc_info[tclass].qcount; in i40e_lan_select_queue() 3764 hash = i40e_swdcb_skb_tx_hash(netdev, skb, qcount); in i40e_lan_select_queue()
|
| H A D | i40e.h | 335 u16 qcount; /* Total Queues */ member
|
| H A D | i40e_debugfs.c | 416 vsi->tc_config.tc_info[i].qcount, in i40e_dbg_dump_vsi_seid()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_pf.c | 3130 int err, qcount, qos_txqs; in otx2_probe() local 3156 qcount = min_t(int, num_online_cpus(), OTX2_MAX_CQ_CNT); in otx2_probe() 3157 qos_txqs = min_t(int, qcount, OTX2_QOS_MAX_LEAF_NODES); in otx2_probe() 3159 netdev = alloc_etherdev_mqs(sizeof(*pf), qcount + qos_txqs, qcount); in otx2_probe() 3174 hw->rx_queues = qcount; in otx2_probe() 3175 hw->tx_queues = qcount; in otx2_probe() 3176 hw->non_qos_queues = qcount; in otx2_probe() 3177 hw->max_queues = qcount; in otx2_probe() 3293 pf->af_xdp_zc_qidx = bitmap_zalloc(qcount, GFP_KERNEL); in otx2_probe()
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| H A D | nicvf_queues.c | 383 int tail, qcount; in nicvf_refill_rbdr() local 400 qcount = nicvf_queue_reg_read(nic, NIC_QSET_RBDR_0_1_STATUS0, rbdr_idx); in nicvf_refill_rbdr() 401 qcount &= 0x7FFFF; in nicvf_refill_rbdr() 403 if (qcount >= (qs->rbdr_len - 1)) in nicvf_refill_rbdr() 406 refill_rb_cnt = qs->rbdr_len - qcount - 1; in nicvf_refill_rbdr()
|
| H A D | nicvf_main.c | 2098 int err, qcount; in nicvf_probe() local 2127 qcount = netif_get_num_default_rss_queues(); in nicvf_probe() 2132 qcount = min_t(int, num_online_cpus(), in nicvf_probe() 2136 netdev = alloc_etherdev_mqs(sizeof(struct nicvf), qcount, qcount); in nicvf_probe() 2150 nic->max_queues = qcount; in nicvf_probe()
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | qplib_res.c | 391 if (!tqmctx->qcount[i]) in bnxt_qplib_alloc_tqm_rings() 393 hwq_attr.depth = ctx->qpc_count * tqmctx->qcount[i]; in bnxt_qplib_alloc_tqm_rings()
|
| H A D | main.c | 248 rdev->qplib_ctx.tqm_ctx.qcount[i] = in bnxt_re_limit_pf_res()
|
| /linux/drivers/net/ethernet/broadcom/bnx2x/ |
| H A D | bnx2x_sriov.c | 2445 int vf_idx, sb_idx, vfq_idx, qcount, first_vf; in bnx2x_enable_sriov() local 2489 qcount = 0; in bnx2x_enable_sriov() 2494 vf->vfqs = &bp->vfdb->vfqs[qcount]; in bnx2x_enable_sriov() 2495 qcount += vf_sb_count(vf); in bnx2x_enable_sriov()
|
| /linux/net/core/ |
| H A D | dev.c | 3515 u16 qcount = dev->real_num_tx_queues; in skb_tx_hash() local 3521 qcount = sb_dev->tc_to_txq[tc].count; in skb_tx_hash() 3522 if (unlikely(!qcount)) { in skb_tx_hash() 3526 qcount = dev->real_num_tx_queues; in skb_tx_hash() 3531 DEBUG_NET_WARN_ON_ONCE(qcount == 0); in skb_tx_hash() 3535 while (unlikely(hash >= qcount)) in skb_tx_hash() 3536 hash -= qcount; in skb_tx_hash() 3540 return (u16) reciprocal_scale(skb_get_hash(skb), qcount) + qoffset; in skb_tx_hash()
|
| /linux/drivers/net/wireless/broadcom/brcm80211/brcmfmac/ |
| H A D | sdio.c | 536 static int qcount[NUMPRIO]; variable 2845 if (pktq_plen(&bus->txq, prec) > qcount[prec]) in brcmf_sdio_bus_txdata() 2846 qcount[prec] = pktq_plen(&bus->txq, prec); in brcmf_sdio_bus_txdata()
|
| /linux/drivers/net/ethernet/intel/iavf/ |
| H A D | iavf_main.c | 3784 u16 qcount = mqprio_qopt->qopt.count[i]; in __iavf_setup_tc() local 3788 netdev_set_tc_queue(netdev, netdev_tc++, qcount, in __iavf_setup_tc()
|