| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_xsk.c | 36 static int otx2_xsk_ctx_disable(struct otx2_nic *pfvf, u16 qidx, int aura_id) in otx2_xsk_ctx_disable() argument 47 cn10k_rq_aq->qidx = qidx; in otx2_xsk_ctx_disable() 56 rq_aq->qidx = qidx; in otx2_xsk_ctx_disable() 91 static void otx2_clean_up_rq(struct otx2_nic *pfvf, int qidx) in otx2_clean_up_rq() argument 102 cq = &qset->cq[qidx]; in otx2_clean_up_rq() 104 otx2_cleanup_rx_cqes(pfvf, cq, qidx); in otx2_clean_up_rq() 106 pool = &pfvf->qset.pool[qidx]; in otx2_clean_up_rq() 107 iova = otx2_aura_allocptr(pfvf, qidx); in otx2_clean_up_rq() 111 iova = otx2_aura_allocptr(pfvf, qidx); in otx2_clean_up_rq() 115 otx2_xsk_ctx_disable(pfvf, qidx, qidx); in otx2_clean_up_rq() [all …]
|
| H A D | qos_sq.c | 33 static int otx2_qos_sq_aura_pool_init(struct otx2_nic *pfvf, int qidx) in otx2_qos_sq_aura_pool_init() argument 56 pool_id = otx2_get_pool_idx(pfvf, AURA_NIX_SQ, qidx); in otx2_qos_sq_aura_pool_init() 76 sq = &qset->sq[qidx]; in otx2_qos_sq_aura_pool_init() 116 static void otx2_qos_sq_free_sqbs(struct otx2_nic *pfvf, int qidx) in otx2_qos_sq_free_sqbs() argument 124 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 140 sq = &qset->sq[qidx]; in otx2_qos_sq_free_sqbs() 151 static void otx2_qos_sqb_flush(struct otx2_nic *pfvf, int qidx) in otx2_qos_sqb_flush() argument 158 incr = (u64)qidx << 32; in otx2_qos_sqb_flush() 166 static int otx2_qos_ctx_disable(struct otx2_nic *pfvf, u16 qidx, int aura_id) in otx2_qos_ctx_disable() argument 177 cn10k_sq_aq->qidx = qidx; in otx2_qos_ctx_disable() [all …]
|
| H A D | otx2_common.c | 28 struct otx2_nic *pfvf, int qidx) in otx2_nix_rq_op_stats() argument 30 u64 incr = (u64)qidx << 32; in otx2_nix_rq_op_stats() 41 struct otx2_nic *pfvf, int qidx) in otx2_nix_sq_op_stats() argument 43 u64 incr = (u64)qidx << 32; in otx2_nix_sq_op_stats() 84 int otx2_update_rq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_rq_stats() argument 86 struct otx2_rcv_queue *rq = &pfvf->qset.rq[qidx]; in otx2_update_rq_stats() 91 otx2_nix_rq_op_stats(&rq->stats, pfvf, qidx); in otx2_update_rq_stats() 96 int otx2_update_sq_stats(struct otx2_nic *pfvf, int qidx) in otx2_update_sq_stats() argument 98 struct otx2_snd_queue *sq = &pfvf->qset.sq[qidx]; in otx2_update_sq_stats() 103 if (qidx >= pfvf->hw.non_qos_queues) { in otx2_update_sq_stats() [all …]
|
| H A D | rep.c | 156 u16 qidx = rep->rep_id; in rvu_rep_sp_stats64() local 158 otx2_update_rq_stats(priv, qidx); in rvu_rep_sp_stats64() 159 rq = &priv->qset.rq[qidx]; in rvu_rep_sp_stats64() 161 otx2_update_sq_stats(priv, qidx); in rvu_rep_sp_stats64() 162 sq = &priv->qset.sq[qidx]; in rvu_rep_sp_stats64() 504 int err = 0, qidx, vec; in rvu_rep_napi_init() local 512 for (qidx = 0; qidx < hw->cint_cnt; qidx++) { in rvu_rep_napi_init() 513 cq_poll = &qset->napi[qidx]; in rvu_rep_napi_init() 514 cq_poll->cint_idx = qidx; in rvu_rep_napi_init() 516 (qidx < hw->rx_queues) ? qidx : CINT_INVALID_CQ; in rvu_rep_napi_init() [all …]
|
| H A D | otx2_txrx.c | 188 int qidx) in otx2_skb_add_frag() argument 219 pfvf->hw_ops->aura_freeptr(pfvf, qidx, iova & ~0x07ULL); in otx2_skb_add_frag() 247 int qidx) in otx2_free_rcv_seg() argument 260 pfvf->hw_ops->aura_freeptr(pfvf, qidx, in otx2_free_rcv_seg() 267 struct nix_cqe_rx_s *cqe, int qidx) in otx2_check_rcv_errors() argument 275 qidx, parse->errlev, parse->errcode); in otx2_check_rcv_errors() 326 otx2_free_rcv_seg(pfvf, cqe, qidx); in otx2_check_rcv_errors() 453 int *xsk_frames, int qidx, int budget) in otx2_zc_submit_pkts() argument 461 otx2_zc_napi_handler(pfvf, xsk_pool, qidx, budget); in otx2_zc_submit_pkts() 467 int tx_pkts = 0, tx_bytes = 0, qidx; in otx2_tx_napi_handler() local [all …]
|
| H A D | otx2_pf.c | 1398 u64 val, qidx = 0; in otx2_q_intr_handler() local 1401 for (qidx = 0; qidx < pf->qset.cq_cnt; qidx++) { in otx2_q_intr_handler() 1403 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler() 1405 otx2_write64(pf, NIX_LF_CQ_OP_INT, (qidx << 44) | in otx2_q_intr_handler() 1413 qidx, otx2_read64(pf, NIX_LF_ERR_INT)); in otx2_q_intr_handler() 1417 qidx); in otx2_q_intr_handler() 1421 qidx); in otx2_q_intr_handler() 1428 for (qidx = 0; qidx < otx2_get_total_tx_queues(pf); qidx++) { in otx2_q_intr_handler() 1432 sq = &pf->qset.sq[qidx]; in otx2_q_intr_handler() 1442 val = otx2_atomic64_add((qidx << 44), ptr); in otx2_q_intr_handler() [all …]
|
| H A D | otx2_ethtool.c | 89 int qidx, stats; in otx2_get_qset_strings() local 91 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) in otx2_get_qset_strings() 93 ethtool_sprintf(data, "rxq%d: %s", qidx + start_qidx, in otx2_get_qset_strings() 96 for (qidx = 0; qidx < otx2_get_total_tx_queues(pfvf); qidx++) in otx2_get_qset_strings() 98 if (qidx >= pfvf->hw.non_qos_queues) in otx2_get_qset_strings() 100 qidx + start_qidx - in otx2_get_qset_strings() 105 qidx + start_qidx, in otx2_get_qset_strings() 141 int stat, qidx; in otx2_get_qset_stats() local 145 for (qidx = 0; qidx < pfvf->hw.rx_queues; qidx++) { in otx2_get_qset_stats() 146 if (!otx2_update_rq_stats(pfvf, qidx)) { in otx2_get_qset_stats() [all …]
|
| H A D | qos.h | 27 void otx2_qos_free_qid(struct otx2_nic *pfvf, int qidx); 28 int otx2_qos_enable_sq(struct otx2_nic *pfvf, int qidx); 29 void otx2_qos_disable_sq(struct otx2_nic *pfvf, int qidx);
|
| H A D | otx2_txrx.h | 184 struct sk_buff *skb, u16 qidx); 186 int size, int qidx); 188 int size, int qidx);
|
| H A D | cn10k.h | 28 void cn10k_sqe_flush(void *dev, struct otx2_snd_queue *sq, int size, int qidx); 29 int cn10k_sq_aq_init(void *dev, u16 qidx, u8 chan_offset, u16 sqb_aura);
|
| H A D | otx2_xsk.h | 22 void otx2_attach_xsk_buff(struct otx2_nic *pfvf, struct otx2_snd_queue *sq, int qidx);
|
| /linux/drivers/net/ethernet/cavium/thunder/ |
| H A D | nicvf_queues.c | 32 static int nicvf_poll_reg(struct nicvf *nic, int qidx, in nicvf_poll_reg() argument 43 reg_val = nicvf_queue_reg_read(nic, reg, qidx); in nicvf_poll_reg() 504 struct snd_queue *sq, int q_len, int qidx) in nicvf_init_snd_queue() argument 524 qidx += ((nic->sqs_id + 1) * MAX_SND_QUEUES_PER_QS); in nicvf_init_snd_queue() 525 if (qidx < nic->pnicvf->xdp_tx_queues) { in nicvf_init_snd_queue() 627 struct queue_set *qs, int qidx) in nicvf_reclaim_snd_queue() argument 630 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, 0); in nicvf_reclaim_snd_queue() 632 if (nicvf_poll_reg(nic, qidx, NIC_QSET_SQ_0_7_STATUS, 21, 1, 0x01)) in nicvf_reclaim_snd_queue() 635 nicvf_queue_reg_write(nic, NIC_QSET_SQ_0_7_CFG, qidx, NICVF_SQ_RESET); in nicvf_reclaim_snd_queue() 639 struct queue_set *qs, int qidx) in nicvf_reclaim_rcv_queue() argument [all …]
|
| H A D | nicvf_main.c | 75 static inline u8 nicvf_netdev_qidx(struct nicvf *nic, u8 qidx) in nicvf_netdev_qidx() argument 78 return qidx + ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS); in nicvf_netdev_qidx() 80 return qidx; in nicvf_netdev_qidx() 104 u64 qidx, u64 val) in nicvf_queue_reg_write() argument 108 writeq_relaxed(val, addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_write() 111 u64 nicvf_queue_reg_read(struct nicvf *nic, u64 offset, u64 qidx) in nicvf_queue_reg_read() argument 115 return readq_relaxed(addr + (qidx << NIC_Q_NUM_SHIFT)); in nicvf_queue_reg_read() 989 int qidx; in nicvf_handle_qs_err() local 995 for (qidx = 0; qidx < qs->cq_cnt; qidx++) { in nicvf_handle_qs_err() 997 qidx); in nicvf_handle_qs_err() [all …]
|
| H A D | nicvf_queues.h | 336 int qidx, bool enable); 338 void nicvf_sq_enable(struct nicvf *nic, struct snd_queue *sq, int qidx); 339 void nicvf_sq_disable(struct nicvf *nic, int qidx); 342 struct snd_queue *sq, int qidx); 363 u64 qidx, u64 val); 365 u64 offset, u64 qidx);
|
| /linux/drivers/net/ethernet/chelsio/inline_crypto/ch_ipsec/ |
| H A D | chcr_ipsec.c | 417 u32 qidx; in copy_esn_pktxt() local 425 qidx = skb->queue_mapping; in copy_esn_pktxt() 426 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_esn_pktxt() 467 u32 ctrl0, qidx; in copy_cpltx_pktxt() local 473 qidx = skb->queue_mapping; in copy_cpltx_pktxt() 474 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_cpltx_pktxt() 512 unsigned int qidx; in copy_key_cpltx_pktxt() local 516 qidx = skb->queue_mapping; in copy_key_cpltx_pktxt() 517 q = &adap->sge.ethtxq[qidx + pi->first_qset]; in copy_key_cpltx_pktxt() 572 int qidx = skb_get_queue_mapping(skb); in ch_ipsec_crypto_wreq() local [all …]
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_rx.c | 432 skb_record_rx_queue(skb, q->qidx); in fun_handle_cqe_pkt() 614 unsigned int qidx, in fun_rxq_create_sw() argument 629 q->qidx = qidx; in fun_rxq_create_sw() 673 netdev_err(dev, "Unable to allocate memory for Rx queue %u\n", qidx); in fun_rxq_create_sw() 704 err = xdp_rxq_info_reg(&q->xdp_rxq, q->netdev, q->qidx, in fun_rxq_create_dev() 750 q->qidx, ncqe, nrqe, q->hw_cqid, q->hw_sqid, irq->irq_idx, in fun_rxq_create_dev() 761 q->qidx, err); in fun_rxq_create_dev() 776 q->qidx, q->hw_cqid, q->hw_sqid, irq->irq_idx); in fun_rxq_free_dev() 788 int funeth_rxq_create(struct net_device *dev, unsigned int qidx, in funeth_rxq_create() argument 796 q = fun_rxq_create_sw(dev, qidx, ncqe, nrqe, irq); in funeth_rxq_create()
|
| H A D | funeth_tx.c | 624 unsigned int qidx, in fun_txq_create_sw() argument 635 numa_node = cpu_to_node(qidx); /* XDP Tx queue */ in fun_txq_create_sw() 651 q->qidx = qidx; in fun_txq_create_sw() 661 irq ? "Tx" : "XDP", qidx); in fun_txq_create_sw() 709 q->ndq = netdev_get_tx_queue(q->netdev, q->qidx); in fun_txq_create_dev() 718 irq ? "Tx" : "XDP", q->qidx, ndesc, q->hw_qid, irq_idx, in fun_txq_create_dev() 727 irq ? "Tx" : "XDP", q->qidx, err); in fun_txq_create_dev() 740 q->irq ? "Tx" : "XDP", q->qidx, q->hw_qid, in fun_txq_free_dev() 759 int funeth_txq_create(struct net_device *dev, unsigned int qidx, in funeth_txq_create() argument 767 q = fun_txq_create_sw(dev, qidx, ndesc, irq); in funeth_txq_create()
|
| H A D | funeth_txrx.h | 117 u16 qidx; /* queue index within net_device */ member 173 u16 qidx; /* queue index within net_device */ member 254 int funeth_txq_create(struct net_device *dev, unsigned int qidx, 259 int funeth_rxq_create(struct net_device *dev, unsigned int qidx,
|
| /linux/drivers/net/ethernet/broadcom/bnxt/ |
| H A D | bnxt_dcb.c | 55 u8 qidx; in bnxt_hwrm_queue_pri2cos_cfg() local 60 qidx = bp->tc_to_qidx[ets->prio_tc[i]]; in bnxt_hwrm_queue_pri2cos_cfg() 61 pri2cos[i] = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_pri2cos_cfg() 108 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_cos2bw_cfg() local 112 qidx); in bnxt_hwrm_queue_cos2bw_cfg() 115 cos2bw.queue_id = bp->q_info[qidx].queue_id; in bnxt_hwrm_queue_cos2bw_cfg() 131 if (qidx == 0) { in bnxt_hwrm_queue_cos2bw_cfg() 277 u8 qidx = bp->tc_to_qidx[i]; in bnxt_hwrm_queue_pfc_cfg() local 279 if (!BNXT_LLQ(bp->q_info[qidx].queue_profile)) { in bnxt_hwrm_queue_pfc_cfg()
|
| /linux/drivers/scsi/csiostor/ |
| H A D | csio_wr.c | 744 csio_wr_cleanup_eq_stpg(struct csio_hw *hw, int qidx) in csio_wr_cleanup_eq_stpg() argument 746 struct csio_q *q = csio_hw_to_wrm(hw)->q_arr[qidx]; in csio_wr_cleanup_eq_stpg() 761 csio_wr_cleanup_iq_ftr(struct csio_hw *hw, int qidx) in csio_wr_cleanup_iq_ftr() argument 764 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_cleanup_iq_ftr() 862 csio_wr_get(struct csio_hw *hw, int qidx, uint32_t size, in csio_wr_get() argument 866 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_get() 877 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_get() 981 csio_wr_issue(struct csio_hw *hw, int qidx, bool prio) in csio_wr_issue() argument 984 struct csio_q *q = wrm->q_arr[qidx]; in csio_wr_issue() 986 CSIO_DB_ASSERT((qidx >= 0) && (qidx < wrm->free_qidx)); in csio_wr_issue() [all …]
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| H A D | amdgpu_amdkfd_gfx_v9.c | 1026 int qidx; in kgd_gfx_v9_get_cu_occupancy() local 1055 for (qidx = 0; qidx < max_queue_cnt; qidx++) { in kgd_gfx_v9_get_cu_occupancy() 1059 if (!test_bit(qidx, cp_queue_bitmap)) in kgd_gfx_v9_get_cu_occupancy() 1062 if (!(queue_map & (1 << qidx))) in kgd_gfx_v9_get_cu_occupancy() 1066 get_wave_count(adev, qidx, &cu_occupancy[qidx], in kgd_gfx_v9_get_cu_occupancy()
|
| /linux/drivers/scsi/qla2xxx/ |
| H A D | qla_nvme.c | 115 unsigned int qidx, u16 qsize, void **handle) in qla_nvme_alloc_queue() argument 122 if (qidx) in qla_nvme_alloc_queue() 123 qidx--; in qla_nvme_alloc_queue() 130 __func__, handle, qidx, qsize); in qla_nvme_alloc_queue() 132 if (qidx > qla_nvme_fc_transport.max_hw_queues) { in qla_nvme_alloc_queue() 135 __func__, qidx, qla_nvme_fc_transport.max_hw_queues); in qla_nvme_alloc_queue() 143 if (ha->queue_pair_map[qidx]) { in qla_nvme_alloc_queue() 144 *handle = ha->queue_pair_map[qidx]; in qla_nvme_alloc_queue() 147 *handle, qidx); in qla_nvme_alloc_queue()
|
| /linux/drivers/net/ethernet/marvell/octeontx2/af/ |
| H A D | rvu_nix.c | 1077 if (!pfvf->rq_ctx || req->qidx >= pfvf->rq_ctx->qsize) in rvu_nix_blk_aq_enq_inst() 1081 if (!pfvf->sq_ctx || req->qidx >= pfvf->sq_ctx->qsize) in rvu_nix_blk_aq_enq_inst() 1085 if (!pfvf->cq_ctx || req->qidx >= pfvf->cq_ctx->qsize) in rvu_nix_blk_aq_enq_inst() 1092 (req->qidx >= (256UL << (cfg & 0xF)))) in rvu_nix_blk_aq_enq_inst() 1100 (req->qidx >= (256UL << (cfg & 0xF)))) in rvu_nix_blk_aq_enq_inst() 1134 inst.cindex = req->qidx; in rvu_nix_blk_aq_enq_inst() 1210 __set_bit(req->qidx, pfvf->rq_bmap); in rvu_nix_blk_aq_enq_inst() 1212 __set_bit(req->qidx, pfvf->sq_bmap); in rvu_nix_blk_aq_enq_inst() 1214 __set_bit(req->qidx, pfvf->cq_bmap); in rvu_nix_blk_aq_enq_inst() 1220 (test_bit(req->qidx, pfvf->rq_bmap) & in rvu_nix_blk_aq_enq_inst() [all …]
|
| /linux/include/linux/ |
| H A D | nvme-fc-driver.h | 480 unsigned int qidx, u16 qsize, 483 unsigned int qidx, void *handle);
|
| /linux/drivers/net/ethernet/chelsio/cxgb3/ |
| H A D | cxgb3_main.c | 401 int i, j, err, qidx = 0; in request_msix_data_irqs() local 407 err = request_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 409 adap->sge.qs[qidx]. in request_msix_data_irqs() 411 adap->msix_info[qidx + 1].desc, in request_msix_data_irqs() 412 &adap->sge.qs[qidx]); in request_msix_data_irqs() 414 while (--qidx >= 0) in request_msix_data_irqs() 415 free_irq(adap->msix_info[qidx + 1].vec, in request_msix_data_irqs() 416 &adap->sge.qs[qidx]); in request_msix_data_irqs() 419 qidx++; in request_msix_data_irqs() 939 static int send_pktsched_cmd(struct adapter *adap, int sched, int qidx, int lo, in send_pktsched_cmd() argument [all …]
|