| /linux/drivers/net/ethernet/huawei/hinic3/ |
| H A D | hinic3_tx.c | 59 txq->q_depth = nic_dev->q_params.sq_depth; in hinic3_alloc_txqs() 60 txq->q_mask = nic_dev->q_params.sq_depth - 1; in hinic3_alloc_txqs() 163 static void free_all_tx_skbs(struct net_device *netdev, u32 sq_depth, in free_all_tx_skbs() argument 169 for (idx = 0; idx < sq_depth; idx++) { in free_all_tx_skbs() 676 u32 sq_depth, struct hinic3_dyna_txq_res *txqs_res) in hinic3_alloc_txqs_res() argument 684 tqres->tx_info = kzalloc_objs(*tqres->tx_info, sq_depth); in hinic3_alloc_txqs_res() 689 sq_depth * HINIC3_BDS_PER_SQ_WQEBB + HINIC3_MAX_SQ_SGE); in hinic3_alloc_txqs_res() 711 u32 sq_depth, struct hinic3_dyna_txq_res *txqs_res) in hinic3_free_txqs_res() argument 719 free_all_tx_skbs(netdev, sq_depth, tqres->tx_info); in hinic3_free_txqs_res() 726 u32 sq_depth, struct hinic3_dyna_txq_res *txqs_res) in hinic3_configure_txqs() argument [all …]
|
| H A D | hinic3_netdev_ops.c | 148 q_params->sq_depth, q_params->txqs_res); in hinic3_alloc_txrxq_resources() 164 hinic3_free_txqs_res(netdev, q_params->num_qps, q_params->sq_depth, in hinic3_alloc_txrxq_resources() 184 hinic3_free_txqs_res(netdev, q_params->num_qps, q_params->sq_depth, in hinic3_free_txrxq_resources() 203 q_params->sq_depth, q_params->txqs_res); in hinic3_configure_txrxqs() 262 qp_params->sq_depth = trxq_params->sq_depth; in hinic3_alloc_channel_resources()
|
| H A D | hinic3_nic_io.c | 318 u16 q_id, u32 sq_depth, u16 sq_msix_idx) in hinic3_create_sq() argument 328 err = hinic3_wq_create(hwdev, &sq->wq, sq_depth, in hinic3_create_sq() 362 struct hinic3_io_queue *rq, u16 q_id, u32 sq_depth, in hinic3_create_qp() argument 367 err = hinic3_create_sq(hwdev, sq, q_id, sq_depth, qp_msix_idx); in hinic3_create_qp() 425 qp_params->sq_depth, qp_params->rq_depth, in hinic3_alloc_qps()
|
| /linux/net/9p/ |
| H A D | trans_rdma.c | 79 int sq_depth; member 115 if (rdma->sq_depth != P9_RDMA_SQ_DEPTH) in p9_rdma_show_options() 116 seq_printf(m, ",sq=%u", rdma->sq_depth); in p9_rdma_show_options() 469 rdma->sq_depth = opts->sq_depth; in alloc_rdma() 474 sema_init(&rdma->sq_sem, rdma->sq_depth); in alloc_rdma() 585 opts.sq_depth + opts.rq_depth + 1, in rdma_create_trans() 599 qp_attr.cap.max_send_wr = opts.sq_depth; in rdma_create_trans()
|
| /linux/drivers/net/ethernet/fungible/funcore/ |
| H A D | fun_queue.c | 76 u8 sqe_size_log2, u32 sq_depth, dma_addr_t dma_addr, in fun_sq_create() argument 89 if (sq_depth > fdev->q_depth) in fun_sq_create() 94 wb_addr = dma_addr + (sq_depth << sqe_size_log2); in fun_sq_create() 101 sq_depth - 1, dma_addr, 0, in fun_sq_create() 356 funq->sq_cmds = fun_alloc_ring_mem(funq->fdev->dev, funq->sq_depth, in fun_alloc_sqes() 390 fun_free_ring_mem(dev, funq->sq_depth, 1 << funq->sqe_size_log2, in fun_free_queue() 442 funq->sq_depth = req->sq_depth; in fun_alloc_queue()
|
| H A D | fun_dev.c | 231 .sq_depth = areq->sq_depth, in fun_enable_admin_queue() 234 unsigned int ntags = areq->sq_depth - 1; in fun_enable_admin_queue() 241 if (areq->sq_depth < AQA_MIN_QUEUE_SIZE || in fun_enable_admin_queue() 242 areq->sq_depth > AQA_MAX_QUEUE_SIZE || in fun_enable_admin_queue() 269 writel((funq->sq_depth - 1) << AQA_ASQS_SHIFT | in fun_enable_admin_queue() 401 if (++funq->sq_tail == funq->sq_depth) in fun_submit_admin_cmd()
|
| H A D | fun_dev.h | 93 u16 sq_depth; member
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_dev.c | 270 static int set_hw_ioctxt(struct hinic_hwdev *hwdev, unsigned int sq_depth, in set_hw_ioctxt() argument 289 hw_ioctxt.sq_depth = ilog2(sq_depth); in set_hw_ioctxt() 438 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth) in hinic_hwdev_ifup() argument 461 func_to_io->sq_depth = sq_depth; in hinic_hwdev_ifup() 489 err = set_hw_ioctxt(hwdev, sq_depth, rq_depth); in hinic_hwdev_ifup()
|
| H A D | hinic_hw_io.h | 73 u16 sq_depth; member
|
| H A D | hinic_dev.h | 99 u16 sq_depth; member
|
| H A D | hinic_hw_dev.h | 288 u16 sq_depth; member 627 int hinic_hwdev_ifup(struct hinic_hwdev *hwdev, u16 sq_depth, u16 rq_depth);
|
| H A D | hinic_ethtool.c | 558 ring->tx_pending = nic_dev->sq_depth; in hinic_get_ringparam() 599 if (new_sq_depth == nic_dev->sq_depth && in hinic_set_ringparam() 605 nic_dev->sq_depth, nic_dev->rq_depth, in hinic_set_ringparam() 608 nic_dev->sq_depth = new_sq_depth; in hinic_set_ringparam()
|
| H A D | hinic_hw_mbox.c | 1291 (hw_ctxt)->sq_depth >= HINIC_QUEUE_MIN_DEPTH && \ 1292 (hw_ctxt)->sq_depth <= HINIC_QUEUE_MAX_DEPTH && \ 1300 if (!hw_ctxt->rq_depth && !hw_ctxt->sq_depth && in hw_ctxt_qps_param_valid()
|
| /linux/drivers/crypto/hisilicon/ |
| H A D | qm.c | 275 #define QM_MK_SQC_DW3_V2(sqe_sz, sq_depth) \ argument 276 ((((u32)sq_depth) - 1) | ((u32)ilog2(sqe_sz) << QM_SQ_SQE_SIZE_SHIFT)) 2037 if (unlikely(atomic_read(&qp->qp_status.used) == qp->sq_depth - 1)) in qm_get_avail_sqe() 2174 sqc.w8 = cpu_to_le16(qp->sq_depth - 1); in qm_sq_ctx_cfg() 2176 sqc.dw3 = cpu_to_le32(QM_MK_SQC_DW3_V2(qm->sqe_size, qp->sq_depth)); in qm_sq_ctx_cfg() 2288 u16 sq_depth = qp->sq_depth; in qp_stop_fail_cb() local 2289 u16 cur_head = (cur_tail + sq_depth - qp_used) % sq_depth; in qp_stop_fail_cb() 2295 pos = (i + cur_head) % sq_depth; in qp_stop_fail_cb() 2466 sq_tail_next = (sq_tail + 1) % qp->sq_depth; in hisi_qp_send() 2693 qp_info.sq_depth = qp->sq_depth; in hisi_qm_uacce_ioctl() [all …]
|
| H A D | debugfs.c | 322 u16 sq_depth = qm->qp_array->sq_depth; in qm_sq_dump() local 328 ret = q_dump_param_parse(qm, s, &sqe_id, &qp_id, sq_depth); in qm_sq_dump()
|
| /linux/drivers/net/ethernet/fungible/funeth/ |
| H A D | funeth_main.c | 494 xdpqs = alloc_xdpqs(netdev, qset->nxdpqs, qset->sq_depth, in fun_alloc_rings() 503 err = alloc_txqs(netdev, txqs, qset->ntxqs, qset->sq_depth, in fun_alloc_rings() 843 .sq_depth = fp->sq_depth, in funeth_open() 1072 xdpqs = alloc_xdpqs(dev, nqs, fp->sq_depth, 0, FUN_QSTATE_INIT_FULL); in fun_enter_xdp() 1632 .sq_depth = fp->sq_depth, in fun_change_num_queues() 1769 fp->sq_depth = min(SQ_DEPTH, fdev->q_depth); in fun_create_netdev() 1990 .sq_depth = ADMIN_SQ_DEPTH, in funeth_probe()
|
| /linux/tools/testing/selftests/ublk/ |
| H A D | null.c | 42 dev->tgt.sq_depth = dev->tgt.cq_depth = 2 * info->queue_depth; in ublk_null_tgt_init()
|
| H A D | kublk.h | 160 unsigned int sq_depth; member
|
| H A D | stripe.c | 350 dev->tgt.sq_depth = mul * dev->dev_info.queue_depth * conf->nr_files; in ublk_stripe_tgt_init()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | user.h | 446 u32 *sq_depth, u8 *sq_shift); 559 u32 sq_depth; member
|
| H A D | verbs.c | 661 ret = irdma_uk_calc_depth_shift_sq(ukinfo, &ukinfo->sq_depth, in irdma_setup_umode_qp() 672 (ukinfo->sq_depth - IRDMA_SQ_RSVD) >> ukinfo->sq_shift; in irdma_setup_umode_qp() 675 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_umode_qp() 701 status = irdma_uk_calc_depth_shift_sq(ukinfo, &ukinfo->sq_depth, in irdma_setup_kmode_qp() 712 kzalloc_objs(*iwqp->kqp.sq_wrid_mem, ukinfo->sq_depth); in irdma_setup_kmode_qp() 728 size = (ukinfo->sq_depth + ukinfo->rq_depth) * IRDMA_QP_WQE_MIN_SIZE; in irdma_setup_kmode_qp() 744 ukinfo->rq = &ukinfo->sq[ukinfo->sq_depth]; in irdma_setup_kmode_qp() 745 info->rq_pa = info->sq_pa + (ukinfo->sq_depth * IRDMA_QP_WQE_MIN_SIZE); in irdma_setup_kmode_qp() 749 ukinfo->sq_size = ukinfo->sq_depth >> ukinfo->sq_shift; in irdma_setup_kmode_qp() 753 iwqp->max_send_wr = (ukinfo->sq_depth - IRDMA_SQ_RSVD) >> ukinfo->sq_shift; in irdma_setup_kmode_qp()
|
| /linux/include/net/9p/ |
| H A D | client.h | 176 int sq_depth; member
|
| /linux/drivers/crypto/hisilicon/zip/ |
| H A D | zip_crypto.c | 464 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_req_q() 515 u16 q_depth = ctx->qp_ctx[0].qp->sq_depth; in hisi_zip_create_sgl_pool()
|
| /linux/fs/9p/ |
| H A D | vfs_super.c | 343 ctx->rdma_opts.sq_depth = P9_RDMA_SQ_DEPTH; in v9fs_init_fs_context()
|
| /linux/drivers/crypto/hisilicon/sec2/ |
| H A D | sec_crypto.c | 130 req_id = idr_alloc_cyclic(&qp_ctx->req_idr, NULL, 0, qp_ctx->qp->sq_depth, GFP_ATOMIC); in sec_alloc_req_id() 140 if (unlikely(req_id < 0 || req_id >= qp_ctx->qp->sq_depth)) { in sec_free_req_id() 209 if (atomic_read(&qp_ctx->qp->qp_status.used) == qp_ctx->qp->sq_depth - 1) in qp_send_message() 213 if (atomic_read(&qp_ctx->qp->qp_status.used) == qp_ctx->qp->sq_depth - 1) { in qp_send_message() 229 qp_ctx->send_head = (qp_ctx->send_head + 1) % qp_ctx->qp->sq_depth; in qp_send_message() 568 u16 q_depth = qp_ctx->qp->sq_depth; in sec_alloc_qp_ctx_resource()
|