Home
last modified time | relevance | path

Searched refs:cq_size (Results 1 – 25 of 25) sorted by relevance

/linux/drivers/infiniband/ulp/iser/
H A Diser_verbs.c235 unsigned int max_send_wr, cq_size; in iser_create_ib_conn_res() local
250 cq_size = max_send_wr + ISER_QP_MAX_RECV_DTOS; in iser_create_ib_conn_res()
251 ib_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_SOFTIRQ); in iser_create_ib_conn_res()
256 ib_conn->cq_size = cq_size; in iser_create_ib_conn_res()
285 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size); in iser_create_ib_conn_res()
382 ib_cq_pool_put(ib_conn->cq, ib_conn->cq_size); in iser_free_ib_conn_res()
H A Discsi_iser.h371 u32 cq_size; member
/linux/drivers/net/ethernet/mellanox/mlxbf_gige/
H A Dmlxbf_gige_rx.c122 size_t wq_size, cq_size; in mlxbf_gige_rx_init() local
156 cq_size = MLXBF_GIGE_RX_CQE_SZ * priv->rx_q_entries; in mlxbf_gige_rx_init()
157 priv->rx_cqe_base = dma_alloc_coherent(priv->dev, cq_size, in mlxbf_gige_rx_init()
/linux/drivers/net/ethernet/amd/pds_core/
H A Dcore.c153 dma_free_coherent(dev, qcq->cq_size, in pdsc_qcq_free()
239 qcq->q_size + qcq->cq_size, in pdsc_qcq_alloc()
272 qcq->cq_size = PDS_PAGE_SIZE + (num_descs * cq_desc_size); in pdsc_qcq_alloc()
273 qcq->cq_base = dma_alloc_coherent(dev, qcq->cq_size, in pdsc_qcq_alloc()
H A Ddebugfs.c122 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size); in pdsc_debugfs_add_qcq()
/linux/include/uapi/rdma/
H A Dirdma-abi.h108 __u32 cq_size; member
/linux/drivers/infiniband/hw/irdma/
H A Duk.c1043 void irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) in irdma_uk_cq_resize() argument
1046 cq->cq_size = cq_size; in irdma_uk_cq_resize()
1047 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_resize()
1761 cq->cq_size = info->cq_size; in irdma_uk_cq_init()
1766 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_init()
H A Dpuda.c734 set_64bit_val(wqe, 0, cq->cq_uk.cq_size); in irdma_puda_cq_wqe()
776 cqsize = rsrc->cq_size * (sizeof(struct irdma_cqe)); in irdma_puda_cq_create()
788 info.shadow_read_threshold = rsrc->cq_size >> 2; in irdma_puda_cq_create()
793 init_info->cq_size = rsrc->cq_size; in irdma_puda_cq_create()
1044 rsrc->cq_size = info->rq_size + info->sq_size; in irdma_puda_create_rsrc()
1047 rsrc->cq_size += info->rq_size; in irdma_puda_create_rsrc()
H A Duser.h538 u32 cq_size; member
575 u32 cq_size; member
H A Dverbs.c2063 info.cq_size = max(entries, 4); in irdma_resize_cq()
2065 if (info.cq_size == iwcq->sc_cq.cq_uk.cq_size - 1) in irdma_resize_cq()
2102 rsize = info.cq_size * sizeof(struct irdma_cqe); in irdma_resize_cq()
2152 ibcq->cqe = info.cq_size - 1; in irdma_resize_cq()
2475 #define IRDMA_CREATE_CQ_MIN_RESP_LEN offsetofend(struct irdma_create_cq_resp, cq_size) in irdma_create_cq()
2515 ukinfo->cq_size = max(entries, 4); in irdma_create_cq()
2521 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; in irdma_create_cq()
2602 ukinfo->cq_size = entries; in irdma_create_cq()
2605 rsize = info.cq_uk_init_info.cq_size * sizeof(struct irdma_extended_cqe); in irdma_create_cq()
2607 rsize = info.cq_uk_init_info.cq_size * sizeof(struct irdma_cqe); in irdma_create_cq()
[all …]
H A Dctrl.c2958 set_64bit_val(wqe, 0, cq->cq_uk.cq_size); in irdma_sc_cq_create()
3015 set_64bit_val(wqe, 0, cq->cq_uk.cq_size); in irdma_sc_cq_destroy()
3055 irdma_uk_cq_resize(&cq->cq_uk, info->cq_base, info->cq_size); in irdma_sc_cq_resize()
3084 set_64bit_val(wqe, 0, info->cq_size); in irdma_sc_cq_modify()
4895 cq->cq_uk.cq_size = info->num_elem; in irdma_sc_ccq_init()
4977 set_64bit_val(wqe, 0, ccq->cq_uk.cq_size); in irdma_sc_ccq_destroy()
H A Dtype.h722 u32 cq_size; member
/linux/drivers/infiniband/ulp/isert/
H A Dib_isert.h184 u32 cq_size; member
H A Dib_isert.c103 u32 cq_size = ISERT_QP_MAX_REQ_DTOS + ISERT_QP_MAX_RECV_DTOS + 2; in isert_create_qp() local
109 isert_conn->cq = ib_cq_pool_get(ib_dev, cq_size, -1, IB_POLL_WORKQUEUE); in isert_create_qp()
115 isert_conn->cq_size = cq_size; in isert_create_qp()
137 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size); in isert_create_qp()
409 ib_cq_pool_put(isert_conn->cq, isert_conn->cq_size); in isert_destroy_qp()
/linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c411 static int mlx5_fpga_conn_create_cq(struct mlx5_fpga_conn *conn, int cq_size) in mlx5_fpga_conn_create_cq() argument
431 cq_size = roundup_pow_of_two(cq_size); in mlx5_fpga_conn_create_cq()
432 MLX5_SET(cqc, temp_cqc, log_cq_size, ilog2(cq_size)); in mlx5_fpga_conn_create_cq()
462 MLX5_SET(cqc, cqc, log_cq_size, ilog2(cq_size)); in mlx5_fpga_conn_create_cq()
/linux/drivers/infiniband/ulp/srpt/
H A Dib_srpt.h304 u32 cq_size; member
/linux/drivers/net/ethernet/pensando/ionic/
H A Dionic_debugfs.c133 debugfs_create_x32("cq_size", 0400, qcq_dentry, &qcq->cq_size); in ionic_debugfs_add_qcq()
/linux/drivers/dma/
H A Dhisi_dma.c586 size_t cq_size = sizeof(struct hisi_dma_cqe) * hdma_dev->chan_depth; in hisi_dma_alloc_qps_mem() local
598 chan->cq = dmam_alloc_coherent(dev, cq_size, &chan->cq_dma, in hisi_dma_alloc_qps_mem()
/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/sws/
H A Ddr_send.c1206 int cq_size; in mlx5dr_send_ring_alloc() local
1214 cq_size = QUEUE_SIZE + 1; in mlx5dr_send_ring_alloc()
1215 dmn->send_ring->cq = dr_create_cq(dmn->mdev, dmn->uar, cq_size); in mlx5dr_send_ring_alloc()
/linux/include/linux/qed/
H A Dqed_rdma_if.h257 u32 cq_size; member
/linux/drivers/net/ethernet/brocade/bna/
H A Dbna_tx_rx.c2138 u32 cq_size, hq_size, dq_size; in bna_rx_res_req() local
2149 cq_size = cq_depth * BFI_CQ_WI_SIZE; in bna_rx_res_req()
2150 cq_size = ALIGN(cq_size, PAGE_SIZE); in bna_rx_res_req()
2151 cpage_count = SIZE_TO_PAGES(cq_size); in bna_rx_res_req()
/linux/drivers/infiniband/hw/cxgb4/
H A Dt4.h832 u64 cq_size; member
/linux/include/uapi/drm/
H A Dhabanalabs_accel.h1193 __u32 cq_size; member
/linux/drivers/net/ethernet/qlogic/qed/
H A Dqed_rdma.c1079 p_ramrod->max_cqes = cpu_to_le32(params->cq_size); in qed_rdma_create_cq()
/linux/drivers/infiniband/hw/bnxt_re/
H A Dqplib_fp.c2227 req.cq_size = cpu_to_le32(cq->max_wqe); in bnxt_qplib_create_cq()