/freebsd/contrib/ofed/libirdma/ |
H A D | abi.h | 60 __u32 cq_size; member 72 __u32 cq_size; member
|
H A D | irdma_uverbs.c | 391 static inline size_t get_cq_total_bytes(u32 cq_size) { in get_cq_total_bytes() argument 392 return roundup(cq_size * sizeof(struct irdma_cqe), IRDMA_HW_PAGE_SIZE); in get_cq_total_bytes() 453 info.cq_size = get_cq_size(attr_ex->cqe, hw_rev); in ucreate_cq() 454 total_size = get_cq_total_bytes(info.cq_size); in ucreate_cq() 510 attr_ex->cqe = info.cq_size; in ucreate_cq() 581 irdma_free_hw_buf(cq_buf->cq.cq_base, get_cq_total_bytes(cq_buf->cq.cq_size)); in irdma_free_cq_buf() 1987 size_t cq_size; in irdma_uresize_cq() local 2003 if (cqe_needed == iwucq->cq.cq_size) in irdma_uresize_cq() 2006 cq_size = get_cq_total_bytes(cqe_needed); in irdma_uresize_cq() 2007 cq_pages = cq_size >> IRDMA_HW_PAGE_SHIFT; in irdma_uresize_cq() [all …]
|
H A D | irdma-abi.h | 129 __u32 cq_size; member
|
H A D | irdma_user.h | 411 u32 cq_size; member 450 u32 cq_size; member
|
H A D | irdma_uk.c | 1059 irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) in irdma_uk_cq_resize() argument 1062 cq->cq_size = cq_size; in irdma_uk_cq_resize() 1063 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_resize() 1780 cq->cq_size = info->cq_size; in irdma_uk_cq_init() 1785 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_init()
|
/freebsd/sys/dev/mana/ |
H A D | hw_channel.c | 407 uint32_t eq_size, cq_size; in mana_hwc_create_cq() local 414 cq_size = roundup_pow_of_two(GDMA_CQE_SIZE * q_depth); in mana_hwc_create_cq() 415 if (cq_size < MINIMUM_SUPPORTED_PAGE_SIZE) in mana_hwc_create_cq() 416 cq_size = MINIMUM_SUPPORTED_PAGE_SIZE; in mana_hwc_create_cq() 428 err = mana_hwc_create_gdma_cq(hwc, cq_size, hwc_cq, in mana_hwc_create_cq()
|
H A D | mana_en.c | 1226 req.cq_size = cq_spec->queue_size; in mana_create_wq_obj() 2078 uint32_t cq_size; in mana_create_txq() local 2098 cq_size = apc->tx_queue_size * COMP_ENTRY_SIZE; in mana_create_txq() 2099 cq_size = ALIGN(cq_size, PAGE_SIZE); in mana_create_txq() 2131 spec.queue_size = cq_size; in mana_create_txq() 2292 struct mana_rxq *rxq, uint32_t *rxq_size, uint32_t *cq_size) in mana_alloc_rx_wqe() argument 2304 *cq_size = 0; in mana_alloc_rx_wqe() 2337 *cq_size += COMP_ENTRY_SIZE; in mana_alloc_rx_wqe() 2371 uint32_t cq_size, rq_size; in mana_create_rxq() local 2414 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size); in mana_create_rxq() [all …]
|
H A D | mana.h | 659 uint32_t cq_size; member
|
/freebsd/sys/dev/irdma/ |
H A D | irdma_kcompat.c | 933 #define IRDMA_CREATE_CQ_MIN_RESP_LEN offsetofend(struct irdma_create_cq_resp, cq_size) in irdma_create_cq() 969 ukinfo->cq_size = max(entries, 4); in irdma_create_cq() 973 iwcq->ibcq.cqe = info.cq_uk_init_info.cq_size; in irdma_create_cq() 1048 ukinfo->cq_size = entries; in irdma_create_cq() 1051 rsize = info.cq_uk_init_info.cq_size * sizeof(struct irdma_extended_cqe); in irdma_create_cq() 1053 rsize = info.cq_uk_init_info.cq_size * sizeof(struct irdma_cqe); in irdma_create_cq() 1078 info.shadow_read_threshold = min(info.cq_uk_init_info.cq_size / 2, in irdma_create_cq() 1108 resp.cq_size = info.cq_uk_init_info.cq_size; in irdma_create_cq()
|
H A D | irdma-abi.h | 129 __u32 cq_size; member
|
H A D | irdma_uk.c | 975 irdma_uk_cq_resize(struct irdma_cq_uk *cq, void *cq_base, int cq_size) in irdma_uk_cq_resize() argument 978 cq->cq_size = cq_size; in irdma_uk_cq_resize() 979 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_resize() 1698 cq->cq_size = info->cq_size; in irdma_uk_cq_init() 1703 IRDMA_RING_INIT(cq->cq_ring, cq->cq_size); in irdma_uk_cq_init()
|
H A D | irdma_puda.c | 772 set_64bit_val(wqe, IRDMA_BYTE_0, cq->cq_uk.cq_size); in irdma_puda_cq_wqe() 823 cqsize = rsrc->cq_size * (sizeof(struct irdma_cqe)); in irdma_puda_cq_create() 835 info.shadow_read_threshold = rsrc->cq_size >> 2; in irdma_puda_cq_create() 840 init_info->cq_size = rsrc->cq_size; in irdma_puda_cq_create() 1159 rsrc->cq_size = info->rq_size + info->sq_size; in irdma_puda_create_rsrc() 1162 rsrc->cq_size += info->rq_size; in irdma_puda_create_rsrc()
|
H A D | irdma_user.h | 512 u32 cq_size; member 551 u32 cq_size; member
|
H A D | irdma_puda.h | 149 u32 cq_size; member
|
H A D | irdma_verbs.h | 158 u16 cq_size; member
|
H A D | irdma_verbs.c | 1489 info.cq_size = max(entries, 4); in irdma_resize_cq() 1491 if (info.cq_size == iwcq->sc_cq.cq_uk.cq_size - 1) in irdma_resize_cq() 1527 rsize = info.cq_size * sizeof(struct irdma_cqe); in irdma_resize_cq() 1576 ibcq->cqe = info.cq_size - 1; in irdma_resize_cq()
|
/freebsd/sys/dev/mlx5/mlx5_fpga/ |
H A D | mlx5fpga_conn.c | 425 static int mlx5_fpga_conn_create_cq(struct mlx5_fpga_conn *conn, int cq_size) in mlx5_fpga_conn_create_cq() argument 438 cq_size = roundup_pow_of_two(cq_size); in mlx5_fpga_conn_create_cq() 439 MLX5_SET(cqc, temp_cqc, log_cq_size, ilog2(cq_size)); in mlx5_fpga_conn_create_cq() 467 MLX5_SET(cqc, cqc, log_cq_size, ilog2(cq_size)); in mlx5_fpga_conn_create_cq()
|
/freebsd/sys/dev/qlnx/qlnxe/ |
H A D | ecore_roce_api.h | 314 u32 cq_size; member 330 u32 cq_size; member
|
H A D | ecore_rdma_api.h | 407 u32 cq_size; member 452 u32 cq_size; member
|
/freebsd/sys/dev/cxgbe/iw_cxgbe/ |
H A D | device.c | 159 rdev->status_page->cq_size = sc->vres.cq.size; in c4iw_rdev_open()
|
H A D | t4.h | 715 u64 cq_size; member
|
/freebsd/sys/dev/mlx4/mlx4_ib/ |
H A D | mlx4_ib_mad.c | 1943 int ret, cq_size; in create_pv_resources() local 1969 cq_size = 2 * MLX4_NUM_TUNNEL_BUFS; in create_pv_resources() 1971 cq_size *= 2; in create_pv_resources() 1973 cq_attr.cqe = cq_size; in create_pv_resources()
|
/freebsd/contrib/ofed/libcxgb4/ |
H A D | dev.c | 194 context->status_page->cq_size; in c4iw_alloc_context()
|
H A D | t4.h | 732 u64 cq_size; member
|
/freebsd/sys/ofed/drivers/infiniband/core/ |
H A D | ib_mad.c | 3111 int ret, cq_size; in ib_mad_port_open() local 3136 cq_size = mad_sendq_size + mad_recvq_size; in ib_mad_port_open() 3139 cq_size *= 2; in ib_mad_port_open() 3148 port_priv->cq = ib_alloc_cq(port_priv->device, port_priv, cq_size, 0, in ib_mad_port_open()
|