| /linux/drivers/net/ethernet/mellanox/mlx4/ |
| H A D | cq.c | 290 static int mlx4_init_user_cqes(void *buf, int entries, int cqe_size) in mlx4_init_user_cqes() argument 292 int entries_per_copy = PAGE_SIZE / cqe_size; in mlx4_init_user_cqes() 318 array_size(entries, cqe_size)) ? in mlx4_init_user_cqes() 330 int cqe_size) in mlx4_init_kernel_cqes() argument 335 memset(buf->direct.buf, 0xcc, entries * cqe_size); in mlx4_init_kernel_cqes() 395 dev->caps.cqe_size); in mlx4_cq_alloc() 400 dev->caps.cqe_size); in mlx4_cq_alloc()
|
| H A D | fw.h | 203 u16 cqe_size; /* For use only when CQE stride feature enabled */ member
|
| H A D | en_rx.c | 712 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq() 934 cqe = mlx4_en_get_cqe(cq->buf, index, priv->cqe_size) + factor; in mlx4_en_process_rx_cq()
|
| H A D | fw.c | 1944 dev->caps.cqe_size = 64; in mlx4_INIT_HCA() 1947 dev->caps.cqe_size = 32; in mlx4_INIT_HCA() 1954 dev->caps.cqe_size = cache_line_size(); in mlx4_INIT_HCA() 2178 param->cqe_size = 1 << ((byte_field & in mlx4_QUERY_HCA()
|
| H A D | en_netdev.c | 1682 cqe = mlx4_en_get_cqe(cq->buf, j, priv->cqe_size) + in mlx4_en_start_port() 3230 priv->cqe_factor = (mdev->dev->caps.cqe_size == 64) ? 1 : 0; in mlx4_en_init_netdev() 3231 priv->cqe_size = mdev->dev->caps.cqe_size; in mlx4_en_init_netdev()
|
| H A D | main.c | 1049 dev->caps.cqe_size = 64; in mlx4_slave_cap() 1052 dev->caps.cqe_size = 32; in mlx4_slave_cap() 1061 dev->caps.cqe_size = hca_param->cqe_size; in mlx4_slave_cap()
|
| /linux/include/uapi/rdma/ |
| H A D | hns-abi.h | 42 __u32 cqe_size; member 132 __u32 cqe_size; member
|
| H A D | mlx4-abi.h | 71 __u32 cqe_size; member
|
| /linux/drivers/net/ethernet/marvell/octeontx2/nic/ |
| H A D | otx2_txrx.h | 146 u16 cqe_size; member
|
| H A D | otx2_ethtool.c | 382 kernel_ring->cqe_size = pfvf->hw.xqe_size; in otx2_get_ringparam() 393 u32 xqe_size = kernel_ring->cqe_size; in otx2_set_ringparam()
|
| H A D | otx2_common.c | 1097 cq->cqe_size = pfvf->qset.xqe_size; in otx2_cq_init() 1100 err = qmem_alloc(pfvf->dev, &cq->cqe, cq->cqe_cnt, cq->cqe_size); in otx2_cq_init()
|
| H A D | otx2_txrx.c | 25 #define CQE_ADDR(CQ, idx) ((CQ)->cqe_base + ((CQ)->cqe_size * (idx)))
|
| /linux/drivers/net/ethernet/huawei/hinic/ |
| H A D | hinic_hw_qp.c | 320 size_t cqe_dma_size, cqe_size; in alloc_rq_cqe() local 324 cqe_size = wq->q_depth * sizeof(*rq->cqe); in alloc_rq_cqe() 325 rq->cqe = vzalloc(cqe_size); in alloc_rq_cqe()
|
| /linux/drivers/infiniband/hw/ocrdma/ |
| H A D | ocrdma_hw.c | 1785 u32 hw_pages, cqe_size, page_size, cqe_count; in ocrdma_mbx_create_cq() local 1798 cqe_size = OCRDMA_DPP_CQE_SIZE; in ocrdma_mbx_create_cq() 1803 cqe_size = sizeof(struct ocrdma_cqe); in ocrdma_mbx_create_cq() 1807 cq->len = roundup(max_hw_cqe * cqe_size, OCRDMA_MIN_Q_PAGE_SIZE); in ocrdma_mbx_create_cq() 1827 cqe_count = cq->len / cqe_size; in ocrdma_mbx_create_cq() 1856 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size); in ocrdma_mbx_create_cq() 1858 cmd->cmd.pdid_cqecnt = (cq->len / cqe_size) - 1; in ocrdma_mbx_create_cq()
|
| /linux/drivers/scsi/bnx2i/ |
| H A D | bnx2i.h | 661 u32 cqe_size; member
|
| H A D | bnx2i_hwi.c | 171 if (cq_index > ep->qp.cqe_size * 2) in bnx2i_arm_cq_event_coalescing() 172 cq_index -= ep->qp.cqe_size * 2; in bnx2i_arm_cq_event_coalescing() 1123 ep->qp.cqe_size = hba->max_cqes; in bnx2i_alloc_qp_resc() 2063 if (qp->cqe_exp_seq_sn == (qp->cqe_size * 2 + 1)) in bnx2i_process_new_cqes()
|
| /linux/drivers/infiniband/hw/irdma/ |
| H A D | verbs.c | 2035 u8 cqe_size; in irdma_resize_cq() local 2061 cqe_size = iwcq->sc_cq.cq_uk.avoid_mem_cflct ? 64 : 32; in irdma_resize_cq() 2062 if (entries * cqe_size == IRDMA_HW_PAGE_SIZE) in irdma_resize_cq() 2495 u8 cqe_size; in irdma_create_cq() local 2522 cqe_size = cqe_64byte_ena ? 64 : 32; in irdma_create_cq() 2602 if (entries * cqe_size == IRDMA_HW_PAGE_SIZE) in irdma_create_cq()
|
| /linux/io_uring/ |
| H A D | io_uring.c | 533 size_t cqe_size = sizeof(struct io_uring_cqe); in __io_cqring_overflow_flush() local 543 cqe_size <<= 1; in __io_cqring_overflow_flush() 551 memcpy(cqe, &ocqe->cqe, cqe_size); in __io_cqring_overflow_flush()
|
| /linux/drivers/scsi/lpfc/ |
| H A D | lpfc_sli4.h | 541 uint32_t cqe_size; member
|
| /linux/include/linux/mlx4/ |
| H A D | device.h | 618 u32 cqe_size; member
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 3777 return hns_roce_buf_offset(hr_cq->mtr.kmem, n * hr_cq->cqe_size); in get_cqe_v2() 3838 memcpy(dest, cqe, hr_cq->cqe_size); in __hns_roce_v2_cq_clean() 3872 if (hr_cq->cqe_size == HNS_ROCE_V3_CQE_SIZE) in hns_roce_v2_write_cqc() 4047 cq->cqe_size, false); in get_cqe_status()
|
| /linux/drivers/vdpa/mlx5/net/ |
| H A D | mlx5_vnet.c | 66 int cqe_size; member 319 buf->cqe_size = MLX5_VDPA_CQE_SIZE; in cq_frag_buf_alloc()
|