/linux/drivers/net/ethernet/qlogic/netxen/ |
H A D | netxen_nic_ctx.c | 270 size_t rq_size, rsp_size; in nx_fw_cmd_create_rx_ctx() local 280 rq_size = in nx_fw_cmd_create_rx_ctx() 285 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_rx_ctx() 350 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_rx_ctx() 394 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in nx_fw_cmd_create_rx_ctx() 424 size_t rq_size, rsp_size; in nx_fw_cmd_create_tx_ctx() local 433 rq_size = SIZEOF_HOSTRQ_TX(nx_hostrq_tx_ctx_t); in nx_fw_cmd_create_tx_ctx() 434 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in nx_fw_cmd_create_tx_ctx() 476 cmd.req.arg3 = rq_size; in nx_fw_cmd_create_tx_ctx() 500 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in nx_fw_cmd_create_tx_ctx()
|
/linux/drivers/infiniband/hw/irdma/ |
H A D | puda.c | 127 rsrc->rx_wqe_idx = ((rsrc->rx_wqe_idx + 1) % rsrc->rq_size); in irdma_puda_replenish_rq() 649 u32 sq_size, rq_size; in irdma_puda_qp_create() local 653 rq_size = rsrc->rq_size * IRDMA_QP_WQE_MIN_SIZE; in irdma_puda_qp_create() 654 rsrc->qpmem.size = ALIGN((sq_size + rq_size + (IRDMA_SHADOW_AREA_SIZE << 3) + IRDMA_QP_CTX_SIZE), in irdma_puda_qp_create() 665 qp->hw_rq_size = irdma_get_encoded_wqe_size(rsrc->rq_size, IRDMA_QUEUE_TYPE_SQ_RQ); in irdma_puda_qp_create() 675 ukqp->shadow_area = ukqp->rq_base[rsrc->rq_size].elem; in irdma_puda_qp_create() 677 qp->shadow_area_pa = qp->rq_pa + rq_size; in irdma_puda_qp_create() 685 ukqp->rq_size = rsrc->rq_size; in irdma_puda_qp_create() 689 IRDMA_RING_INIT(ukqp->rq_ring, ukqp->rq_size); in irdma_puda_qp_create() 1013 rqwridsize = info->rq_size * 8; in irdma_puda_create_rsrc() [all …]
|
H A D | user.h | 328 u32 rq_size; member 374 u32 rq_size; member 410 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift,
|
H A D | puda.h | 91 u32 rq_size; member 114 u32 rq_size; member
|
H A D | uk.c | 1288 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, in irdma_get_rqdepth() argument 1293 *rqdepth = irdma_qp_round_up((rq_size << shift) + IRDMA_RQ_RSVD); in irdma_get_rqdepth() 1404 status = irdma_get_rqdepth(ukinfo->uk_attrs, ukinfo->rq_size, in irdma_uk_calc_depth_shift_rq() 1454 qp->rq_size = info->rq_size; in irdma_uk_qp_init() 1458 IRDMA_RING_INIT(qp->rq_ring, qp->rq_size); in irdma_uk_qp_init()
|
/linux/drivers/scsi/ |
H A D | stex.c | 346 u16 rq_size; member 360 u16 rq_size; member 411 struct req_msg *req = hba->dma_mem + hba->req_head * hba->rq_size; in stex_alloc_req() 422 hba->req_head * hba->rq_size + sizeof(struct st_msg_header)); in stex_ss_alloc_req() 548 addr = hba->dma_handle + hba->req_head * hba->rq_size; in stex_ss_send_cmd() 1062 h->req_sz = cpu_to_le16(hba->rq_size); in stex_common_handshake() 1074 status_phys = hba->dma_handle + (hba->rq_count+1) * hba->rq_size; in stex_common_handshake() 1153 h->req_sz = cpu_to_le16(hba->rq_size); in stex_ss_handshake() 1544 .rq_size = 1048, 1557 .rq_size = 1048, [all …]
|
/linux/drivers/scsi/bnx2i/ |
H A D | bnx2i_init.c | 65 unsigned int rq_size = BNX2I_RQ_WQES_DEFAULT; variable 66 module_param(rq_size, int, 0664); 67 MODULE_PARM_DESC(rq_size, "Configure RQ size");
|
H A D | bnx2i.h | 796 extern unsigned int rq_size;
|
/linux/drivers/net/ethernet/qlogic/qlcnic/ |
H A D | qlcnic_ctx.c | 250 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() local 260 rq_size = SIZEOF_HOSTRQ_RX(struct qlcnic_hostrq_rx_ctx, nrds_rings, in qlcnic_82xx_fw_cmd_create_rx_ctx() 265 addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_rx_ctx() 343 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_rx_ctx() 387 dma_free_coherent(&adapter->pdev->dev, rq_size, prq, hostrq_phys_addr); in qlcnic_82xx_fw_cmd_create_rx_ctx() 426 size_t rq_size, rsp_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() local 435 rq_size = SIZEOF_HOSTRQ_TX(struct qlcnic_hostrq_tx_ctx); in qlcnic_82xx_fw_cmd_create_tx_ctx() 436 rq_addr = dma_alloc_coherent(&adapter->pdev->dev, rq_size, in qlcnic_82xx_fw_cmd_create_tx_ctx() 489 cmd.req.arg[3] = rq_size; in qlcnic_82xx_fw_cmd_create_tx_ctx() 518 dma_free_coherent(&adapter->pdev->dev, rq_size, rq_addr, rq_phys_addr); in qlcnic_82xx_fw_cmd_create_tx_ctx()
|
/linux/include/uapi/rdma/ |
H A D | cxgb4-abi.h | 83 __u32 rq_size; member
|
/linux/drivers/infiniband/ulp/srpt/ |
H A D | ib_srpt.c | 1856 WARN_ON(ch->rq_size < 1); in srpt_create_ch_ib() 1864 ch->cq = ib_cq_pool_get(sdev->device, ch->rq_size + sq_size, -1, in srpt_create_ch_ib() 1869 ch->rq_size + sq_size, ret); in srpt_create_ch_ib() 1872 ch->cq_size = ch->rq_size + sq_size; in srpt_create_ch_ib() 1895 qp_init->cap.max_recv_wr = ch->rq_size; in srpt_create_ch_ib() 1933 for (i = 0; i < ch->rq_size; i++) in srpt_create_ch_ib() 2178 ch->sport->sdev, ch->rq_size, in srpt_release_channel_work() 2184 sdev, ch->rq_size, in srpt_release_channel_work() 2301 ch->rq_size = min(MAX_SRPT_RQ_SIZE, sdev->device->attrs.max_qp_wr); in srpt_cm_req_recv() 2312 srpt_alloc_ioctx_ring(ch->sport->sdev, ch->rq_size, in srpt_cm_req_recv() [all …]
|
H A D | ib_srpt.h | 309 int rq_size; member
|
/linux/drivers/infiniband/hw/efa/ |
H A D | efa_verbs.c | 467 qp->rq_cpu_addr, qp->rq_size, in efa_destroy_qp() 470 qp->rq_size, DMA_TO_DEVICE); in efa_destroy_qp() 535 if (qp->rq_size) { in qp_mmap_entries_setup() 551 address, qp->rq_size, in qp_mmap_entries_setup() 557 resp->rq_mmap_size = qp->rq_size; in qp_mmap_entries_setup() 720 qp->rq_size = PAGE_ALIGN(create_qp_params.rq_ring_size_in_bytes); in efa_create_qp() 721 if (qp->rq_size) { in efa_create_qp() 723 qp->rq_size, DMA_TO_DEVICE); in efa_create_qp() 731 qp->rq_cpu_addr, qp->rq_size, &qp->rq_dma_addr); in efa_create_qp() 782 if (qp->rq_size) in efa_create_qp() [all …]
|
H A D | efa.h | 116 size_t rq_size; member
|
/linux/drivers/infiniband/hw/erdma/ |
H A D | erdma_verbs.c | 55 ilog2(qp->attrs.rq_size)) | in create_qp_cmd() 493 qp->attrs.rq_size << RQE_SHIFT, in free_kernel_qp() 520 kqp->rwr_tbl = vmalloc_array(qp->attrs.rq_size, sizeof(u64)); in init_kernel_qp() 535 size = qp->attrs.rq_size << RQE_SHIFT; in init_kernel_qp() 886 qp->attrs.rq_size * RQE_SIZE)) in init_user_qp() 899 qp->attrs.rq_size << RQE_SHIFT, 0, va + rq_offset, in init_user_qp() 969 qp->attrs.rq_size = roundup_pow_of_two(attrs->cap.max_recv_wr); in erdma_create_qp() 985 uresp.num_rqe = qp->attrs.rq_size; in erdma_create_qp() 1584 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in erdma_query_qp()
|
H A D | erdma_cq.c | 155 depth = qp->attrs.rq_size; in erdma_poll_one_cqe()
|
H A D | erdma_qp.c | 545 qp->attrs.rq_size, RQE_SHIFT); in erdma_post_recv_one() 563 qp->kern_qp.rwr_tbl[qp->kern_qp.rq_pi & (qp->attrs.rq_size - 1)] = in erdma_post_recv_one()
|
H A D | erdma_verbs.h | 208 u32 rq_size; member
|
/linux/drivers/infiniband/sw/siw/ |
H A D | siw_verbs.c | 407 qp->attrs.rq_size = 0; in siw_create_qp() 421 qp->attrs.rq_size = num_rqe; in siw_create_qp() 520 qp_attr->cap.max_recv_wr = qp->attrs.rq_size; in siw_query_qp() 1003 if (qp->srq || qp->attrs.rq_size == 0) { in siw_post_receive() 1063 u32 idx = qp->rq_put % qp->attrs.rq_size; in siw_post_receive()
|
H A D | siw_qp.c | 1299 while (qp->attrs.rq_size) { in siw_rq_flush() 1301 &qp->recvq[qp->rq_get % qp->attrs.rq_size]; in siw_rq_flush()
|
H A D | siw.h | 259 u32 rq_size; member
|
/linux/drivers/vfio/pci/mlx5/ |
H A D | cmd.c | 1438 u64 rq_size) in mlx5vf_alloc_qp_recv_resources() argument 1440 unsigned int npages = DIV_ROUND_UP_ULL(rq_size, PAGE_SIZE); in mlx5vf_alloc_qp_recv_resources() 1517 u64 rq_size = SZ_2M; in mlx5vf_start_page_tracker() local 1544 max_recv_wr = DIV_ROUND_UP_ULL(rq_size, max_msg_size); in mlx5vf_start_page_tracker() 1568 rq_size); in mlx5vf_start_page_tracker()
|
/linux/block/ |
H A D | blk-mq.c | 3410 size_t rq_size, left; in blk_mq_alloc_rqs() local 3421 rq_size = round_up(sizeof(struct request) + set->cmd_size, in blk_mq_alloc_rqs() 3423 left = rq_size * depth; in blk_mq_alloc_rqs() 3442 if (order_to_size(this_order) < rq_size) in blk_mq_alloc_rqs() 3458 entries_per_page = order_to_size(this_order) / rq_size; in blk_mq_alloc_rqs() 3460 left -= to_do * rq_size; in blk_mq_alloc_rqs() 3470 p += rq_size; in blk_mq_alloc_rqs()
|
/linux/drivers/net/ethernet/microsoft/mana/ |
H A D | mana_en.c | 2194 u32 cq_size, rq_size; in mana_create_rxq() local 2220 err = mana_alloc_rx_wqe(apc, rxq, &rq_size, &cq_size); in mana_create_rxq() 2224 rq_size = MANA_PAGE_ALIGN(rq_size); in mana_create_rxq() 2231 spec.queue_size = rq_size; in mana_create_rxq()
|
/linux/drivers/infiniband/hw/bnxt_re/ |
H A D | roce_hsi.h | 479 __le32 rq_size; member 670 __le32 rq_size; member 823 __le32 rq_size; member 1758 __le32 rq_size; member
|