| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | srq.c | 45 if (srq_init_attr->attr.max_sge == 0 || in rvt_create_srq() 46 srq_init_attr->attr.max_sge > dev->dparms.props.max_srq_sge || in rvt_create_srq() 55 srq->rq.max_sge = srq_init_attr->attr.max_sge; in rvt_create_srq() 56 sz = sizeof(struct ib_sge) * srq->rq.max_sge + in rvt_create_srq() 145 srq->rq.max_sge * sizeof(struct ib_sge); in rvt_modify_srq() 284 attr->max_sge = srq->rq.max_sge; in rvt_query_srq()
|
| /linux/drivers/infiniband/core/ |
| H A D | uverbs_std_types_wq.c | 49 ret = uverbs_copy_from(&wq_init_attr.max_sge, attrs, in UVERBS_HANDLER() 99 &wq_init_attr.max_sge, in UVERBS_HANDLER() 100 sizeof(wq_init_attr.max_sge)); in UVERBS_HANDLER()
|
| H A D | uverbs_std_types_srq.c | 50 ret = uverbs_copy_from(&attr.attr.max_sge, attrs, in UVERBS_HANDLER() 123 &attr.attr.max_sge, in UVERBS_HANDLER() 124 sizeof(attr.attr.max_sge)); in UVERBS_HANDLER()
|
| H A D | rw.c | 291 u32 max_sge = dir == DMA_TO_DEVICE ? qp->max_write_sge : in rdma_rw_init_map_wrs() local 296 ctx->nr_ops = DIV_ROUND_UP(sg_cnt, max_sge); in rdma_rw_init_map_wrs() 308 u32 nr_sge = min(sg_cnt, max_sge); in rdma_rw_init_map_wrs() 408 u32 max_sge = dir == DMA_TO_DEVICE ? qp->max_write_sge : in rdma_rw_init_map_wrs_bvec() local 413 u32 nr_ops = DIV_ROUND_UP(nr_bvec, max_sge); in rdma_rw_init_map_wrs_bvec() 432 u32 nr_sge = min(nr_bvec - mapped_bvecs, max_sge); in rdma_rw_init_map_wrs_bvec()
|
| /linux/include/uapi/rdma/ |
| H A D | ib_user_verbs.h | 199 __u32 max_sge; member 283 __u32 max_sge; member 1194 __u32 max_sge; member 1205 __u32 max_sge; member 1216 __u32 max_sge; member 1238 __u32 max_sge; member 1260 __u32 max_sge; member 1270 __u32 max_sge; member
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_qp.c | 245 qp->sq.max_sge = wqe_size / sizeof(struct ib_sge); in rxe_init_sq() 270 init->cap.max_send_sge = qp->sq.max_sge; in rxe_init_sq() 336 qp->rq.max_sge = init->cap.max_recv_sge; in rxe_init_rq() 338 qp->rq.max_sge*sizeof(struct ib_sge); in rxe_init_rq() 471 init->cap.max_send_sge = qp->sq.max_sge; in rxe_qp_to_init() 476 init->cap.max_recv_sge = qp->rq.max_sge; in rxe_qp_to_init() 812 attr->cap.max_send_sge = qp->sq.max_sge; in rxe_qp_to_attr() 817 attr->cap.max_recv_sge = qp->rq.max_sge; in rxe_qp_to_attr()
|
| /linux/net/rds/ |
| H A D | ib.c | 154 rds_ibdev->max_sge = min(device->attrs.max_send_sge, RDS_IB_MAX_SGE); in rds_ib_add_one() 207 rdsdebug("RDS/IB: max_mr = %d, max_wrs = %d, max_sge = %d, max_1m_mrs = %d, max_8k_mrs = %d\n", in rds_ib_add_one() 208 device->attrs.max_mr, rds_ibdev->max_wrs, rds_ibdev->max_sge, in rds_ib_add_one() 322 iinfo->max_send_sge = rds_ibdev->max_sge; in rds_ib_conn_info_visitor() 359 iinfo6->max_send_sge = rds_ibdev->max_sge; in rds6_ib_conn_info_visitor()
|
| H A D | ib_send.c | 888 u32 max_sge = ic->rds_ibdev->max_sge; in rds_ib_xmit_rdma() 926 i = DIV_ROUND_UP(op->op_count, max_sge); in rds_ib_xmit_rdma() 956 if (num_sge > max_sge) { in rds_ib_xmit_rdma() 957 send->s_rdma_wr.wr.num_sge = max_sge; in rds_ib_xmit_rdma() 958 num_sge -= max_sge; in rds_ib_xmit_rdma() 862 u32 max_sge = ic->rds_ibdev->max_sge; rds_ib_xmit_rdma() local
|
| H A D | ib.h | 256 int max_sge; member
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | frwr_ops.c | 176 unsigned int max_sge; in frwr_query_device() local 185 max_sge = min_t(unsigned int, attrs->max_send_sge, in frwr_query_device() 187 if (max_sge < RPCRDMA_MIN_SEND_SGES) { in frwr_query_device() 188 pr_err("rpcrdma: HCA provides only %u send SGEs\n", max_sge); in frwr_query_device() 191 ep->re_attr.cap.max_send_sge = max_sge; in frwr_query_device()
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 150 u32 max_sge; in pds_vfio_dirty_alloc_sgl() local 152 max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8); in pds_vfio_dirty_alloc_sgl() 153 sgl_size = max_sge * sizeof(struct pds_lm_sg_elem); in pds_vfio_dirty_alloc_sgl() 166 region->num_sge = max_sge; in pds_vfio_dirty_alloc_sgl()
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_srq.c | 216 attr->max_sge > dev->limits.max_srq_sge) in mthca_alloc_srq() 220 srq->max_gs = attr->max_sge; in mthca_alloc_srq() 298 attr->max_sge = srq->max_gs; in mthca_alloc_srq() 422 srq_attr->max_sge = srq->max_gs; in mthca_query_srq()
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_controlpath.c | 1782 struct ionic_tbl_buf *buf, int max_wr, int max_sge, in ionic_qp_sq_init() argument 1808 if (max_sge < 1) in ionic_qp_sq_init() 1811 if (max_sge > min(ionic_v1_send_wqe_max_sge(dev->lif_cfg.max_stride, 0, in ionic_qp_sq_init() 1846 qp->sq_spec = ionic_v1_use_spec_sge(max_sge, sq_spec); in ionic_qp_sq_init() 1850 max_sge); in ionic_qp_sq_init() 1853 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 1862 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 2008 struct ionic_tbl_buf *buf, int max_wr, int max_sge, in ionic_qp_rq_init() argument 2031 if (max_sge < 1) in ionic_qp_rq_init() 2034 if (max_sge > min(ionic_v1_recv_wqe_max_sge(dev->lif_cfg.max_stride, 0, false), in ionic_qp_rq_init() [all …]
|
| /linux/drivers/scsi/elx/efct/ |
| H A D | efct_lio.h | 37 u32 max_sge; member
|
| /linux/drivers/scsi/csiostor/ |
| H A D | csio_scsi.h | 149 uint8_t max_sge; /* Max SGE */ member
|
| H A D | csio_scsi.c | 614 if (dma_buf == NULL || i > scsim->max_sge) { in csio_setup_ddp() 1816 if (unlikely(nsge > scsim->max_sge)) { in csio_queuecommand() 1819 " SGEs: %d, Max SGEs: %d\n", nsge, scsim->max_sge); in csio_queuecommand() 2430 scm->max_sge = CSIO_SCSI_MAX_SGE; in csio_scsim_init()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw.h | 57 int max_sge; member 244 u32 max_sge; member
|
| H A D | siw_main.c | 334 sdev->attrs.max_sge = SIW_MAX_SGE; in siw_device_create()
|
| /linux/include/rdma/ |
| H A D | rdmavt_qp.h | 262 u8 max_sge; member 550 rq->max_sge * sizeof(struct ib_sge)) * n); in rvt_get_rwqe_ptr()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs.c | 256 u32 max_send_wr, u32 max_recv_wr, u32 max_sge) in create_qp() argument 267 init_attr.cap.max_send_sge = max_sge; in create_qp()
|
| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_hw_v2.c | 788 u32 max_sge, bool rsv) in fill_recv_sge_to_wqe() argument 808 if (cnt < max_sge) in fill_recv_sge_to_wqe() 810 (max_sge - cnt) * HNS_ROCE_SGE_SIZE); in fill_recv_sge_to_wqe() 815 u32 wqe_idx, u32 max_sge) in fill_rq_wqe() argument 820 fill_recv_sge_to_wqe(wr, wqe, max_sge, hr_qp->rq.rsv_sge); in fill_rq_wqe() 833 u32 wqe_idx, nreq, max_sge; in hns_roce_v2_post_recv() local 846 max_sge = hr_qp->rq.max_gs - hr_qp->rq.rsv_sge; in hns_roce_v2_post_recv() 855 if (unlikely(wr->num_sge > max_sge)) { in hns_roce_v2_post_recv() 857 wr->num_sge, max_sge); in hns_roce_v2_post_recv() 864 fill_rq_wqe(hr_qp, wr, wqe_idx, max_sge); in hns_roce_v2_post_recv() [all …]
|
| /linux/drivers/scsi/ |
| H A D | mvumi.c | 200 if (*sg_count > mhba->max_sge) { in mvumi_make_sgl() 203 *sg_count, mhba->max_sge); in mvumi_make_sgl() 2324 mhba->max_sge = MVUMI_MAX_SG_ENTRY; in mvumi_init_fw() 2330 mhba->max_sge = MVUMI_MAX_SG_ENTRY; in mvumi_init_fw() 2396 host->sg_tablesize = mhba->max_sge > max_sg ? max_sg : mhba->max_sge; in mvumi_io_attach()
|
| H A D | mvumi.h | 516 unsigned short max_sge; member
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_dev_api.h | 229 u32 max_sge; member
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | main.c | 603 attr->max_sge = qed_attr->max_sge; in qedr_set_device_attr()
|