| /linux/drivers/infiniband/hw/hns/ |
| H A D | hns_roce_srq.c | 270 u32 max_sge = dev->caps.max_srq_sges; in proc_srq_sge() local 273 return max_sge; in proc_srq_sge() 282 max_sge = roundup_pow_of_two(max_sge + 1); in proc_srq_sge() 286 return max_sge; in proc_srq_sge() 295 u32 max_sge; in set_srq_basic_param() local 297 max_sge = proc_srq_sge(hr_dev, srq, !!udata); in set_srq_basic_param() 299 attr->max_sge > max_sge || !attr->max_sge) { in set_srq_basic_param() 302 attr->max_wr, attr->max_sge); in set_srq_basic_param() 308 srq->max_gs = roundup_pow_of_two(attr->max_sge + srq->rsv_sge); in set_srq_basic_param() 311 attr->max_sge = srq->max_gs - srq->rsv_sge; in set_srq_basic_param()
|
| /linux/drivers/infiniband/sw/rdmavt/ |
| H A D | srq.c | 45 if (srq_init_attr->attr.max_sge == 0 || in rvt_create_srq() 46 srq_init_attr->attr.max_sge > dev->dparms.props.max_srq_sge || in rvt_create_srq() 55 srq->rq.max_sge = srq_init_attr->attr.max_sge; in rvt_create_srq() 56 sz = sizeof(struct ib_sge) * srq->rq.max_sge + in rvt_create_srq() 145 srq->rq.max_sge * sizeof(struct ib_sge); in rvt_modify_srq() 284 attr->max_sge = srq->rq.max_sge; in rvt_query_srq()
|
| /linux/drivers/infiniband/hw/vmw_pvrdma/ |
| H A D | pvrdma_srq.c | 86 srq_attr->max_sge = resp->attrs.max_sge; in pvrdma_query_srq() 128 init_attr->attr.max_sge > dev->dsr->caps.max_srq_sge) { in pvrdma_create_srq() 179 cmd->attrs.max_sge = init_attr->attr.max_sge; in pvrdma_create_srq()
|
| /linux/drivers/infiniband/core/ |
| H A D | uverbs_std_types_wq.c | 49 ret = uverbs_copy_from(&wq_init_attr.max_sge, attrs, in UVERBS_HANDLER() 99 &wq_init_attr.max_sge, in UVERBS_HANDLER() 100 sizeof(wq_init_attr.max_sge)); in UVERBS_HANDLER()
|
| H A D | uverbs_std_types_srq.c | 50 ret = uverbs_copy_from(&attr.attr.max_sge, attrs, in UVERBS_HANDLER() 123 &attr.attr.max_sge, in UVERBS_HANDLER() 124 sizeof(attr.attr.max_sge)); in UVERBS_HANDLER()
|
| H A D | rw.c | 291 u32 max_sge = dir == DMA_TO_DEVICE ? qp->max_write_sge : in rdma_rw_init_map_wrs() local 296 ctx->nr_ops = DIV_ROUND_UP(sg_cnt, max_sge); in rdma_rw_init_map_wrs() 308 u32 nr_sge = min(sg_cnt, max_sge); in rdma_rw_init_map_wrs() 408 u32 max_sge = dir == DMA_TO_DEVICE ? qp->max_write_sge : in rdma_rw_init_map_wrs_bvec() local 413 u32 nr_ops = DIV_ROUND_UP(nr_bvec, max_sge); in rdma_rw_init_map_wrs_bvec() 432 u32 nr_sge = min(nr_bvec - mapped_bvecs, max_sge); in rdma_rw_init_map_wrs_bvec()
|
| H A D | uverbs_cmd.c | 360 resp->max_sge = min(attr->max_send_sge, attr->max_recv_sge); in copy_query_dev_fields() 2949 wq_init_attr.max_sge = cmd.max_sge; in ib_uverbs_ex_create_wq() 2982 resp.max_sge = wq_init_attr.max_sge; in ib_uverbs_ex_create_wq() 3446 attr.attr.max_sge = cmd->max_sge; in __uverbs_create_xsrq() 3479 resp.max_sge = attr.attr.max_sge; in __uverbs_create_xsrq() 3516 xcmd.max_sge = cmd.max_sge; in ib_uverbs_create_srq() 3588 resp.max_sge = attr.max_sge; in ib_uverbs_query_srq() 3671 resp.tm_caps.max_sge = attr.tm_caps.max_sge; in ib_uverbs_ex_query_device() 4006 max_sge,
|
| /linux/include/uapi/rdma/ |
| H A D | ib_user_verbs.h | 199 __u32 max_sge; member 283 __u32 max_sge; member 1194 __u32 max_sge; member 1205 __u32 max_sge; member 1216 __u32 max_sge; member 1238 __u32 max_sge; member 1260 __u32 max_sge; member 1270 __u32 max_sge; member
|
| /linux/drivers/infiniband/sw/rxe/ |
| H A D | rxe_qp.c | 245 qp->sq.max_sge = wqe_size / sizeof(struct ib_sge); in rxe_init_sq() 270 init->cap.max_send_sge = qp->sq.max_sge; in rxe_init_sq() 336 qp->rq.max_sge = init->cap.max_recv_sge; in rxe_init_rq() 338 qp->rq.max_sge*sizeof(struct ib_sge); in rxe_init_rq() 471 init->cap.max_send_sge = qp->sq.max_sge; in rxe_qp_to_init() 476 init->cap.max_recv_sge = qp->rq.max_sge; in rxe_qp_to_init() 812 attr->cap.max_send_sge = qp->sq.max_sge; in rxe_qp_to_attr() 817 attr->cap.max_recv_sge = qp->rq.max_sge; in rxe_qp_to_attr()
|
| /linux/drivers/infiniband/sw/siw/ |
| H A D | siw_verbs.c | 154 attr->max_recv_sge = sdev->attrs.max_sge; in siw_query_device() 156 attr->max_send_sge = sdev->attrs.max_sge; in siw_query_device() 1415 u32 max_sge) in siw_alloc_mr() argument 1432 if (max_sge > SIW_MAX_SGE_PBL) { in siw_alloc_mr() 1433 siw_dbg_pd(pd, "too many sge's: %d\n", max_sge); in siw_alloc_mr() 1437 pbl = siw_pbl_alloc(max_sge); in siw_alloc_mr() 1449 rv = siw_mr_add_mem(mr, pd, pbl, 0, max_sge * PAGE_SIZE, 0); in siw_alloc_mr() 1613 attrs->max_sge > SIW_MAX_SGE || attrs->srq_limit > attrs->max_wr) { in siw_create_srq() 1617 srq->max_sge = attrs->max_sge; in siw_create_srq() 1728 attrs->max_sge = srq->max_sge; in siw_query_srq() [all …]
|
| H A D | siw.h | 57 int max_sge; member 244 u32 max_sge; member
|
| /linux/net/sunrpc/xprtrdma/ |
| H A D | frwr_ops.c | 176 unsigned int max_sge; in frwr_query_device() local 185 max_sge = min_t(unsigned int, attrs->max_send_sge, in frwr_query_device() 187 if (max_sge < RPCRDMA_MIN_SEND_SGES) { in frwr_query_device() 188 pr_err("rpcrdma: HCA provides only %u send SGEs\n", max_sge); in frwr_query_device() 191 ep->re_attr.cap.max_send_sge = max_sge; in frwr_query_device()
|
| /linux/net/rds/ |
| H A D | ib.c | 154 rds_ibdev->max_sge = min(device->attrs.max_send_sge, RDS_IB_MAX_SGE); in rds_ib_add_one() 208 device->attrs.max_mr, rds_ibdev->max_wrs, rds_ibdev->max_sge, in rds_ib_add_one() 322 iinfo->max_send_sge = rds_ibdev->max_sge; in rds_ib_conn_info_visitor() 359 iinfo6->max_send_sge = rds_ibdev->max_sge; in rds6_ib_conn_info_visitor()
|
| H A D | ib_send.c | 888 u32 max_sge = ic->rds_ibdev->max_sge; in rds_ib_xmit_rdma() 926 i = DIV_ROUND_UP(op->op_count, max_sge); in rds_ib_xmit_rdma() 956 if (num_sge > max_sge) { in rds_ib_xmit_rdma() 957 send->s_rdma_wr.wr.num_sge = max_sge; in rds_ib_xmit_rdma() 958 num_sge -= max_sge; in rds_ib_xmit_rdma() 862 u32 max_sge = ic->rds_ibdev->max_sge; rds_ib_xmit_rdma() local
|
| /linux/drivers/infiniband/hw/bnxt_re/ |
| H A D | ib_verbs.c | 1103 sq->max_sge = BNXT_STATIC_MAX_SGE; in bnxt_re_setup_swqe_size() 1105 sq->wqe_size = bnxt_re_get_wqe_size(ilsize, sq->max_sge); in bnxt_re_setup_swqe_size() 1260 qp->qplib_qp.sq.max_sge = 2; in bnxt_re_create_shadow_qp() 1272 qp->qplib_qp.rq.max_sge = qp1_qp->rq.max_sge; in bnxt_re_create_shadow_qp() 1322 rq->max_sge = init_attr->cap.max_recv_sge; in bnxt_re_init_rq_attr() 1323 if (rq->max_sge > dev_attr->max_qp_sges) in bnxt_re_init_rq_attr() 1324 rq->max_sge = dev_attr->max_qp_sges; in bnxt_re_init_rq_attr() 1325 init_attr->cap.max_recv_sge = rq->max_sge; in bnxt_re_init_rq_attr() 1326 rq->wqe_size = bnxt_re_setup_rwqe_size(qplqp, rq->max_sge, in bnxt_re_init_rq_attr() 1353 qplqp->rq.max_sge = dev_attr->max_qp_sges; in bnxt_re_adjust_gsi_rq_attr() [all …]
|
| /linux/drivers/infiniband/hw/mlx5/ |
| H A D | srq.c | 212 init_attr->attr.max_sge > max_sge_sz) { in mlx5_ib_create_srq() 215 init_attr->attr.max_sge, max_sge_sz); in mlx5_ib_create_srq() 226 srq->msrq.max_gs = init_attr->attr.max_sge; in mlx5_ib_create_srq() 371 srq_attr->max_sge = srq->msrq.max_gs; in mlx5_ib_query_srq()
|
| /linux/drivers/infiniband/hw/mlx4/ |
| H A D | srq.c | 95 init_attr->attr.max_sge > dev->dev->caps.max_srq_sge) in mlx4_ib_create_srq() 101 srq->msrq.max_gs = init_attr->attr.max_sge; in mlx4_ib_create_srq() 263 srq_attr->max_sge = srq->msrq.max_gs; in mlx4_ib_query_srq()
|
| /linux/drivers/vfio/pci/pds/ |
| H A D | dirty.c | 150 u32 max_sge; in pds_vfio_dirty_alloc_sgl() local 152 max_sge = DIV_ROUND_UP(page_count, PAGE_SIZE * 8); in pds_vfio_dirty_alloc_sgl() 153 sgl_size = max_sge * sizeof(struct pds_lm_sg_elem); in pds_vfio_dirty_alloc_sgl() 166 region->num_sge = max_sge; in pds_vfio_dirty_alloc_sgl()
|
| /linux/drivers/infiniband/hw/mthca/ |
| H A D | mthca_srq.c | 216 attr->max_sge > dev->limits.max_srq_sge) in mthca_alloc_srq() 220 srq->max_gs = attr->max_sge; in mthca_alloc_srq() 298 attr->max_sge = srq->max_gs; in mthca_alloc_srq() 422 srq_attr->max_sge = srq->max_gs; in mthca_query_srq()
|
| /linux/drivers/infiniband/hw/ionic/ |
| H A D | ionic_controlpath.c | 1780 struct ionic_tbl_buf *buf, int max_wr, int max_sge, in ionic_qp_sq_init() argument 1806 if (max_sge < 1) in ionic_qp_sq_init() 1809 if (max_sge > min(ionic_v1_send_wqe_max_sge(dev->lif_cfg.max_stride, 0, in ionic_qp_sq_init() 1844 qp->sq_spec = ionic_v1_use_spec_sge(max_sge, sq_spec); in ionic_qp_sq_init() 1848 max_sge); in ionic_qp_sq_init() 1851 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 1860 wqe_size = ionic_v1_send_wqe_min_size(max_sge, max_data, in ionic_qp_sq_init() 2006 struct ionic_tbl_buf *buf, int max_wr, int max_sge, in ionic_qp_rq_init() argument 2029 if (max_sge < 1) in ionic_qp_rq_init() 2032 if (max_sge > min(ionic_v1_recv_wqe_max_sge(dev->lif_cfg.max_stride, 0, false), in ionic_qp_rq_init() [all …]
|
| /linux/drivers/scsi/elx/efct/ |
| H A D | efct_lio.h | 37 u32 max_sge; member
|
| /linux/drivers/infiniband/hw/qedr/ |
| H A D | verbs.c | 106 srq_attr->max_sge = qattr->max_sge; in qedr_query_srq() 142 attr->max_send_sge = qattr->max_sge; in qedr_query_device() 143 attr->max_recv_sge = qattr->max_sge; in qedr_query_device() 144 attr->max_sge_rd = qattr->max_sge; in qedr_query_device() 1222 if (attrs->cap.max_send_sge > qattr->max_sge) { in qedr_check_qp_attrs() 1225 attrs->cap.max_send_sge, qattr->max_sge); in qedr_check_qp_attrs() 1229 if (attrs->cap.max_recv_sge > qattr->max_sge) { in qedr_check_qp_attrs() 1232 attrs->cap.max_recv_sge, qattr->max_sge); in qedr_check_qp_attrs() 1434 if (attrs->attr.max_sge > qattr->max_sge) { in qedr_check_srq_params() 1437 attrs->attr.max_sge, qattr->max_sge); in qedr_check_srq_params() [all …]
|
| /linux/drivers/scsi/csiostor/ |
| H A D | csio_scsi.h | 149 uint8_t max_sge; /* Max SGE */ member
|
| /linux/include/rdma/ |
| H A D | rdmavt_qp.h | 262 u8 max_sge; member 550 rq->max_sge * sizeof(struct ib_sge)) * n); in rvt_get_rwqe_ptr()
|
| /linux/drivers/infiniband/ulp/rtrs/ |
| H A D | rtrs.c | 256 u32 max_send_wr, u32 max_recv_wr, u32 max_sge) in create_qp() argument 267 init_attr.cap.max_send_sge = max_sge; in create_qp()
|