/freebsd/contrib/ofed/libmlx4/ |
H A D | qp.c | 61 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe() 66 return qp->buf.buf + qp->sq.offset + (n << qp->sq.wqe_shift); in get_send_wqe() 100 ctrl->fence_size = 1 << (qp->sq.wqe_shift - 4); in mlx4_qp_init_sq_ownership() 648 for (qp->sq.wqe_shift = 6; 1 << qp->sq.wqe_shift < size; in mlx4_calc_sq_wqe_size() 649 qp->sq.wqe_shift++) in mlx4_calc_sq_wqe_size() 672 for (qp->rq.wqe_shift = 4; in mlx4_alloc_qp_buf() 673 1 << qp->rq.wqe_shift < qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg); in mlx4_alloc_qp_buf() 674 qp->rq.wqe_shift++) in mlx4_alloc_qp_buf() 677 qp->buf_size = (qp->rq.wqe_cnt << qp->rq.wqe_shift) + in mlx4_alloc_qp_buf() 678 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in mlx4_alloc_qp_buf() [all …]
|
H A D | srq.c | 46 return srq->buf.buf + (n << srq->wqe_shift); in get_wqe() 141 for (srq->wqe_shift = 5; 1 << srq->wqe_shift < size; ++srq->wqe_shift) in mlx4_alloc_srq_buf() 144 buf_size = srq->max << srq->wqe_shift; in mlx4_alloc_srq_buf() 164 (void *) scatter < (void *) next + (1 << srq->wqe_shift); in mlx4_alloc_srq_buf()
|
H A D | mlx4.h | 201 int wqe_shift; member 217 int wqe_shift; member
|
H A D | verbs.c | 833 qp->sq_spare_wqes = (2048 >> qp->sq.wqe_shift) + 1; in mlx4_create_qp_ex() 870 cmd.log_sq_stride = qp->sq.wqe_shift; in mlx4_create_qp_ex()
|
/freebsd/sys/dev/mthca/ |
H A D | mthca_srq.c | 77 return srq->queue.direct.buf + (n << srq->wqe_shift); in get_wqe() 79 return srq->queue.page_list[(n << srq->wqe_shift) >> PAGE_SHIFT].buf + in get_wqe() 80 ((n << srq->wqe_shift) & (PAGE_SIZE - 1)); in get_wqe() 108 context->wqe_base_ds = cpu_to_be64(1 << (srq->wqe_shift - 4)); in mthca_tavor_init_srq_context() 133 context->logstride_usrpage = cpu_to_be32((srq->wqe_shift - 4) << 29); in mthca_arbel_init_srq_context() 143 mthca_buf_free(dev, srq->max << srq->wqe_shift, &srq->queue, in mthca_free_srq_buf() 163 err = mthca_buf_alloc(dev, srq->max << srq->wqe_shift, in mthca_alloc_srq_buf() 183 next->nda_op = htonl(((i + 1) << srq->wqe_shift) | 1); in mthca_alloc_srq_buf() 190 (void *) scatter < wqe + (1 << srq->wqe_shift); in mthca_alloc_srq_buf() 229 srq->wqe_shift = ilog2(ds); in mthca_alloc_srq() [all …]
|
H A D | mthca_qp.c | 211 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); in get_recv_wqe() 213 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + in get_recv_wqe() 214 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); in get_recv_wqe() 221 (n << qp->sq.wqe_shift); in get_send_wqe() 224 (n << qp->sq.wqe_shift)) >> in get_send_wqe() 226 ((qp->send_wqe_offset + (n << qp->sq.wqe_shift)) & in get_send_wqe() 605 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mthca_modify_qp() 609 qp_context->sq_size_stride |= qp->sq.wqe_shift - 4; in __mthca_modify_qp() 955 1 << qp->sq.wqe_shift)); in mthca_adjust_qp_caps() 962 (min(dev->limits.max_desc_sz, 1 << qp->rq.wqe_shift) - in mthca_adjust_qp_caps() [all …]
|
H A D | mthca_provider.h | 231 int wqe_shift; member 257 int wqe_shift; member
|
H A D | mthca_cq.c | 541 >> wq->wqe_shift); in mthca_poll_one() 548 wqe_index = wqe >> srq->wqe_shift; in mthca_poll_one() 555 wqe_index = wqe >> wq->wqe_shift; in mthca_poll_one()
|
/freebsd/contrib/ofed/libmlx5/ |
H A D | srq.c | 46 return srq->buf.buf + (n << srq->wqe_shift); in get_wqe() 55 int max = 1 << (srq->wqe_shift - 4); in mlx5_copy_to_recv_srq() 181 srq->wqe_shift = mlx5_ilog2(size); in mlx5_alloc_srq_buf()
|
H A D | mlx5.c | 660 qp_out->sq.stride = 1 << mqp->sq.wqe_shift; in mlx5dv_get_qp() 664 qp_out->rq.stride = 1 << mqp->rq.wqe_shift; in mlx5dv_get_qp() 704 rwq_out->stride = 1 << mrwq->rq.wqe_shift; in mlx5dv_get_rwq() 719 srq_out->stride = 1 << msrq->wqe_shift; in mlx5dv_get_srq()
|
H A D | verbs.c | 924 qp->sq.wqe_shift = mlx5_ilog2(MLX5_SEND_WQE_BB); in mlx5_calc_sq_size() 957 rwq->rq.wqe_shift = mlx5_ilog2(wqe_size); in mlx5_calc_rwq_size() 992 qp->rq.wqe_shift = mlx5_ilog2(wqe_size); in mlx5_calc_rq_size() 999 qp->rq.wqe_shift = 0; in mlx5_calc_rq_size() 1316 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in create_qp() 1320 (qp->sq.wqe_cnt << qp->sq.wqe_shift); in create_qp() 1346 cmd.rq_wqe_shift = qp->rq.wqe_shift; in create_qp() 2114 cmd.drv.rq_wqe_shift = rwq->rq.wqe_shift; in mlx5_create_wq()
|
H A D | qp.c | 63 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe() 68 return rwq->pbuff + (n << rwq->rq.wqe_shift); in get_wq_recv_wqe() 96 int max = 1 << (qp->rq.wqe_shift - 4); in mlx5_copy_to_recv_wqe() 1093 memset(sig, 0, 1 << rwq->rq.wqe_shift); in mlx5_post_wq_recv() 1166 memset(sig, 0, 1 << qp->rq.wqe_shift); in mlx5_post_recv()
|
H A D | mlx5.h | 353 int wqe_shift; member 376 int wqe_shift; member
|
/freebsd/sys/dev/mlx5/mlx5_core/ |
H A D | mlx5_srq.c | 65 MLX5_SET(wq, wq, log_wq_stride, in->wqe_shift + 4); in set_wq() 77 MLX5_SET(srqc, srqc, log_rq_stride, in->wqe_shift); in set_srqc() 92 in->wqe_shift = MLX5_GET(wq, wq, log_wq_stride) - 4; in get_wq() 105 in->wqe_shift = MLX5_GET(srqc, srqc, log_rq_stride); in get_srqc() 134 u32 log_rq_stride = in->wqe_shift; in get_pas_size()
|
/freebsd/sys/dev/mlx4/mlx4_ib/ |
H A D | mlx4_ib_qp.c | 199 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 204 return get_wqe(qp, qp->sq.offset + (n << qp->sq.wqe_shift)); in get_send_wqe() 227 s = roundup(size, 1U << qp->sq.wqe_shift); in stamp_send_wqe() 229 ind = (i >> qp->sq.wqe_shift) + n; in stamp_send_wqe() 233 wqe = buf + (i & ((1 << qp->sq.wqe_shift) - 1)); in stamp_send_wqe() 288 post_nop_wqe(qp, ind, s << qp->sq.wqe_shift); in pad_wraparound() 406 qp->rq.wqe_shift = ilog2(qp->rq.max_gs * sizeof (struct mlx4_wqe_data_seg)); in set_rq_size() 489 qp->sq.wqe_shift = ilog2(64); in set_kernel_sq_size() 491 qp->sq.wqe_shift = order_base_2(s); in set_kernel_sq_size() 494 qp->sq_max_wqes_per_wr = DIV_ROUND_UP(s, 1U << qp->sq.wqe_shift); in set_kernel_sq_size() [all …]
|
H A D | mlx4_ib_srq.c | 44 return mlx4_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 103 srq->msrq.wqe_shift = ilog2(desc_size); in mlx4_ib_create_srq() 285 mlx4_buf_free(dev->dev, msrq->msrq.max << msrq->msrq.wqe_shift, in mlx4_ib_destroy_srq()
|
H A D | mlx4_ib.h | 175 int wqe_shift; member
|
/freebsd/sys/dev/mlx5/mlx5_ib/ |
H A D | mlx5_ib_srq.c | 41 return mlx5_buf_offset(&srq->buf, n << srq->msrq.wqe_shift); in get_wqe() 275 srq->msrq.wqe_shift = ilog2(desc_size); in mlx5_ib_create_srq() 294 in.wqe_shift = srq->msrq.wqe_shift - 4; in mlx5_ib_create_srq()
|
H A D | mlx5_ib_qp.c | 107 return get_wqe(qp, qp->rq.offset + (n << qp->rq.wqe_shift)); in get_recv_wqe() 152 offset = wq->offset + ((wqe_index % wq->wqe_cnt) << wq->wqe_shift); in mlx5_ib_read_user_wqe() 153 wq_end = wq->offset + (wq->wqe_cnt << wq->wqe_shift); in mlx5_ib_read_user_wqe() 173 wqe_length = 1 << wq->wqe_shift; in mlx5_ib_read_user_wqe() 247 qp->rq.wqe_shift = 0; in set_rq_size() 253 qp->rq.wqe_shift = ucmd->rq_wqe_shift; in set_rq_size() 254 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; in set_rq_size() 270 qp->rq.wqe_shift = ilog2(wqe_size); in set_rq_size() 271 qp->rq.max_gs = (1 << qp->rq.wqe_shift) / sizeof(struct mlx5_wqe_data_seg) - qp->wq_sig; in set_rq_size() 409 qp->sq.wqe_shift = ilog2(MLX5_SEND_WQE_BB); in calc_sq_size() [all …]
|
H A D | mlx5_ib.h | 237 int wqe_shift; member 260 u32 wqe_shift; member
|
/freebsd/sys/dev/mlx5/ |
H A D | srq.h | 40 u32 wqe_shift; member
|
H A D | driver.h | 487 int wqe_shift; member
|
/freebsd/sys/dev/mlx4/mlx4_core/ |
H A D | mlx4_srq.c | 190 srq_context->logstride = srq->wqe_shift - 4; in mlx4_srq_alloc()
|
/freebsd/sys/dev/mlx4/ |
H A D | device.h | 767 int wqe_shift; member
|