/freebsd/contrib/ofed/libirdma/ |
H A D | irdma_umain.c | 162 iwvctx->uk_attrs.feature_flags = resp.feature_flags; in irdma_init_context() 163 iwvctx->uk_attrs.hw_rev = resp.hw_rev; in irdma_init_context() 164 iwvctx->uk_attrs.max_hw_wq_frags = resp.max_hw_wq_frags; in irdma_init_context() 165 iwvctx->uk_attrs.max_hw_read_sges = resp.max_hw_read_sges; in irdma_init_context() 166 iwvctx->uk_attrs.max_hw_inline = resp.max_hw_inline; in irdma_init_context() 167 iwvctx->uk_attrs.max_hw_rq_quanta = resp.max_hw_rq_quanta; in irdma_init_context() 168 iwvctx->uk_attrs.max_hw_wq_quanta = resp.max_hw_wq_quanta; in irdma_init_context() 169 iwvctx->uk_attrs.max_hw_sq_chunk = resp.max_hw_sq_chunk; in irdma_init_context() 170 iwvctx->uk_attrs.max_hw_cq_size = resp.max_hw_cq_size; in irdma_init_context() 171 iwvctx->uk_attrs.min_hw_cq_size = resp.min_hw_cq_size; in irdma_init_context() [all …]
|
H A D | irdma_uverbs.c | 413 struct irdma_uk_attrs *uk_attrs; in ucreate_cq() local 422 uk_attrs = &iwvctx->uk_attrs; in ucreate_cq() 423 hw_rev = uk_attrs->hw_rev; in ucreate_cq() 434 if (attr_ex->cqe < uk_attrs->min_hw_cq_size || attr_ex->cqe > uk_attrs->max_hw_cq_size - 1) { in ucreate_cq() 459 if (!(uk_attrs->feature_flags & IRDMA_FEATURE_CQ_RESIZE)) in ucreate_cq() 483 if (uk_attrs->feature_flags & IRDMA_FEATURE_CQ_RESIZE) { in ucreate_cq() 616 struct irdma_uk_attrs *uk_attrs; in irdma_udestroy_cq() local 623 uk_attrs = &iwvctx->uk_attrs; in irdma_udestroy_cq() 637 if (uk_attrs->feature_flags & IRDMA_FEATURE_CQ_RESIZE) { in irdma_udestroy_cq() 1362 struct irdma_uk_attrs *uk_attrs; in irdma_ucreate_qp() local [all …]
|
H A D | irdma_uk.c | 240 avail_quanta = qp->uk_attrs->max_hw_sq_chunk - in irdma_qp_get_next_send_wqe() 242 qp->uk_attrs->max_hw_sq_chunk); in irdma_qp_get_next_send_wqe() 273 if (qp->uk_attrs->hw_rev == IRDMA_GEN_1 && wqe_quanta == 1 && in irdma_qp_get_next_send_wqe() 380 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_rdma_write() 384 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_write() 469 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && in irdma_uk_rdma_read() 473 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_read() 561 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_send() 565 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_send() 1033 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(info->num_sges & 0x01) && in irdma_uk_post_receive() [all …]
|
H A D | irdma_user.h | 359 struct irdma_uk_attrs *uk_attrs; member 420 struct irdma_uk_attrs *uk_attrs; member 463 void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge, 465 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, u32 *sqdepth); 466 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, u32 *rqdepth);
|
H A D | irdma.h | 66 struct irdma_uk_attrs uk_attrs; member
|
H A D | irdma_umain.h | 86 struct irdma_uk_attrs uk_attrs; member
|
/freebsd/sys/dev/irdma/ |
H A D | irdma_uk.c | 240 avail_quanta = qp->uk_attrs->max_hw_sq_chunk - in irdma_qp_get_next_send_wqe() 242 qp->uk_attrs->max_hw_sq_chunk); in irdma_qp_get_next_send_wqe() 273 if (qp->uk_attrs->hw_rev == IRDMA_GEN_1 && wqe_quanta == 1 && in irdma_qp_get_next_send_wqe() 380 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_rdma_write() 384 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_write() 469 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && in irdma_uk_rdma_read() 473 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_rdma_read() 561 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(frag_cnt & 0x01) && in irdma_uk_send() 565 if (qp->uk_attrs->hw_rev == IRDMA_GEN_2) in irdma_uk_send() 949 if (qp->uk_attrs->hw_rev >= IRDMA_GEN_2 && !(info->num_sges & 0x01) && in irdma_uk_post_receive() [all …]
|
H A D | irdma_kcompat.c | 180 struct irdma_uk_attrs *uk_attrs = &iwdev->rf->sc_dev.hw_attrs.uk_attrs; in irdma_alloc_ucontext() local 200 if (uk_attrs->hw_rev != IRDMA_GEN_1) in irdma_alloc_ucontext() 214 uresp.feature_flags = uk_attrs->feature_flags; in irdma_alloc_ucontext() 215 uresp.max_hw_wq_frags = uk_attrs->max_hw_wq_frags; in irdma_alloc_ucontext() 216 uresp.max_hw_read_sges = uk_attrs->max_hw_read_sges; in irdma_alloc_ucontext() 217 uresp.max_hw_inline = uk_attrs->max_hw_inline; in irdma_alloc_ucontext() 218 uresp.max_hw_rq_quanta = uk_attrs->max_hw_rq_quanta; in irdma_alloc_ucontext() 219 uresp.max_hw_wq_quanta = uk_attrs->max_hw_wq_quanta; in irdma_alloc_ucontext() 220 uresp.max_hw_sq_chunk = uk_attrs->max_hw_sq_chunk; in irdma_alloc_ucontext() 221 uresp.max_hw_cq_size = uk_attrs->max_hw_cq_size; in irdma_alloc_ucontext() [all …]
|
H A D | icrdma_hw.c | 234 dev->hw_attrs.uk_attrs.max_hw_wq_frags = ICRDMA_MAX_WQ_FRAGMENT_COUNT; in icrdma_init_hw() 235 dev->hw_attrs.uk_attrs.max_hw_read_sges = ICRDMA_MAX_SGE_RD; in icrdma_init_hw() 236 dev->hw_attrs.uk_attrs.min_hw_wq_size = ICRDMA_MIN_WQ_SIZE; in icrdma_init_hw() 237 dev->hw_attrs.uk_attrs.max_hw_sq_chunk = IRDMA_MAX_QUANTA_PER_WR; in icrdma_init_hw() 241 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_RELAX_RQ_ORDER; in icrdma_init_hw() 242 dev->hw_attrs.uk_attrs.feature_flags |= IRDMA_FEATURE_RTS_AE | in icrdma_init_hw()
|
H A D | irdma_user.h | 460 struct irdma_uk_attrs *uk_attrs; member 521 struct irdma_uk_attrs *uk_attrs; member 564 void irdma_get_wqe_shift(struct irdma_uk_attrs *uk_attrs, u32 sge, 566 int irdma_get_sqdepth(struct irdma_uk_attrs *uk_attrs, u32 sq_size, u8 shift, u32 *sqdepth); 567 int irdma_get_rqdepth(struct irdma_uk_attrs *uk_attrs, u32 rq_size, u8 shift, u32 *rqdepth);
|
H A D | irdma_puda.c | 136 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) { in irdma_puda_post_recvbuf() 276 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_puda_poll_info() 318 if (cq->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_puda_poll_info() 344 if (cq->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) { in irdma_puda_poll_info() 490 if (qp->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_send() 576 if (rsrc->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_send_buf() 618 if (qp->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_puda_qp_setctx() 715 ukqp->uk_attrs = &qp->dev->hw_attrs.uk_attrs; in irdma_puda_qp_create() 1160 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_puda_create_rsrc() 1228 if (qp->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_ilq_putback_rcvbuf() [all …]
|
H A D | irdma_verbs.c | 76 props->max_sge_rd = hw_attrs->uk_attrs.max_hw_read_sges; in irdma_query_device() 83 if (hw_attrs->uk_attrs.hw_rev == IRDMA_GEN_2) { in irdma_query_device() 90 if (hw_attrs->uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_query_device() 285 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2) { in irdma_compute_push_wqe_offset() 320 WARN_ON_ONCE(iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev < IRDMA_GEN_2); in irdma_setup_push_mmap_entries() 444 iwdev->rf->sc_dev.hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) in irdma_setup_umode_qp() 622 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_iw_fill_and_set_qpctx_info() 642 struct irdma_uk_attrs *uk_attrs = &dev->hw_attrs.uk_attrs; in irdma_validate_qp_attrs() local 647 if (init_attr->cap.max_inline_data > uk_attrs->max_hw_inline || in irdma_validate_qp_attrs() 648 init_attr->cap.max_send_sge > uk_attrs->max_hw_wq_frags || in irdma_validate_qp_attrs() [all …]
|
H A D | irdma_ctrl.c | 134 if (vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_set_qos_info() 136 if (vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2) in irdma_set_qos_info() 439 info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags || in irdma_sc_qp_init() 441 info->pd->dev->hw_attrs.uk_attrs.max_hw_wq_frags) in irdma_sc_qp_init() 473 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) in irdma_sc_qp_init() 978 if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_sc_qp_setctx() 1007 if (dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_sc_qp_setctx() 1050 if (dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) { in irdma_sc_qp_setctx() 1498 if (qp->qp_uk.uk_attrs->hw_rev == IRDMA_GEN_1) { in irdma_sc_send_lsmm() 1521 if (qp->dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) in irdma_sc_send_lsmm() [all …]
|
H A D | irdma.h | 204 struct irdma_uk_attrs uk_attrs; member
|
H A D | fbsd_kcompat.h | 61 ((props)->max_sge = (rf)->sc_dev.hw_attrs.uk_attrs.max_hw_wq_frags)
|
H A D | irdma_utils.c | 1413 if (buf->vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_ieq_update_tcpip_info() 1491 if (buf->vsi->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_puda_get_tcpip_info() 1534 if (sc_vsi->dev->hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) in irdma_hw_stats_timeout()
|
H A D | irdma_cm.c | 3022 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev >= IRDMA_GEN_2) { in irdma_receive_ilq() 3575 if (cm_node->dev->hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2) in irdma_accept() 3616 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) { in irdma_accept() 3765 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_2) in irdma_connect() 4104 if (dev->hw_attrs.uk_attrs.feature_flags & IRDMA_FEATURE_RTS_AE) { in irdma_cm_event_connected()
|
H A D | icrdma.c | 514 rf->sc_dev.hw_attrs.uk_attrs.hw_rev = IRDMA_GEN_2; in irdma_fill_device_info()
|
H A D | irdma_hw.c | 1767 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_deinit_hw() 1988 if (iwdev->rf->sc_dev.hw_attrs.uk_attrs.hw_rev == IRDMA_GEN_1) in irdma_rt_init_hw() 2149 rf->max_cqe = rf->sc_dev.hw_attrs.uk_attrs.max_hw_cq_size; in irdma_initialize_hw_rsrc()
|