/freebsd/sys/dev/enic/ |
H A D | vnic_rq.c | 10 void vnic_rq_init_start(struct vnic_rq *rq, unsigned int cq_index, in vnic_rq_init_start() argument 16 unsigned int count = rq->ring.desc_count; in vnic_rq_init_start() 18 paddr = (u64)rq->ring.base_addr | VNIC_PADDR_TARGET; in vnic_rq_init_start() 19 ENIC_BUS_WRITE_8(rq->ctrl, RX_RING_BASE, paddr); in vnic_rq_init_start() 20 ENIC_BUS_WRITE_4(rq->ctrl, RX_RING_SIZE, count); in vnic_rq_init_start() 21 ENIC_BUS_WRITE_4(rq->ctrl, RX_CQ_INDEX, cq_index); in vnic_rq_init_start() 22 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_INTR_ENABLE, error_interrupt_enable); in vnic_rq_init_start() 23 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_INTR_OFFSET, error_interrupt_offset); in vnic_rq_init_start() 24 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_STATUS, 0); in vnic_rq_init_start() 25 ENIC_BUS_WRITE_4(rq->ctrl, RX_FETCH_INDEX, fetch_index); in vnic_rq_init_start() [all …]
|
H A D | vnic_rq.h | 81 static inline unsigned int vnic_rq_desc_avail(struct vnic_rq *rq) in vnic_rq_desc_avail() argument 84 return rq->ring.desc_avail; in vnic_rq_desc_avail() 87 static inline unsigned int vnic_rq_desc_used(struct vnic_rq *rq) in vnic_rq_desc_used() argument 90 return rq->ring.desc_count - rq->ring.desc_avail - 1; in vnic_rq_desc_used() 98 static inline int vnic_rq_fill(struct vnic_rq *rq, in vnic_rq_fill() argument 99 int (*buf_fill)(struct vnic_rq *rq)) in vnic_rq_fill() argument 103 while (vnic_rq_desc_avail(rq) > 0) { in vnic_rq_fill() 105 err = (*buf_fill)(rq); in vnic_rq_fill() 113 static inline int vnic_rq_fill_count(struct vnic_rq *rq, in vnic_rq_fill_count() argument 114 int (*buf_fill)(struct vnic_rq *rq), unsigned int count) in vnic_rq_fill_count() argument [all …]
|
H A D | enic_txrx.c | 260 struct vnic_rq *rq; in enic_isc_rxd_refill() local 276 rq = &softc->enic.rq[iru->iru_qsidx]; in enic_isc_rxd_refill() 277 rqd = rq->ring.descs; in enic_isc_rxd_refill() 282 if (idx == rq->ring.desc_count) in enic_isc_rxd_refill() 290 rq->in_use = 1; in enic_isc_rxd_refill() 292 if (rq->need_initial_post) { in enic_isc_rxd_refill() 293 ENIC_BUS_WRITE_4(rq->ctrl, RX_FETCH_INDEX, 0); in enic_isc_rxd_refill() 296 enic_initial_post_rx(&softc->enic, rq); in enic_isc_rxd_refill() 304 struct vnic_rq *rq; in enic_isc_rxd_flush() local 307 rq = &softc->enic.rq[rxqid]; in enic_isc_rxd_flush() [all …]
|
/freebsd/sys/dev/mlx5/mlx5_en/ |
H A D | mlx5_en_rx.c | 34 mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, in mlx5e_alloc_rx_wqe() argument 44 if (rq->mbuf[ix].mbuf != NULL) in mlx5e_alloc_rx_wqe() 55 for (i = 1; i < rq->nsegs; i++) { in mlx5e_alloc_rx_wqe() 56 if (mb_head->m_pkthdr.len >= rq->wqe_sz) in mlx5e_alloc_rx_wqe() 73 err = mlx5_accel_ipsec_rx_tag_add(rq->ifp, &rq->mbuf[ix]); in mlx5e_alloc_rx_wqe() 76 err = -bus_dmamap_load_mbuf_sg(rq->dma_tag, rq->mbuf[ix].dma_map, in mlx5e_alloc_rx_wqe() 81 bus_dmamap_unload(rq->dma_tag, rq->mbuf[ix].dma_map); in mlx5e_alloc_rx_wqe() 92 for (; i < rq->nsegs; i++) { in mlx5e_alloc_rx_wqe() 97 rq->mbuf[ix].mbuf = mb; in mlx5e_alloc_rx_wqe() 98 rq->mbuf[ix].data = mb->m_data; in mlx5e_alloc_rx_wqe() [all …]
|
H A D | mlx5_en_dim.c | 64 struct mlx5e_rq *rq = container_of(dim, struct mlx5e_rq, dim); in mlx5e_dim_work() local 65 struct mlx5e_channel *c = container_of(rq, struct mlx5e_channel, rq); in mlx5e_dim_work() 71 mtx_lock(&rq->mtx); in mlx5e_dim_work() 75 mtx_unlock(&rq->mtx); in mlx5e_dim_work() 91 mlx5_core_modify_cq_moderation(c->priv->mdev, &rq->cq.mcq, in mlx5e_dim_work()
|
H A D | mlx5_en_main.c | 49 struct mlx5e_rq_param rq; member 848 struct mlx5e_rq *rq = &pch->rq; in mlx5e_update_stats_locked() local 849 struct mlx5e_rq_stats *rq_stats = &pch->rq.stats; in mlx5e_update_stats_locked() 852 rq_stats->sw_lro_queued = rq->lro.lro_queued; in mlx5e_update_stats_locked() 853 rq_stats->sw_lro_flushed = rq->lro.lro_flushed; in mlx5e_update_stats_locked() 1205 struct mlx5e_rq *rq) in mlx5e_create_rq() argument 1234 &rq->dma_tag))) in mlx5e_create_rq() 1237 err = mlx5_wq_ll_create(mdev, ¶m->wq, rqc_wq, &rq->wq, in mlx5e_create_rq() 1238 &rq->wq_ctrl); in mlx5e_create_rq() 1242 rq->wq.db = &rq->wq.db[MLX5_RCV_DBR]; in mlx5e_create_rq() [all …]
|
/freebsd/sys/dev/oce/ |
H A D | oce_queue.c | 57 static int oce_rq_create(struct oce_rq *rq, uint32_t if_id, struct oce_eq *eq); 58 static void oce_rq_free(struct oce_rq *rq); 59 static void oce_rq_del(struct oce_rq *rq); 90 struct oce_rq *rq; in oce_queue_init_all() local 102 for_all_rq_queues(sc, rq, i) { in oce_queue_init_all() 103 sc->rq[i] = oce_rq_init(sc, sc->rx_ring_size, sc->rq_frag_size, in oce_queue_init_all() 106 if (!sc->rq[i]) in oce_queue_init_all() 139 for_all_rq_queues(sc, rq, i) { in oce_queue_init_all() 140 rc = oce_rq_create(rq, sc->if_id, in oce_queue_init_all() 144 rq->queue_index = i; in oce_queue_init_all() [all …]
|
H A D | oce_if.c | 165 static void oce_rx(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe); 166 static void oce_check_rx_bufs(POCE_SOFTC sc, uint32_t num_cqes, struct oce_rq *rq); 169 static void oce_rx_lro(struct oce_rq *rq, struct nic_hwlro_singleton_cqe *cqe, struct nic_hwlro_cqe… 170 static void oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf … 1511 oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf **m) in oce_rx_mbuf_chain() argument 1513 POCE_SOFTC sc = (POCE_SOFTC) rq->parent; in oce_rx_mbuf_chain() 1520 if (rq->ring->cidx == rq->ring->pidx) { in oce_rx_mbuf_chain() 1525 pd = &rq->pckts[rq->ring->cidx]; in oce_rx_mbuf_chain() 1527 bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE); in oce_rx_mbuf_chain() 1528 bus_dmamap_unload(rq->tag, pd->map); in oce_rx_mbuf_chain() [all …]
|
/freebsd/sys/kern/ |
H A D | kern_switch.c | 260 runq_init(struct runq *rq) in runq_init() argument 264 bzero(rq, sizeof *rq); in runq_init() 266 TAILQ_INIT(&rq->rq_queues[i]); in runq_init() 274 runq_clrbit(struct runq *rq, int pri) in runq_clrbit() argument 278 rqb = &rq->rq_status; in runq_clrbit() 291 runq_findbit(struct runq *rq) in runq_findbit() argument 297 rqb = &rq->rq_status; in runq_findbit() 310 runq_findbit_from(struct runq *rq, u_char pri) in runq_findbit_from() argument 320 rqb = &rq->rq_status; in runq_findbit_from() 346 runq_setbit(struct runq *rq, int pri) in runq_setbit() argument [all …]
|
/freebsd/sys/dev/ice/ |
H A D | ice_controlq.c | 45 (qinfo)->rq.head = prefix##_ARQH; \ 46 (qinfo)->rq.tail = prefix##_ARQT; \ 47 (qinfo)->rq.len = prefix##_ARQLEN; \ 48 (qinfo)->rq.bah = prefix##_ARQBAH; \ 49 (qinfo)->rq.bal = prefix##_ARQBAL; \ 50 (qinfo)->rq.len_mask = prefix##_ARQLEN_ARQLEN_M; \ 51 (qinfo)->rq.len_ena_mask = prefix##_ARQLEN_ARQENABLE_M; \ 52 (qinfo)->rq.len_crit_mask = prefix##_ARQLEN_ARQCRIT_M; \ 53 (qinfo)->rq.head_mask = prefix##_ARQH_ARQH_M; \ 144 cq->rq in ice_free_cq_ring() [all...] |
/freebsd/sys/contrib/openzfs/include/os/linux/kernel/linux/ |
H A D | blkdev_compat.h | 550 io_data_dir(struct bio *bio, struct request *rq) in io_data_dir() argument 552 if (rq != NULL) { in io_data_dir() 553 if (op_is_write(req_op(rq))) { in io_data_dir() 563 io_is_flush(struct bio *bio, struct request *rq) in io_is_flush() argument 565 if (rq != NULL) in io_is_flush() 566 return (req_op(rq) == REQ_OP_FLUSH); in io_is_flush() 571 io_is_discard(struct bio *bio, struct request *rq) in io_is_discard() argument 573 if (rq != NULL) in io_is_discard() 574 return (req_op(rq) == REQ_OP_DISCARD); in io_is_discard() 579 io_is_secure_erase(struct bio *bio, struct request *rq) in io_is_secure_erase() argument [all …]
|
/freebsd/contrib/ofed/libcxgb4/ |
H A D | verbs.c | 342 qhp->wq.rq.msn = 1; in create_qp_v0() 343 qhp->wq.rq.qid = resp.rqid; in create_qp_v0() 344 qhp->wq.rq.size = resp.rq_size; in create_qp_v0() 345 qhp->wq.rq.memsize = resp.rq_memsize; in create_qp_v0() 366 qhp->wq.rq.udb = dbva; in create_qp_v0() 367 qhp->wq.rq.queue = mmap(NULL, qhp->wq.rq.memsize, in create_qp_v0() 370 if (qhp->wq.rq.queue == MAP_FAILED) in create_qp_v0() 377 qhp->wq.rq.sw_rq = calloc(qhp->wq.rq.size, sizeof (uint64_t)); in create_qp_v0() 378 if (!qhp->wq.rq.sw_rq) in create_qp_v0() 386 qhp->wq.rq.udb, qhp->wq.rq.queue, in create_qp_v0() [all …]
|
H A D | t4.h | 355 struct t4_rq rq; member 365 return wq->rq.in_use; in t4_rqes_posted() 370 return wq->rq.in_use == 0; in t4_rq_empty() 375 return wq->rq.in_use == (wq->rq.size - 1); in t4_rq_full() 380 return wq->rq.size - 1 - wq->rq.in_use; in t4_rq_avail() 385 wq->rq.in_use++; in t4_rq_produce() 386 if (++wq->rq.pidx == wq->rq.size) in t4_rq_produce() 387 wq->rq.pidx = 0; in t4_rq_produce() 388 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce() 389 if (wq->rq.wq_pidx >= wq->rq.size * T4_RQ_NUM_SLOTS) in t4_rq_produce() [all …]
|
H A D | dev.c | 300 qhp->wq.rq.qid, in dump_qp() 301 qhp->wq.rq.queue, in dump_qp() 302 qhp->wq.rq.sw_rq, in dump_qp() 303 qhp->wq.rq.cidx, in dump_qp() 304 qhp->wq.rq.pidx, in dump_qp() 305 qhp->wq.rq.in_use, in dump_qp() 306 qhp->wq.rq.size); in dump_qp() 343 cidx = qhp->wq.rq.cidx; in dump_qp() 344 pidx = qhp->wq.rq.pidx; in dump_qp() 348 swrqe = &qhp->wq.rq.sw_rq[cidx]; in dump_qp() [all …]
|
/freebsd/sys/dev/ocs_fc/ |
H A D | ocs_hw_queues.c | 71 hw_rq_t *rq = NULL; in ocs_hw_init_queues() local 236 rq = hw_new_rq(cq, len, hw->ulp_start + qt->ulp); in ocs_hw_init_queues() 237 if (rq == NULL) { in ocs_hw_init_queues() 241 rq->filter_mask = qt->filter_mask; in ocs_hw_init_queues() 296 rq = hw_new_rq(cq, mrq.rq_cfg[i].len, mrq.rq_cfg[i].ulp); in ocs_hw_init_queues() 297 if (rq == NULL) { in ocs_hw_init_queues() 301 rq->filter_mask = mrq.rq_cfg[i].filter_mask; in ocs_hw_init_queues() 593 hw_rq_t *rq = ocs_malloc(hw->os, sizeof(*rq), OCS_M_ZERO | OCS_M_NOWAIT); in hw_new_rq() local 598 if (rq != NULL) { in hw_new_rq() 599 rq->instance = hw->hw_rq_count++; in hw_new_rq() [all …]
|
/freebsd/sys/dev/cxgbe/iw_cxgbe/ |
H A D | t4.h | 362 struct t4_rq rq; member 369 return wq->rq.in_use; in t4_rqes_posted() 374 return wq->rq.in_use == 0; in t4_rq_empty() 379 return wq->rq.in_use == (wq->rq.size - 1); in t4_rq_full() 384 return wq->rq.size - 1 - wq->rq.in_use; in t4_rq_avail() 389 wq->rq.in_use++; in t4_rq_produce() 390 if (++wq->rq.pidx == wq->rq.size) in t4_rq_produce() 391 wq->rq.pidx = 0; in t4_rq_produce() 392 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce() 393 if (wq->rq.wq_pidx >= wq->rq.size * T4_RQ_NUM_SLOTS) in t4_rq_produce() [all …]
|
H A D | qp.c | 112 wq->rq.memsize, wq->rq.queue, in destroy_qp() 113 dma_unmap_addr(&wq->rq, mapping)); in destroy_qp() 117 c4iw_rqtpool_free(rdev, wq->rq.rqt_hwaddr, wq->rq.rqt_size); in destroy_qp() 118 kfree(wq->rq.sw_rq); in destroy_qp() 120 c4iw_put_qpid(rdev, wq->rq.qid, uctx); in destroy_qp() 148 wq->rq.qid = c4iw_get_qpid(rdev, uctx); in create_qp() 149 if (!wq->rq.qid) { in create_qp() 162 wq->rq.sw_rq = kzalloc(wq->rq.size * sizeof *wq->rq.sw_rq, in create_qp() 164 if (!wq->rq.sw_rq) { in create_qp() 173 wq->rq.rqt_size = roundup_pow_of_two(max_t(u16, wq->rq.size, 16)); in create_qp() [all …]
|
/freebsd/sys/dev/bnxt/bnxt_re/ |
H A D | qplib_fp.c | 74 if (!qp->rq.flushed) { in __bnxt_qplib_add_flush_qp() 79 qp->rq.flushed = true; in __bnxt_qplib_add_flush_qp() 120 if (qp->rq.flushed) { in __bnxt_qplib_del_flush_qp() 121 qp->rq.flushed = false; in __bnxt_qplib_del_flush_qp() 137 qp->rq.hwq.prod = 0; in bnxt_qplib_clean_qp() 138 qp->rq.hwq.cons = 0; in bnxt_qplib_clean_qp() 139 qp->rq.swq_start = 0; in bnxt_qplib_clean_qp() 140 qp->rq.swq_last = 0; in bnxt_qplib_clean_qp() 229 qp->rq.max_wqe); in bnxt_qplib_alloc_hdr_buf() 912 struct bnxt_qplib_q *rq = &qp->rq; in bnxt_qplib_create_qp1() local [all …]
|
/freebsd/contrib/ofed/libmlx4/ |
H A D | qp.c | 61 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe() 88 qp->rq.head = 0; in mlx4_init_qp_indices() 89 qp->rq.tail = 0; in mlx4_init_qp_indices() 522 pthread_spin_lock(&qp->rq.lock); in mlx4_post_recv() 526 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx4_post_recv() 529 if (wq_overflow(&qp->rq, nreq, to_mcq(ibqp->recv_cq))) { in mlx4_post_recv() 535 if (wr->num_sge > qp->rq.max_gs) { in mlx4_post_recv() 546 if (i < qp->rq.max_gs) { in mlx4_post_recv() 552 qp->rq.wrid[ind] = wr->wr_id; in mlx4_post_recv() 554 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx4_post_recv() [all …]
|
/freebsd/sys/arm/include/ |
H A D | sysreg.h | 271 #define CP15_CNTPCT(rq, rr) p15, 0, rq, rr, c14 /* Physical Count Register */ argument 272 #define CP15_CNTVCT(rq, rr) p15, 1, rq, rr, c14 /* Virtual Count Register */ argument 273 #define CP15_CNTP_CVAL(rq, rr) p15, 2, rq, rr, c14 /* PL1 Physical Timer Compare Value Register */ argument 274 #define CP15_CNTV_CVAL(rq, rr) p15, 3, rq, rr, c14 /* Virtual Timer Compare Value Register */ argument 275 #define CP15_CNTVOFF(rq, rr) p15, 4, rq, rr, c14 /* Virtual Offset Register */ argument 276 #define CP15_CNTHP_CVAL(rq, rr) p15, 6, rq, rr, c14 /* PL2 Physical Timer Compare Value Register */ argument 278 #define CP15_VTTBR(rq, rr) p15, 6, rq, rr, c2 /* Virtualization Translation Table Base Register */ argument 279 #define CP15_HTTBR(rq, rr) p15, 4, rq, rr, c2 /* Hyp Translation Table Base Register */ argument 280 #define CP15_TTBR0_2(rq, rr) p15, 0, rq, rr, c2 /* Translation Table Base Register 0 */ argument 281 #define CP15_TTBR1_2(rq, rr) p15, 1, rq, rr, c2 /* Translation Table Base Register 1 */ argument [all …]
|
/freebsd/sys/dev/mthca/ |
H A D | mthca_qp.c | 211 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); in get_recv_wqe() 213 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + in get_recv_wqe() 214 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); in get_recv_wqe() 500 qp_attr->cap.max_recv_wr = qp->rq.max; in mthca_query_qp() 502 qp_attr->cap.max_recv_sge = qp->rq.max_gs; in mthca_query_qp() 603 if (qp->rq.max) in __mthca_modify_qp() 604 qp_context->rq_size_stride = ilog2(qp->rq.max) << 3; in __mthca_modify_qp() 605 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mthca_modify_qp() 765 qp_context->rcv_db_index = cpu_to_be32(qp->rq.db_index); in __mthca_modify_qp() 831 mthca_wq_reset(&qp->rq); in __mthca_modify_qp() [all …]
|
/freebsd/sys/contrib/openzfs/module/os/linux/zfs/ |
H A D | zvol_os.c | 50 struct request *rq, boolean_t force_sync); 95 zvol_end_io(struct bio *bio, struct request *rq, int error) in zvol_end_io() argument 101 blk_mq_end_request(rq, errno_to_bi_status(error)); in zvol_end_io() 129 struct request *rq; member 133 struct request *rq; member 165 struct request *rq = bd->rq; in zvol_mq_queue_rq() local 166 zvol_state_t *zv = rq->q->queuedata; in zvol_mq_queue_rq() 169 blk_mq_start_request(rq); in zvol_mq_queue_rq() 171 if (blk_rq_is_passthrough(rq)) { in zvol_mq_queue_rq() 173 blk_mq_end_request(rq, BLK_STS_IOERR); in zvol_mq_queue_rq() [all …]
|
/freebsd/sys/dev/cfi/ |
H A D | cfi_dev.c | 265 struct cfiocqry *rq; in cfi_devioctl() local 279 rq = (struct cfiocqry *)data; in cfi_devioctl() 280 if (rq->offset >= sc->sc_size / sc->sc_width) in cfi_devioctl() 282 if (rq->offset > ULONG_MAX - rq->count || in cfi_devioctl() 283 rq->offset + rq->count > sc->sc_size / sc->sc_width) in cfi_devioctl() 286 while (!error && rq->count--) { in cfi_devioctl() 287 val = cfi_read_qry(sc, rq->offset++); in cfi_devioctl() 288 error = copyout(&val, rq->buffer++, 1); in cfi_devioctl()
|
/freebsd/contrib/ofed/libmlx5/ |
H A D | qp.c | 63 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe() 68 return rwq->pbuff + (n << rwq->rq.wqe_shift); in get_wq_recv_wqe() 96 int max = 1 << (qp->rq.wqe_shift - 4); in mlx5_copy_to_recv_wqe() 160 rwq->rq.head = 0; in mlx5_init_rwq_indices() 161 rwq->rq.tail = 0; in mlx5_init_rwq_indices() 168 qp->rq.head = 0; in mlx5_init_qp_indices() 169 qp->rq.tail = 0; in mlx5_init_qp_indices() 1072 mlx5_spin_lock(&rwq->rq.lock); in mlx5_post_wq_recv() 1074 ind = rwq->rq.head & (rwq->rq.wqe_cnt - 1); in mlx5_post_wq_recv() 1077 if (unlikely(mlx5_wq_overflow(&rwq->rq, nreq, in mlx5_post_wq_recv() [all …]
|
/freebsd/sys/dev/qlnx/qlnxr/ |
H A D | qlnxr_cm.c | 125 qp->rqe_wr_id[qp->rq.gsi_cons].rc = in qlnxr_ll2_complete_rx_packet() 127 qp->rqe_wr_id[qp->rq.gsi_cons].vlan_id = data->vlan; in qlnxr_ll2_complete_rx_packet() 129 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qlnxr_ll2_complete_rx_packet() 131 *((u32 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[0]) = in qlnxr_ll2_complete_rx_packet() 133 *((u16 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[4]) = in qlnxr_ll2_complete_rx_packet() 136 qlnxr_inc_sw_gsi_cons(&qp->rq); in qlnxr_ll2_complete_rx_packet() 411 qp->rq.max_wr = attrs->cap.max_recv_wr; in qlnxr_create_gsi_qp() 414 qp->rqe_wr_id = kzalloc(qp->rq.max_wr * sizeof(*qp->rqe_wr_id), in qlnxr_create_gsi_qp() 780 memset(&qp->rqe_wr_id[qp->rq.prod], 0, in qlnxr_gsi_post_recv() 781 sizeof(qp->rqe_wr_id[qp->rq.prod])); in qlnxr_gsi_post_recv() [all …]
|