Home
last modified time | relevance | path

Searched refs:rq (Results 1 – 25 of 155) sorted by relevance

1234567

/freebsd/sys/dev/enic/
H A Dvnic_rq.c10 void vnic_rq_init_start(struct vnic_rq *rq, unsigned int cq_index, in vnic_rq_init_start() argument
16 unsigned int count = rq->ring.desc_count; in vnic_rq_init_start()
18 paddr = (u64)rq->ring.base_addr | VNIC_PADDR_TARGET; in vnic_rq_init_start()
19 ENIC_BUS_WRITE_8(rq->ctrl, RX_RING_BASE, paddr); in vnic_rq_init_start()
20 ENIC_BUS_WRITE_4(rq->ctrl, RX_RING_SIZE, count); in vnic_rq_init_start()
21 ENIC_BUS_WRITE_4(rq->ctrl, RX_CQ_INDEX, cq_index); in vnic_rq_init_start()
22 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_INTR_ENABLE, error_interrupt_enable); in vnic_rq_init_start()
23 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_INTR_OFFSET, error_interrupt_offset); in vnic_rq_init_start()
24 ENIC_BUS_WRITE_4(rq->ctrl, RX_ERROR_STATUS, 0); in vnic_rq_init_start()
25 ENIC_BUS_WRITE_4(rq->ctrl, RX_FETCH_INDEX, fetch_index); in vnic_rq_init_start()
[all …]
H A Dvnic_rq.h81 static inline unsigned int vnic_rq_desc_avail(struct vnic_rq *rq) in vnic_rq_desc_avail() argument
84 return rq->ring.desc_avail; in vnic_rq_desc_avail()
87 static inline unsigned int vnic_rq_desc_used(struct vnic_rq *rq) in vnic_rq_desc_used() argument
90 return rq->ring.desc_count - rq->ring.desc_avail - 1; in vnic_rq_desc_used()
98 static inline int vnic_rq_fill(struct vnic_rq *rq, in vnic_rq_fill() argument
99 int (*buf_fill)(struct vnic_rq *rq)) in vnic_rq_fill() argument
103 while (vnic_rq_desc_avail(rq) > 0) { in vnic_rq_fill()
105 err = (*buf_fill)(rq); in vnic_rq_fill()
113 static inline int vnic_rq_fill_count(struct vnic_rq *rq, in vnic_rq_fill_count() argument
114 int (*buf_fill)(struct vnic_rq *rq), unsigned int count) in vnic_rq_fill_count() argument
[all …]
H A Denic_txrx.c260 struct vnic_rq *rq; in enic_isc_rxd_refill() local
276 rq = &softc->enic.rq[iru->iru_qsidx]; in enic_isc_rxd_refill()
277 rqd = rq->ring.descs; in enic_isc_rxd_refill()
282 if (idx == rq->ring.desc_count) in enic_isc_rxd_refill()
290 rq->in_use = 1; in enic_isc_rxd_refill()
292 if (rq->need_initial_post) { in enic_isc_rxd_refill()
293 ENIC_BUS_WRITE_4(rq->ctrl, RX_FETCH_INDEX, 0); in enic_isc_rxd_refill()
296 enic_initial_post_rx(&softc->enic, rq); in enic_isc_rxd_refill()
304 struct vnic_rq *rq; in enic_isc_rxd_flush() local
307 rq = &softc->enic.rq[rxqid]; in enic_isc_rxd_flush()
[all …]
/freebsd/sys/dev/mlx5/mlx5_en/
H A Dmlx5_en_rx.c34 mlx5e_alloc_rx_wqe(struct mlx5e_rq *rq, in mlx5e_alloc_rx_wqe() argument
44 if (rq->mbuf[ix].mbuf != NULL) in mlx5e_alloc_rx_wqe()
55 for (i = 1; i < rq->nsegs; i++) { in mlx5e_alloc_rx_wqe()
56 if (mb_head->m_pkthdr.len >= rq->wqe_sz) in mlx5e_alloc_rx_wqe()
73 err = mlx5_accel_ipsec_rx_tag_add(rq->ifp, &rq->mbuf[ix]); in mlx5e_alloc_rx_wqe()
76 err = -bus_dmamap_load_mbuf_sg(rq->dma_tag, rq->mbuf[ix].dma_map, in mlx5e_alloc_rx_wqe()
81 bus_dmamap_unload(rq->dma_tag, rq->mbuf[ix].dma_map); in mlx5e_alloc_rx_wqe()
92 for (; i < rq->nsegs; i++) { in mlx5e_alloc_rx_wqe()
97 rq->mbuf[ix].mbuf = mb; in mlx5e_alloc_rx_wqe()
98 rq->mbuf[ix].data = mb->m_data; in mlx5e_alloc_rx_wqe()
[all …]
H A Dmlx5_en_dim.c64 struct mlx5e_rq *rq = container_of(dim, struct mlx5e_rq, dim); in mlx5e_dim_work() local
65 struct mlx5e_channel *c = container_of(rq, struct mlx5e_channel, rq); in mlx5e_dim_work()
71 mtx_lock(&rq->mtx); in mlx5e_dim_work()
75 mtx_unlock(&rq->mtx); in mlx5e_dim_work()
91 mlx5_core_modify_cq_moderation(c->priv->mdev, &rq->cq.mcq, in mlx5e_dim_work()
H A Dmlx5_en_main.c49 struct mlx5e_rq_param rq; member
848 struct mlx5e_rq *rq = &pch->rq; in mlx5e_update_stats_locked() local
849 struct mlx5e_rq_stats *rq_stats = &pch->rq.stats; in mlx5e_update_stats_locked()
852 rq_stats->sw_lro_queued = rq->lro.lro_queued; in mlx5e_update_stats_locked()
853 rq_stats->sw_lro_flushed = rq->lro.lro_flushed; in mlx5e_update_stats_locked()
1205 struct mlx5e_rq *rq) in mlx5e_create_rq() argument
1234 &rq->dma_tag))) in mlx5e_create_rq()
1237 err = mlx5_wq_ll_create(mdev, &param->wq, rqc_wq, &rq->wq, in mlx5e_create_rq()
1238 &rq->wq_ctrl); in mlx5e_create_rq()
1242 rq->wq.db = &rq->wq.db[MLX5_RCV_DBR]; in mlx5e_create_rq()
[all …]
/freebsd/sys/dev/oce/
H A Doce_queue.c57 static int oce_rq_create(struct oce_rq *rq, uint32_t if_id, struct oce_eq *eq);
58 static void oce_rq_free(struct oce_rq *rq);
59 static void oce_rq_del(struct oce_rq *rq);
90 struct oce_rq *rq; in oce_queue_init_all() local
102 for_all_rq_queues(sc, rq, i) { in oce_queue_init_all()
103 sc->rq[i] = oce_rq_init(sc, sc->rx_ring_size, sc->rq_frag_size, in oce_queue_init_all()
106 if (!sc->rq[i]) in oce_queue_init_all()
139 for_all_rq_queues(sc, rq, i) { in oce_queue_init_all()
140 rc = oce_rq_create(rq, sc->if_id, in oce_queue_init_all()
144 rq->queue_index = i; in oce_queue_init_all()
[all …]
H A Doce_if.c165 static void oce_rx(struct oce_rq *rq, struct oce_nic_rx_cqe *cqe);
166 static void oce_check_rx_bufs(POCE_SOFTC sc, uint32_t num_cqes, struct oce_rq *rq);
169 static void oce_rx_lro(struct oce_rq *rq, struct nic_hwlro_singleton_cqe *cqe, struct nic_hwlro_cqe…
170 static void oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf …
1511 oce_rx_mbuf_chain(struct oce_rq *rq, struct oce_common_cqe_info *cqe_info, struct mbuf **m) in oce_rx_mbuf_chain() argument
1513 POCE_SOFTC sc = (POCE_SOFTC) rq->parent; in oce_rx_mbuf_chain()
1520 if (rq->ring->cidx == rq->ring->pidx) { in oce_rx_mbuf_chain()
1525 pd = &rq->pckts[rq->ring->cidx]; in oce_rx_mbuf_chain()
1527 bus_dmamap_sync(rq->tag, pd->map, BUS_DMASYNC_POSTWRITE); in oce_rx_mbuf_chain()
1528 bus_dmamap_unload(rq->tag, pd->map); in oce_rx_mbuf_chain()
[all …]
/freebsd/sys/kern/
H A Dkern_switch.c260 runq_init(struct runq *rq) in runq_init() argument
264 bzero(rq, sizeof *rq); in runq_init()
266 TAILQ_INIT(&rq->rq_queues[i]); in runq_init()
274 runq_clrbit(struct runq *rq, int pri) in runq_clrbit() argument
278 rqb = &rq->rq_status; in runq_clrbit()
291 runq_findbit(struct runq *rq) in runq_findbit() argument
297 rqb = &rq->rq_status; in runq_findbit()
310 runq_findbit_from(struct runq *rq, u_char pri) in runq_findbit_from() argument
320 rqb = &rq->rq_status; in runq_findbit_from()
346 runq_setbit(struct runq *rq, int pri) in runq_setbit() argument
[all …]
/freebsd/sys/dev/ice/
H A Dice_controlq.c45 (qinfo)->rq.head = prefix##_ARQH; \
46 (qinfo)->rq.tail = prefix##_ARQT; \
47 (qinfo)->rq.len = prefix##_ARQLEN; \
48 (qinfo)->rq.bah = prefix##_ARQBAH; \
49 (qinfo)->rq.bal = prefix##_ARQBAL; \
50 (qinfo)->rq.len_mask = prefix##_ARQLEN_ARQLEN_M; \
51 (qinfo)->rq.len_ena_mask = prefix##_ARQLEN_ARQENABLE_M; \
52 (qinfo)->rq.len_crit_mask = prefix##_ARQLEN_ARQCRIT_M; \
53 (qinfo)->rq.head_mask = prefix##_ARQH_ARQH_M; \
144 cq->rq in ice_free_cq_ring()
[all...]
/freebsd/sys/contrib/openzfs/include/os/linux/kernel/linux/
H A Dblkdev_compat.h550 io_data_dir(struct bio *bio, struct request *rq) in io_data_dir() argument
552 if (rq != NULL) { in io_data_dir()
553 if (op_is_write(req_op(rq))) { in io_data_dir()
563 io_is_flush(struct bio *bio, struct request *rq) in io_is_flush() argument
565 if (rq != NULL) in io_is_flush()
566 return (req_op(rq) == REQ_OP_FLUSH); in io_is_flush()
571 io_is_discard(struct bio *bio, struct request *rq) in io_is_discard() argument
573 if (rq != NULL) in io_is_discard()
574 return (req_op(rq) == REQ_OP_DISCARD); in io_is_discard()
579 io_is_secure_erase(struct bio *bio, struct request *rq) in io_is_secure_erase() argument
[all …]
/freebsd/contrib/ofed/libcxgb4/
H A Dverbs.c342 qhp->wq.rq.msn = 1; in create_qp_v0()
343 qhp->wq.rq.qid = resp.rqid; in create_qp_v0()
344 qhp->wq.rq.size = resp.rq_size; in create_qp_v0()
345 qhp->wq.rq.memsize = resp.rq_memsize; in create_qp_v0()
366 qhp->wq.rq.udb = dbva; in create_qp_v0()
367 qhp->wq.rq.queue = mmap(NULL, qhp->wq.rq.memsize, in create_qp_v0()
370 if (qhp->wq.rq.queue == MAP_FAILED) in create_qp_v0()
377 qhp->wq.rq.sw_rq = calloc(qhp->wq.rq.size, sizeof (uint64_t)); in create_qp_v0()
378 if (!qhp->wq.rq.sw_rq) in create_qp_v0()
386 qhp->wq.rq.udb, qhp->wq.rq.queue, in create_qp_v0()
[all …]
H A Dt4.h355 struct t4_rq rq; member
365 return wq->rq.in_use; in t4_rqes_posted()
370 return wq->rq.in_use == 0; in t4_rq_empty()
375 return wq->rq.in_use == (wq->rq.size - 1); in t4_rq_full()
380 return wq->rq.size - 1 - wq->rq.in_use; in t4_rq_avail()
385 wq->rq.in_use++; in t4_rq_produce()
386 if (++wq->rq.pidx == wq->rq.size) in t4_rq_produce()
387 wq->rq.pidx = 0; in t4_rq_produce()
388 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce()
389 if (wq->rq.wq_pidx >= wq->rq.size * T4_RQ_NUM_SLOTS) in t4_rq_produce()
[all …]
H A Ddev.c300 qhp->wq.rq.qid, in dump_qp()
301 qhp->wq.rq.queue, in dump_qp()
302 qhp->wq.rq.sw_rq, in dump_qp()
303 qhp->wq.rq.cidx, in dump_qp()
304 qhp->wq.rq.pidx, in dump_qp()
305 qhp->wq.rq.in_use, in dump_qp()
306 qhp->wq.rq.size); in dump_qp()
343 cidx = qhp->wq.rq.cidx; in dump_qp()
344 pidx = qhp->wq.rq.pidx; in dump_qp()
348 swrqe = &qhp->wq.rq.sw_rq[cidx]; in dump_qp()
[all …]
/freebsd/sys/dev/ocs_fc/
H A Docs_hw_queues.c71 hw_rq_t *rq = NULL; in ocs_hw_init_queues() local
236 rq = hw_new_rq(cq, len, hw->ulp_start + qt->ulp); in ocs_hw_init_queues()
237 if (rq == NULL) { in ocs_hw_init_queues()
241 rq->filter_mask = qt->filter_mask; in ocs_hw_init_queues()
296 rq = hw_new_rq(cq, mrq.rq_cfg[i].len, mrq.rq_cfg[i].ulp); in ocs_hw_init_queues()
297 if (rq == NULL) { in ocs_hw_init_queues()
301 rq->filter_mask = mrq.rq_cfg[i].filter_mask; in ocs_hw_init_queues()
593 hw_rq_t *rq = ocs_malloc(hw->os, sizeof(*rq), OCS_M_ZERO | OCS_M_NOWAIT); in hw_new_rq() local
598 if (rq != NULL) { in hw_new_rq()
599 rq->instance = hw->hw_rq_count++; in hw_new_rq()
[all …]
/freebsd/sys/dev/cxgbe/iw_cxgbe/
H A Dt4.h362 struct t4_rq rq; member
369 return wq->rq.in_use; in t4_rqes_posted()
374 return wq->rq.in_use == 0; in t4_rq_empty()
379 return wq->rq.in_use == (wq->rq.size - 1); in t4_rq_full()
384 return wq->rq.size - 1 - wq->rq.in_use; in t4_rq_avail()
389 wq->rq.in_use++; in t4_rq_produce()
390 if (++wq->rq.pidx == wq->rq.size) in t4_rq_produce()
391 wq->rq.pidx = 0; in t4_rq_produce()
392 wq->rq.wq_pidx += DIV_ROUND_UP(len16*16, T4_EQ_ENTRY_SIZE); in t4_rq_produce()
393 if (wq->rq.wq_pidx >= wq->rq.size * T4_RQ_NUM_SLOTS) in t4_rq_produce()
[all …]
H A Dqp.c112 wq->rq.memsize, wq->rq.queue, in destroy_qp()
113 dma_unmap_addr(&wq->rq, mapping)); in destroy_qp()
117 c4iw_rqtpool_free(rdev, wq->rq.rqt_hwaddr, wq->rq.rqt_size); in destroy_qp()
118 kfree(wq->rq.sw_rq); in destroy_qp()
120 c4iw_put_qpid(rdev, wq->rq.qid, uctx); in destroy_qp()
148 wq->rq.qid = c4iw_get_qpid(rdev, uctx); in create_qp()
149 if (!wq->rq.qid) { in create_qp()
162 wq->rq.sw_rq = kzalloc(wq->rq.size * sizeof *wq->rq.sw_rq, in create_qp()
164 if (!wq->rq.sw_rq) { in create_qp()
173 wq->rq.rqt_size = roundup_pow_of_two(max_t(u16, wq->rq.size, 16)); in create_qp()
[all …]
/freebsd/sys/dev/bnxt/bnxt_re/
H A Dqplib_fp.c74 if (!qp->rq.flushed) { in __bnxt_qplib_add_flush_qp()
79 qp->rq.flushed = true; in __bnxt_qplib_add_flush_qp()
120 if (qp->rq.flushed) { in __bnxt_qplib_del_flush_qp()
121 qp->rq.flushed = false; in __bnxt_qplib_del_flush_qp()
137 qp->rq.hwq.prod = 0; in bnxt_qplib_clean_qp()
138 qp->rq.hwq.cons = 0; in bnxt_qplib_clean_qp()
139 qp->rq.swq_start = 0; in bnxt_qplib_clean_qp()
140 qp->rq.swq_last = 0; in bnxt_qplib_clean_qp()
229 qp->rq.max_wqe); in bnxt_qplib_alloc_hdr_buf()
912 struct bnxt_qplib_q *rq = &qp->rq; in bnxt_qplib_create_qp1() local
[all …]
/freebsd/contrib/ofed/libmlx4/
H A Dqp.c61 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe()
88 qp->rq.head = 0; in mlx4_init_qp_indices()
89 qp->rq.tail = 0; in mlx4_init_qp_indices()
522 pthread_spin_lock(&qp->rq.lock); in mlx4_post_recv()
526 ind = qp->rq.head & (qp->rq.wqe_cnt - 1); in mlx4_post_recv()
529 if (wq_overflow(&qp->rq, nreq, to_mcq(ibqp->recv_cq))) { in mlx4_post_recv()
535 if (wr->num_sge > qp->rq.max_gs) { in mlx4_post_recv()
546 if (i < qp->rq.max_gs) { in mlx4_post_recv()
552 qp->rq.wrid[ind] = wr->wr_id; in mlx4_post_recv()
554 ind = (ind + 1) & (qp->rq.wqe_cnt - 1); in mlx4_post_recv()
[all …]
/freebsd/sys/arm/include/
H A Dsysreg.h271 #define CP15_CNTPCT(rq, rr) p15, 0, rq, rr, c14 /* Physical Count Register */ argument
272 #define CP15_CNTVCT(rq, rr) p15, 1, rq, rr, c14 /* Virtual Count Register */ argument
273 #define CP15_CNTP_CVAL(rq, rr) p15, 2, rq, rr, c14 /* PL1 Physical Timer Compare Value Register */ argument
274 #define CP15_CNTV_CVAL(rq, rr) p15, 3, rq, rr, c14 /* Virtual Timer Compare Value Register */ argument
275 #define CP15_CNTVOFF(rq, rr) p15, 4, rq, rr, c14 /* Virtual Offset Register */ argument
276 #define CP15_CNTHP_CVAL(rq, rr) p15, 6, rq, rr, c14 /* PL2 Physical Timer Compare Value Register */ argument
278 #define CP15_VTTBR(rq, rr) p15, 6, rq, rr, c2 /* Virtualization Translation Table Base Register */ argument
279 #define CP15_HTTBR(rq, rr) p15, 4, rq, rr, c2 /* Hyp Translation Table Base Register */ argument
280 #define CP15_TTBR0_2(rq, rr) p15, 0, rq, rr, c2 /* Translation Table Base Register 0 */ argument
281 #define CP15_TTBR1_2(rq, rr) p15, 1, rq, rr, c2 /* Translation Table Base Register 1 */ argument
[all …]
/freebsd/sys/dev/mthca/
H A Dmthca_qp.c211 return qp->queue.direct.buf + (n << qp->rq.wqe_shift); in get_recv_wqe()
213 return qp->queue.page_list[(n << qp->rq.wqe_shift) >> PAGE_SHIFT].buf + in get_recv_wqe()
214 ((n << qp->rq.wqe_shift) & (PAGE_SIZE - 1)); in get_recv_wqe()
500 qp_attr->cap.max_recv_wr = qp->rq.max; in mthca_query_qp()
502 qp_attr->cap.max_recv_sge = qp->rq.max_gs; in mthca_query_qp()
603 if (qp->rq.max) in __mthca_modify_qp()
604 qp_context->rq_size_stride = ilog2(qp->rq.max) << 3; in __mthca_modify_qp()
605 qp_context->rq_size_stride |= qp->rq.wqe_shift - 4; in __mthca_modify_qp()
765 qp_context->rcv_db_index = cpu_to_be32(qp->rq.db_index); in __mthca_modify_qp()
831 mthca_wq_reset(&qp->rq); in __mthca_modify_qp()
[all …]
/freebsd/sys/contrib/openzfs/module/os/linux/zfs/
H A Dzvol_os.c50 struct request *rq, boolean_t force_sync);
95 zvol_end_io(struct bio *bio, struct request *rq, int error) in zvol_end_io() argument
101 blk_mq_end_request(rq, errno_to_bi_status(error)); in zvol_end_io()
129 struct request *rq; member
133 struct request *rq; member
165 struct request *rq = bd->rq; in zvol_mq_queue_rq() local
166 zvol_state_t *zv = rq->q->queuedata; in zvol_mq_queue_rq()
169 blk_mq_start_request(rq); in zvol_mq_queue_rq()
171 if (blk_rq_is_passthrough(rq)) { in zvol_mq_queue_rq()
173 blk_mq_end_request(rq, BLK_STS_IOERR); in zvol_mq_queue_rq()
[all …]
/freebsd/sys/dev/cfi/
H A Dcfi_dev.c265 struct cfiocqry *rq; in cfi_devioctl() local
279 rq = (struct cfiocqry *)data; in cfi_devioctl()
280 if (rq->offset >= sc->sc_size / sc->sc_width) in cfi_devioctl()
282 if (rq->offset > ULONG_MAX - rq->count || in cfi_devioctl()
283 rq->offset + rq->count > sc->sc_size / sc->sc_width) in cfi_devioctl()
286 while (!error && rq->count--) { in cfi_devioctl()
287 val = cfi_read_qry(sc, rq->offset++); in cfi_devioctl()
288 error = copyout(&val, rq->buffer++, 1); in cfi_devioctl()
/freebsd/contrib/ofed/libmlx5/
H A Dqp.c63 return qp->buf.buf + qp->rq.offset + (n << qp->rq.wqe_shift); in get_recv_wqe()
68 return rwq->pbuff + (n << rwq->rq.wqe_shift); in get_wq_recv_wqe()
96 int max = 1 << (qp->rq.wqe_shift - 4); in mlx5_copy_to_recv_wqe()
160 rwq->rq.head = 0; in mlx5_init_rwq_indices()
161 rwq->rq.tail = 0; in mlx5_init_rwq_indices()
168 qp->rq.head = 0; in mlx5_init_qp_indices()
169 qp->rq.tail = 0; in mlx5_init_qp_indices()
1072 mlx5_spin_lock(&rwq->rq.lock); in mlx5_post_wq_recv()
1074 ind = rwq->rq.head & (rwq->rq.wqe_cnt - 1); in mlx5_post_wq_recv()
1077 if (unlikely(mlx5_wq_overflow(&rwq->rq, nreq, in mlx5_post_wq_recv()
[all …]
/freebsd/sys/dev/qlnx/qlnxr/
H A Dqlnxr_cm.c125 qp->rqe_wr_id[qp->rq.gsi_cons].rc = in qlnxr_ll2_complete_rx_packet()
127 qp->rqe_wr_id[qp->rq.gsi_cons].vlan_id = data->vlan; in qlnxr_ll2_complete_rx_packet()
129 qp->rqe_wr_id[qp->rq.gsi_cons].sg_list[0].length = in qlnxr_ll2_complete_rx_packet()
131 *((u32 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[0]) = in qlnxr_ll2_complete_rx_packet()
133 *((u16 *)&qp->rqe_wr_id[qp->rq.gsi_cons].smac[4]) = in qlnxr_ll2_complete_rx_packet()
136 qlnxr_inc_sw_gsi_cons(&qp->rq); in qlnxr_ll2_complete_rx_packet()
411 qp->rq.max_wr = attrs->cap.max_recv_wr; in qlnxr_create_gsi_qp()
414 qp->rqe_wr_id = kzalloc(qp->rq.max_wr * sizeof(*qp->rqe_wr_id), in qlnxr_create_gsi_qp()
780 memset(&qp->rqe_wr_id[qp->rq.prod], 0, in qlnxr_gsi_post_recv()
781 sizeof(qp->rqe_wr_id[qp->rq.prod])); in qlnxr_gsi_post_recv()
[all …]

1234567