Home
last modified time | relevance | path

Searched refs:rxq (Results 1 – 25 of 73) sorted by relevance

123

/freebsd/sys/dev/sfxge/
H A Dsfxge_rx.c155 sfxge_rx_qflush_done(struct sfxge_rxq *rxq) in sfxge_rx_qflush_done() argument
158 rxq->flush_state = SFXGE_FLUSH_DONE; in sfxge_rx_qflush_done()
162 sfxge_rx_qflush_failed(struct sfxge_rxq *rxq) in sfxge_rx_qflush_failed() argument
165 rxq->flush_state = SFXGE_FLUSH_FAILED; in sfxge_rx_qflush_failed()
183 struct sfxge_rxq *rxq = arg; in sfxge_rx_post_refill() local
189 sc = rxq->sc; in sfxge_rx_post_refill()
190 index = rxq->index; in sfxge_rx_post_refill()
192 magic = sfxge_sw_ev_rxq_magic(SFXGE_SW_EV_RX_QREFILL, rxq); in sfxge_rx_post_refill()
197 KASSERT(rxq->init_state == SFXGE_RXQ_STARTED, in sfxge_rx_post_refill()
203 sfxge_rx_schedule_refill(struct sfxge_rxq *rxq, boolean_t retrying) in sfxge_rx_schedule_refill() argument
[all …]
H A Dsfxge_ev.c53 struct sfxge_rxq *rxq; in sfxge_ev_qcomplete() local
60 rxq = sc->rxq[index]; in sfxge_ev_qcomplete()
82 if (rxq->pending != rxq->completed) in sfxge_ev_qcomplete()
83 sfxge_rx_qcomplete(rxq, eop); in sfxge_ev_qcomplete()
89 struct sfxge_rxq *rxq; in sfxge_get_rxq_by_label() local
93 rxq = evq->sc->rxq[evq->index]; in sfxge_get_rxq_by_label()
95 KASSERT(rxq != NULL, ("rxq == NULL")); in sfxge_get_rxq_by_label()
96 KASSERT(evq->index == rxq->index, ("evq->index != rxq->index")); in sfxge_get_rxq_by_label()
98 return (rxq); in sfxge_get_rxq_by_label()
107 struct sfxge_rxq *rxq; in sfxge_ev_rx() local
[all …]
H A Dsfxge_rx.h187 extern void sfxge_rx_qcomplete(struct sfxge_rxq *rxq, boolean_t eop);
188 extern void sfxge_rx_qrefill(struct sfxge_rxq *rxq);
189 extern void sfxge_rx_qflush_done(struct sfxge_rxq *rxq);
190 extern void sfxge_rx_qflush_failed(struct sfxge_rxq *rxq);
/freebsd/sys/contrib/dev/iwlwifi/pcie/
H A Drx.c55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free.
58 * or reused - added immediately to the iwl->rxq->rx_free list.
66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them
76 * detached from the iwl->rxq. The driver 'processed' index is updated.
77 * + If there are no allocated buffers in iwl->rxq->rx_free,
104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue
108 * rxq
119 iwl_rxq_space(const struct iwl_rxq * rxq) iwl_rxq_space() argument
167 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans * trans,struct iwl_rxq * rxq) iwl_pcie_rxq_inc_wr_ptr() argument
209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local
221 iwl_pcie_restock_bd(struct iwl_trans * trans,struct iwl_rxq * rxq,struct iwl_rx_mem_buffer * rxb) iwl_pcie_restock_bd() argument
[all...]
H A Dinternal.h192 * @rxq: the rxq to get the rb stts from
195 struct iwl_rxq *rxq) in iwl_get_closed_rb_stts()
198 __le16 *rb_stts = rxq->rb_stts; in iwl_get_closed_rb_stts()
202 struct iwl_rb_status *rb_stts = rxq->rb_stts; in iwl_get_closed_rb_stts()
342 * @rxq: all the RX queue data
424 struct iwl_rxq *rxq;
578 void iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq); in iwl_pcie_get_num_sections()
581 struct iwl_rxq *rxq); in iwl_pcie_get_num_sections()
194 iwl_get_closed_rb_stts(struct iwl_trans * trans,struct iwl_rxq * rxq) iwl_get_closed_rb_stts() argument
352 struct iwl_rxq *rxq; global() member
H A Dctxt-info.c215 rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma); in iwl_pcie_ctxt_info_init()
216 rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma); in iwl_pcie_ctxt_info_init()
217 rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma); in iwl_pcie_ctxt_info_init()
/freebsd/sys/dev/xen/netfront/
H A Dnetfront.c221 struct netfront_rxq *rxq; member
294 xn_get_rx_mbuf(struct netfront_rxq *rxq, RING_IDX ri) in xn_get_rx_mbuf() argument
300 m = rxq->mbufs[i]; in xn_get_rx_mbuf()
301 rxq->mbufs[i] = NULL; in xn_get_rx_mbuf()
306 xn_get_rx_ref(struct netfront_rxq *rxq, RING_IDX ri) in xn_get_rx_ref() argument
309 grant_ref_t ref = rxq->grant_ref[i]; in xn_get_rx_ref()
312 rxq->grant_ref[i] = GRANT_REF_INVALID; in xn_get_rx_ref()
479 XN_RX_LOCK(&np->rxq[i]); in netfront_suspend()
484 XN_RX_UNLOCK(&np->rxq[i]); in netfront_suspend()
504 XN_RX_LOCK(&info->rxq[i]); in netfront_resume()
[all …]
/freebsd/sys/dev/netmap/
H A Dif_vtnet_netmap.h164 struct vtnet_rxq *rxq = &sc->vtnet_rxqs[ring_nr]; in vtnet_netmap_kring_refill() local
165 struct virtqueue *vq = rxq->vtnrx_vq; in vtnet_netmap_kring_refill()
171 for (nm_i = rxq->vtnrx_nm_refill; num > 0; in vtnet_netmap_kring_refill()
186 err = sglist_append(&sg, &rxq->vtnrx_shrhdr, sc->vtnet_hdr_size); in vtnet_netmap_kring_refill()
192 err = virtqueue_enqueue(vq, /*cookie=*/rxq, &sg, in vtnet_netmap_kring_refill()
200 rxq->vtnrx_nm_refill = nm_i; in vtnet_netmap_kring_refill()
212 vtnet_netmap_rxq_populate(struct vtnet_rxq *rxq) in vtnet_netmap_rxq_populate() argument
214 struct netmap_adapter *na = NA(rxq->vtnrx_sc->vtnet_ifp); in vtnet_netmap_rxq_populate()
220 slot = netmap_reset(na, NR_RX, rxq->vtnrx_id, 0); in vtnet_netmap_rxq_populate()
223 kring = na->rx_rings[rxq->vtnrx_id]; in vtnet_netmap_rxq_populate()
[all …]
/freebsd/sys/dev/virtio/network/
H A Dif_vtnet.c148 static void vtnet_rx_vq_process(struct vtnet_rxq *rxq, int tries);
821 struct vtnet_rxq *rxq; in vtnet_init_rxq() local
823 rxq = &sc->vtnet_rxqs[id]; in vtnet_init_rxq()
825 snprintf(rxq->vtnrx_name, sizeof(rxq->vtnrx_name), "%s-rx%d", in vtnet_init_rxq()
827 mtx_init(&rxq->vtnrx_mtx, rxq->vtnrx_name, NULL, MTX_DEF); in vtnet_init_rxq()
829 rxq->vtnrx_sc = sc; in vtnet_init_rxq()
830 rxq->vtnrx_id = id; in vtnet_init_rxq()
832 rxq->vtnrx_sg = sglist_alloc(sc->vtnet_rx_nsegs, M_NOWAIT); in vtnet_init_rxq()
833 if (rxq->vtnrx_sg == NULL) in vtnet_init_rxq()
838 if (tcp_lro_init_args(&rxq->vtnrx_lro, sc->vtnet_ifp, in vtnet_init_rxq()
[all …]
/freebsd/sys/dev/mana/
H A Dmana_en.c410 mana_load_rx_mbuf(struct mana_port_context *apc, struct mana_rxq *rxq, in mana_load_rx_mbuf() argument
419 mbuf = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, rxq->datasize); in mana_load_rx_mbuf()
427 mlen = rxq->datasize; in mana_load_rx_mbuf()
446 counter_u64_add(rxq->stats.dma_mapping_err, 1); in mana_load_rx_mbuf()
467 mana_unload_rx_mbuf(struct mana_port_context *apc, struct mana_rxq *rxq, in mana_unload_rx_mbuf() argument
1340 mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) in mana_fence_rq() argument
1346 init_completion(&rxq->fence_event); in mana_fence_rq()
1350 req.wq_obj_handle = rxq->rxobj; in mana_fence_rq()
1356 rxq->rxq_idx, err); in mana_fence_rq()
1363 rxq->rxq_idx, err, resp.hdr.status); in mana_fence_rq()
[all …]
H A Dmana_sysctl.c54 struct mana_rxq *rxq; in mana_sysctl_rx_stat_agg_u64() local
59 rxq = apc->rxqs[i]; in mana_sysctl_rx_stat_agg_u64()
60 stat += *((uint64_t *)((uint8_t *)rxq + offset)); in mana_sysctl_rx_stat_agg_u64()
68 rxq = apc->rxqs[i]; in mana_sysctl_rx_stat_agg_u64()
69 *((uint64_t *)((uint8_t *)rxq + offset)) = 0; in mana_sysctl_rx_stat_agg_u64()
79 struct mana_rxq *rxq; in mana_sysctl_rx_stat_u16() local
83 rxq = apc->rxqs[0]; in mana_sysctl_rx_stat_u16()
84 val = *((uint16_t *)((uint8_t *)rxq + offset)); in mana_sysctl_rx_stat_u16()
99 struct mana_rxq *rxq; in mana_sysctl_rx_stat_u32() local
103 rxq = apc->rxqs[0]; in mana_sysctl_rx_stat_u32()
[all …]
H A Dhw_channel.c173 hwc->rxq->gdma_wq->id = val; in mana_hwc_init_event_handler()
201 hwc->rxq->msg_buf->gpa_mkey = val; in mana_hwc_init_event_handler()
223 struct hwc_wq *hwc_rxq = hwc->rxq; in mana_hwc_rx_event_handler()
646 struct hwc_wq *hwc_rxq = hwc->rxq; in mana_hwc_test_channel()
677 struct gdma_queue *rq = hwc->rxq->gdma_wq; in mana_hwc_establish_channel()
738 hwc->cq, &hwc->rxq); in mana_hwc_init_queues()
841 if (hwc->rxq) in mana_hwc_destroy_channel()
842 mana_hwc_destroy_wq(hwc, hwc->rxq); in mana_hwc_destroy_channel()
/freebsd/sys/dev/ice/
H A Dice_iflib_txrx.c52 static int _ice_ift_rxd_available(struct ice_rx_queue *rxq, qidx_t pidx, qidx_t budget);
53 static int _ice_ift_rxd_pkt_get(struct ice_rx_queue *rxq, if_rxd_info_t ri);
54 static void _ice_ift_rxd_refill(struct ice_rx_queue *rxq, uint32_t pidx,
56 static void _ice_ift_rxd_flush(struct ice_softc *sc, struct ice_rx_queue *rxq,
315 * @rxq: RX queue driver structure in ice_ift_rxd_pkt_get()
323 _ice_ift_rxd_available(struct ice_rx_queue *rxq, qidx_t pidx, qidx_t budget) in ice_ift_rxd_pkt_get()
329 nrxd = rxq->desc_count; in ice_ift_rxd_pkt_get()
332 rxd = &rxq->rx_base[i]; in ice_ift_rxd_pkt_get()
360 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid];
362 return _ice_ift_rxd_available(rxq, pid
251 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid]; ice_ift_rxd_available() local
287 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[ri->iri_qsidx]; ice_ift_rxd_pkt_get() local
367 struct ice_rx_queue *rxq; ice_ift_rxd_refill() local
403 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid]; ice_ift_rxd_flush() local
[all...]
H A Dice_common_txrx.h261 * @rxq: the Rx queue structure
272 ice_rx_checksum(struct ice_rx_queue *rxq, uint32_t *flags, uint32_t *data, in ice_rx_checksum() argument
326 rxq->stats.cso[ICE_CSO_STAT_RX_IP6_ERR]++; in ice_rx_checksum()
337 rxq->stats.cso[ICE_CSO_STAT_RX_IP4_ERR]++; in ice_rx_checksum()
341 rxq->stats.cso[ICE_CSO_STAT_RX_L3_ERR]++; in ice_rx_checksum()
352 rxq->stats.cso[ICE_CSO_STAT_RX_TCP_ERR]++; in ice_rx_checksum()
356 rxq->stats.cso[ICE_CSO_STAT_RX_UDP_ERR]++; in ice_rx_checksum()
360 rxq->stats.cso[ICE_CSO_STAT_RX_SCTP_ERR]++; in ice_rx_checksum()
368 rxq->stats.cso[ICE_CSO_STAT_RX_L4_ERR]++; in ice_rx_checksum()
H A Dif_ice_iflib.c1122 struct ice_rx_queue *rxq; in ice_if_rx_queues_alloc() local
1152 for (i = 0, rxq = vsi->rx_queues; i < nrxqsets; i++, rxq++) { in ice_if_rx_queues_alloc()
1153 rxq->me = i; in ice_if_rx_queues_alloc()
1154 rxq->vsi = vsi; in ice_if_rx_queues_alloc()
1157 rxq->desc_count = sc->scctx->isc_nrxd[0]; in ice_if_rx_queues_alloc()
1160 rxq->tail = QRX_TAIL(vsi->rx_qmap[i]); in ice_if_rx_queues_alloc()
1161 rxq->rx_base = (union ice_32b_rx_flex_desc *)vaddrs[i]; in ice_if_rx_queues_alloc()
1162 rxq->rx_paddr = paddrs[i]; in ice_if_rx_queues_alloc()
1164 ice_add_rxq_sysctls(rxq); in ice_if_rx_queues_alloc()
1243 struct ice_rx_queue __unused *rxq = (struct ice_rx_queue *)arg; in ice_msix_que() local
[all …]
/freebsd/sys/dev/qlnx/qlnxe/
H A Dqlnx_os.c133 static int qlnx_alloc_rx_buffer(qlnx_host_t *ha, struct qlnx_rx_queue *rxq);
134 static void qlnx_reuse_rx_data(struct qlnx_rx_queue *rxq);
136 struct qlnx_rx_queue *rxq);
1462 if (fp->rxq->handle != NULL) { in qlnx_set_rx_coalesce()
1464 0, fp->rxq->handle); in qlnx_set_rx_coalesce()
3851 struct qlnx_rx_queue *rxq; in qlnx_rx_jumbo_chain() local
3854 rxq = fp->rxq; in qlnx_rx_jumbo_chain()
3858 rxq->sw_rx_cons = (rxq->sw_rx_cons + 1) & (RX_RING_SIZE - 1); in qlnx_rx_jumbo_chain()
3860 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_cons]; in qlnx_rx_jumbo_chain()
3866 rxq->sw_rx_cons = in qlnx_rx_jumbo_chain()
[all …]
/freebsd/sys/dev/vmware/vmxnet3/
H A Dif_vmx.c470 struct vmxnet3_rxqueue *rxq; in vmxnet3_msix_intr_assign() local
481 rxq = &sc->vmx_rxq[i]; in vmxnet3_msix_intr_assign()
482 error = iflib_irq_alloc_generic(ctx, &rxq->vxrxq_irq, i + 1, in vmxnet3_msix_intr_assign()
483 IFLIB_INTR_RXTX, vmxnet3_rxq_intr, rxq, i, irq_name); in vmxnet3_msix_intr_assign()
527 struct vmxnet3_rxqueue *rxq; in vmxnet3_free_irqs() local
533 rxq = &sc->vmx_rxq[i]; in vmxnet3_free_irqs()
534 iflib_irq_free(sc->vmx_ctx, &rxq->vxrxq_irq); in vmxnet3_free_irqs()
689 struct vmxnet3_rxqueue *rxq; in vmxnet3_set_interrupt_idx() local
711 rxq = &sc->vmx_rxq[i]; in vmxnet3_set_interrupt_idx()
712 rxs = rxq->vxrxq_rs; in vmxnet3_set_interrupt_idx()
[all …]
/freebsd/sys/net/
H A Diflib.c141 static void iru_init(if_rxd_update_t iru, iflib_rxq_t rxq, uint8_t flid);
701 static int iflib_rxd_avail(if_ctx_t ctx, iflib_rxq_t rxq, qidx_t cidx, qidx_t budget);
768 static int netmap_fl_refill(iflib_rxq_t rxq, struct netmap_kring *kring, bool init);
842 iflib_rxq_t rxq = &ctx->ifc_rxqs[0]; in iflib_netmap_config() local
843 iflib_fl_t fl = &rxq->ifr_fl[0]; in iflib_netmap_config()
858 netmap_fl_refill(iflib_rxq_t rxq, struct netmap_kring *kring, bool init) in netmap_fl_refill() argument
865 if_ctx_t ctx = rxq->ifr_ctx; in netmap_fl_refill()
866 iflib_fl_t fl = &rxq->ifr_fl[0]; in netmap_fl_refill()
900 iru_init(&iru, rxq, 0 /* flid */); in netmap_fl_refill()
964 ctx->isc_rxd_flush(ctx->ifc_softc, rxq->ifr_id, fl->ifl_id, in netmap_fl_refill()
[all …]
/freebsd/sys/dev/iavf/
H A Diavf_vc_iflib.c102 vqpi->rxq.vsi_id = vqci->vsi_id; in iavf_configure_queues()
103 vqpi->rxq.queue_id = i; in iavf_configure_queues()
104 vqpi->rxq.ring_len = scctx->isc_nrxd[0]; in iavf_configure_queues()
105 vqpi->rxq.dma_ring_addr = rxr->rx_paddr; in iavf_configure_queues()
106 vqpi->rxq.max_pkt_size = scctx->isc_max_frame_size; in iavf_configure_queues()
107 vqpi->rxq.databuffer_size = rxr->mbuf_sz; in iavf_configure_queues()
108 vqpi->rxq.splithdr_enabled = 0; in iavf_configure_queues()
/freebsd/sys/dev/gve/
H A Dgve_sysctl.c50 struct sysctl_oid_list *child, struct gve_rx_ring *rxq) in gve_setup_rxq_sysctl() argument
57 snprintf(namebuf, sizeof(namebuf), "rxq%d", rxq->com.id); in gve_setup_rxq_sysctl()
62 stats = &rxq->stats; in gve_setup_rxq_sysctl()
104 &rxq->cnt, 0, "Number of descriptors completed"); in gve_setup_rxq_sysctl()
107 &rxq->fill_cnt, rxq->fill_cnt, in gve_setup_rxq_sysctl()
/freebsd/sys/dev/ral/
H A Drt2560.c252 error = rt2560_alloc_rx_ring(sc, &sc->rxq, RT2560_RX_RING_COUNT); in rt2560_attach()
352 rt2560_free_rx_ring(sc, &sc->rxq); in rt2560_detach()
649 desc = &sc->rxq.desc[i]; in rt2560_alloc_rx_ring()
650 data = &sc->rxq.data[i]; in rt2560_alloc_rx_ring()
1097 hw = RAL_READ(sc, RT2560_SECCSR0) - sc->rxq.physaddr; in rt2560_decryption_intr()
1100 bus_dmamap_sync(sc->rxq.desc_dmat, sc->rxq.desc_map, in rt2560_decryption_intr()
1103 for (; sc->rxq.cur_decrypt != hw;) { in rt2560_decryption_intr()
1104 desc = &sc->rxq.desc[sc->rxq.cur_decrypt]; in rt2560_decryption_intr()
1105 data = &sc->rxq.data[sc->rxq.cur_decrypt]; in rt2560_decryption_intr()
1135 bus_dmamap_sync(sc->rxq.data_dmat, data->map, in rt2560_decryption_intr()
[all …]
H A Drt2661.c255 error = rt2661_alloc_rx_ring(sc, &sc->rxq, RT2661_RX_RING_COUNT); in rt2661_attach()
345 rt2661_free_rx_ring(sc, &sc->rxq); in rt2661_detach()
656 desc = &sc->rxq.desc[i]; in rt2661_alloc_rx_ring()
657 data = &sc->rxq.data[i]; in rt2661_alloc_rx_ring()
967 bus_dmamap_sync(sc->rxq.desc_dmat, sc->rxq.desc_map, in rt2661_rx_intr()
973 desc = &sc->rxq.desc[sc->rxq.cur]; in rt2661_rx_intr()
974 data = &sc->rxq.data[sc->rxq.cur]; in rt2661_rx_intr()
1009 bus_dmamap_sync(sc->rxq.data_dmat, data->map, in rt2661_rx_intr()
1011 bus_dmamap_unload(sc->rxq.data_dmat, data->map); in rt2661_rx_intr()
1013 error = bus_dmamap_load(sc->rxq.data_dmat, data->map, in rt2661_rx_intr()
[all …]
/freebsd/sys/dev/ixl/
H A Dixl_pf_iov.c605 struct i40e_hmc_obj_rxq rxq; in ixl_vf_config_rx_queue() local
611 bzero(&rxq, sizeof(rxq)); in ixl_vf_config_rx_queue()
627 rxq.hsplit_0 = info->rx_split_pos & in ixl_vf_config_rx_queue()
632 rxq.hbuff = info->hdr_size >> I40E_RXQ_CTX_HBUFF_SHIFT; in ixl_vf_config_rx_queue()
634 rxq.dtype = 2; in ixl_vf_config_rx_queue()
641 rxq.base = info->dma_ring_addr / IXL_RX_CTX_BASE_UNITS; in ixl_vf_config_rx_queue()
642 rxq.qlen = info->ring_len; in ixl_vf_config_rx_queue()
644 rxq.dbuff = info->databuffer_size >> I40E_RXQ_CTX_DBUFF_SHIFT; in ixl_vf_config_rx_queue()
646 rxq.dsize = 1; in ixl_vf_config_rx_queue()
647 rxq.crcstrip = 1; in ixl_vf_config_rx_queue()
[all …]
/freebsd/sys/dev/nfe/
H A Dif_nfe.c577 if ((error = nfe_alloc_rx_ring(sc, &sc->rxq)) != 0) in nfe_attach()
762 nfe_free_rx_ring(sc, &sc->rxq); in nfe_detach()
1162 data = &sc->rxq.data[i]; in nfe_alloc_rx_ring()
1914 data = &sc->rxq.data[idx]; in nfe_discard_rxbuf()
1918 desc64 = &sc->rxq.desc64[idx]; in nfe_discard_rxbuf()
1925 desc32 = &sc->rxq.desc32[idx]; in nfe_discard_rxbuf()
1974 if (bus_dmamap_load_mbuf_sg(sc->rxq.rx_data_tag, sc->rxq.rx_spare_map, in nfe_newbuf()
1981 data = &sc->rxq.data[idx]; in nfe_newbuf()
1983 bus_dmamap_sync(sc->rxq.rx_data_tag, data->rx_data_map, in nfe_newbuf()
1985 bus_dmamap_unload(sc->rxq.rx_data_tag, data->rx_data_map); in nfe_newbuf()
[all …]
/freebsd/sys/dev/cxgbe/
H A Dt4_sge.c1109 struct sge_rxq *rxq; in t4_setup_vi_queues() local
1166 for_each_rxq(vi, i, rxq) { in t4_setup_vi_queues()
1167 rc = alloc_rxq(vi, rxq, i, intr_idx, maxp); in t4_setup_vi_queues()
1217 struct sge_rxq *rxq; in t4_teardown_vi_queues() local
1260 for_each_rxq(vi, i, rxq) { in t4_teardown_vi_queues()
1261 free_rxq(vi, rxq); in t4_teardown_vi_queues()
1374 MPASS(irq->rxq != NULL); in t4_vi_intr()
1375 t4_intr(irq->rxq); in t4_vi_intr()
1572 struct sge_rxq *rxq = iq_to_rxq(iq); in service_iq_fl() local
1583 struct lro_ctrl *lro = &rxq->lro; in service_iq_fl()
[all …]

123