/freebsd/sys/dev/sfxge/ |
H A D | sfxge_rx.c | 155 sfxge_rx_qflush_done(struct sfxge_rxq *rxq) in sfxge_rx_qflush_done() argument 158 rxq->flush_state = SFXGE_FLUSH_DONE; in sfxge_rx_qflush_done() 162 sfxge_rx_qflush_failed(struct sfxge_rxq *rxq) in sfxge_rx_qflush_failed() argument 165 rxq->flush_state = SFXGE_FLUSH_FAILED; in sfxge_rx_qflush_failed() 183 struct sfxge_rxq *rxq = arg; in sfxge_rx_post_refill() local 189 sc = rxq->sc; in sfxge_rx_post_refill() 190 index = rxq->index; in sfxge_rx_post_refill() 192 magic = sfxge_sw_ev_rxq_magic(SFXGE_SW_EV_RX_QREFILL, rxq); in sfxge_rx_post_refill() 197 KASSERT(rxq->init_state == SFXGE_RXQ_STARTED, in sfxge_rx_post_refill() 203 sfxge_rx_schedule_refill(struct sfxge_rxq *rxq, boolean_t retrying) in sfxge_rx_schedule_refill() argument [all …]
|
H A D | sfxge_ev.c | 53 struct sfxge_rxq *rxq; in sfxge_ev_qcomplete() local 60 rxq = sc->rxq[index]; in sfxge_ev_qcomplete() 82 if (rxq->pending != rxq->completed) in sfxge_ev_qcomplete() 83 sfxge_rx_qcomplete(rxq, eop); in sfxge_ev_qcomplete() 89 struct sfxge_rxq *rxq; in sfxge_get_rxq_by_label() local 93 rxq = evq->sc->rxq[evq->index]; in sfxge_get_rxq_by_label() 95 KASSERT(rxq != NULL, ("rxq == NULL")); in sfxge_get_rxq_by_label() 96 KASSERT(evq->index == rxq->index, ("evq->index != rxq->index")); in sfxge_get_rxq_by_label() 98 return (rxq); in sfxge_get_rxq_by_label() 107 struct sfxge_rxq *rxq; in sfxge_ev_rx() local [all …]
|
H A D | sfxge_rx.h | 187 extern void sfxge_rx_qcomplete(struct sfxge_rxq *rxq, boolean_t eop); 188 extern void sfxge_rx_qrefill(struct sfxge_rxq *rxq); 189 extern void sfxge_rx_qflush_done(struct sfxge_rxq *rxq); 190 extern void sfxge_rx_qflush_failed(struct sfxge_rxq *rxq);
|
/freebsd/sys/contrib/dev/iwlwifi/pcie/ |
H A D | rx.c | 55 * + A list of pre-allocated RBDs is stored in iwl->rxq->rx_free. 58 * or reused - added immediately to the iwl->rxq->rx_free list. 66 * add them to iwl->rxq->rx_free. If it fails - it continues to claim them 76 * detached from the iwl->rxq. The driver 'processed' index is updated. 77 * + If there are no allocated buffers in iwl->rxq->rx_free, 104 * rxq.pool -> rxq.rx_used -> rxq.rx_free -> rxq.queue 108 * rxq 119 iwl_rxq_space(const struct iwl_rxq * rxq) iwl_rxq_space() argument 167 iwl_pcie_rxq_inc_wr_ptr(struct iwl_trans * trans,struct iwl_rxq * rxq) iwl_pcie_rxq_inc_wr_ptr() argument 209 struct iwl_rxq *rxq = &trans_pcie->rxq[i]; iwl_pcie_rxq_check_wrptr() local 221 iwl_pcie_restock_bd(struct iwl_trans * trans,struct iwl_rxq * rxq,struct iwl_rx_mem_buffer * rxb) iwl_pcie_restock_bd() argument [all...] |
H A D | internal.h | 192 * @rxq: the rxq to get the rb stts from 195 struct iwl_rxq *rxq) in iwl_get_closed_rb_stts() 198 __le16 *rb_stts = rxq->rb_stts; in iwl_get_closed_rb_stts() 202 struct iwl_rb_status *rb_stts = rxq->rb_stts; in iwl_get_closed_rb_stts() 342 * @rxq: all the RX queue data 424 struct iwl_rxq *rxq; 578 void iwl_pcie_rx_init_rxb_lists(struct iwl_rxq *rxq); in iwl_pcie_get_num_sections() 581 struct iwl_rxq *rxq); in iwl_pcie_get_num_sections() 194 iwl_get_closed_rb_stts(struct iwl_trans * trans,struct iwl_rxq * rxq) iwl_get_closed_rb_stts() argument 352 struct iwl_rxq *rxq; global() member
|
H A D | ctxt-info.c | 215 rx_cfg->free_rbd_addr = cpu_to_le64(trans_pcie->rxq->bd_dma); in iwl_pcie_ctxt_info_init() 216 rx_cfg->used_rbd_addr = cpu_to_le64(trans_pcie->rxq->used_bd_dma); in iwl_pcie_ctxt_info_init() 217 rx_cfg->status_wr_ptr = cpu_to_le64(trans_pcie->rxq->rb_stts_dma); in iwl_pcie_ctxt_info_init()
|
/freebsd/sys/dev/xen/netfront/ |
H A D | netfront.c | 221 struct netfront_rxq *rxq; member 294 xn_get_rx_mbuf(struct netfront_rxq *rxq, RING_IDX ri) in xn_get_rx_mbuf() argument 300 m = rxq->mbufs[i]; in xn_get_rx_mbuf() 301 rxq->mbufs[i] = NULL; in xn_get_rx_mbuf() 306 xn_get_rx_ref(struct netfront_rxq *rxq, RING_IDX ri) in xn_get_rx_ref() argument 309 grant_ref_t ref = rxq->grant_ref[i]; in xn_get_rx_ref() 312 rxq->grant_ref[i] = GRANT_REF_INVALID; in xn_get_rx_ref() 479 XN_RX_LOCK(&np->rxq[i]); in netfront_suspend() 484 XN_RX_UNLOCK(&np->rxq[i]); in netfront_suspend() 504 XN_RX_LOCK(&info->rxq[i]); in netfront_resume() [all …]
|
/freebsd/sys/dev/netmap/ |
H A D | if_vtnet_netmap.h | 164 struct vtnet_rxq *rxq = &sc->vtnet_rxqs[ring_nr]; in vtnet_netmap_kring_refill() local 165 struct virtqueue *vq = rxq->vtnrx_vq; in vtnet_netmap_kring_refill() 171 for (nm_i = rxq->vtnrx_nm_refill; num > 0; in vtnet_netmap_kring_refill() 186 err = sglist_append(&sg, &rxq->vtnrx_shrhdr, sc->vtnet_hdr_size); in vtnet_netmap_kring_refill() 192 err = virtqueue_enqueue(vq, /*cookie=*/rxq, &sg, in vtnet_netmap_kring_refill() 200 rxq->vtnrx_nm_refill = nm_i; in vtnet_netmap_kring_refill() 212 vtnet_netmap_rxq_populate(struct vtnet_rxq *rxq) in vtnet_netmap_rxq_populate() argument 214 struct netmap_adapter *na = NA(rxq->vtnrx_sc->vtnet_ifp); in vtnet_netmap_rxq_populate() 220 slot = netmap_reset(na, NR_RX, rxq->vtnrx_id, 0); in vtnet_netmap_rxq_populate() 223 kring = na->rx_rings[rxq->vtnrx_id]; in vtnet_netmap_rxq_populate() [all …]
|
/freebsd/sys/dev/virtio/network/ |
H A D | if_vtnet.c | 148 static void vtnet_rx_vq_process(struct vtnet_rxq *rxq, int tries); 821 struct vtnet_rxq *rxq; in vtnet_init_rxq() local 823 rxq = &sc->vtnet_rxqs[id]; in vtnet_init_rxq() 825 snprintf(rxq->vtnrx_name, sizeof(rxq->vtnrx_name), "%s-rx%d", in vtnet_init_rxq() 827 mtx_init(&rxq->vtnrx_mtx, rxq->vtnrx_name, NULL, MTX_DEF); in vtnet_init_rxq() 829 rxq->vtnrx_sc = sc; in vtnet_init_rxq() 830 rxq->vtnrx_id = id; in vtnet_init_rxq() 832 rxq->vtnrx_sg = sglist_alloc(sc->vtnet_rx_nsegs, M_NOWAIT); in vtnet_init_rxq() 833 if (rxq->vtnrx_sg == NULL) in vtnet_init_rxq() 838 if (tcp_lro_init_args(&rxq->vtnrx_lro, sc->vtnet_ifp, in vtnet_init_rxq() [all …]
|
/freebsd/sys/dev/mana/ |
H A D | mana_en.c | 410 mana_load_rx_mbuf(struct mana_port_context *apc, struct mana_rxq *rxq, in mana_load_rx_mbuf() argument 419 mbuf = m_getjcl(M_NOWAIT, MT_DATA, M_PKTHDR, rxq->datasize); in mana_load_rx_mbuf() 427 mlen = rxq->datasize; in mana_load_rx_mbuf() 446 counter_u64_add(rxq->stats.dma_mapping_err, 1); in mana_load_rx_mbuf() 467 mana_unload_rx_mbuf(struct mana_port_context *apc, struct mana_rxq *rxq, in mana_unload_rx_mbuf() argument 1340 mana_fence_rq(struct mana_port_context *apc, struct mana_rxq *rxq) in mana_fence_rq() argument 1346 init_completion(&rxq->fence_event); in mana_fence_rq() 1350 req.wq_obj_handle = rxq->rxobj; in mana_fence_rq() 1356 rxq->rxq_idx, err); in mana_fence_rq() 1363 rxq->rxq_idx, err, resp.hdr.status); in mana_fence_rq() [all …]
|
H A D | mana_sysctl.c | 54 struct mana_rxq *rxq; in mana_sysctl_rx_stat_agg_u64() local 59 rxq = apc->rxqs[i]; in mana_sysctl_rx_stat_agg_u64() 60 stat += *((uint64_t *)((uint8_t *)rxq + offset)); in mana_sysctl_rx_stat_agg_u64() 68 rxq = apc->rxqs[i]; in mana_sysctl_rx_stat_agg_u64() 69 *((uint64_t *)((uint8_t *)rxq + offset)) = 0; in mana_sysctl_rx_stat_agg_u64() 79 struct mana_rxq *rxq; in mana_sysctl_rx_stat_u16() local 83 rxq = apc->rxqs[0]; in mana_sysctl_rx_stat_u16() 84 val = *((uint16_t *)((uint8_t *)rxq + offset)); in mana_sysctl_rx_stat_u16() 99 struct mana_rxq *rxq; in mana_sysctl_rx_stat_u32() local 103 rxq = apc->rxqs[0]; in mana_sysctl_rx_stat_u32() [all …]
|
H A D | hw_channel.c | 173 hwc->rxq->gdma_wq->id = val; in mana_hwc_init_event_handler() 201 hwc->rxq->msg_buf->gpa_mkey = val; in mana_hwc_init_event_handler() 223 struct hwc_wq *hwc_rxq = hwc->rxq; in mana_hwc_rx_event_handler() 646 struct hwc_wq *hwc_rxq = hwc->rxq; in mana_hwc_test_channel() 677 struct gdma_queue *rq = hwc->rxq->gdma_wq; in mana_hwc_establish_channel() 738 hwc->cq, &hwc->rxq); in mana_hwc_init_queues() 841 if (hwc->rxq) in mana_hwc_destroy_channel() 842 mana_hwc_destroy_wq(hwc, hwc->rxq); in mana_hwc_destroy_channel()
|
/freebsd/sys/dev/ice/ |
H A D | ice_iflib_txrx.c | 52 static int _ice_ift_rxd_available(struct ice_rx_queue *rxq, qidx_t pidx, qidx_t budget); 53 static int _ice_ift_rxd_pkt_get(struct ice_rx_queue *rxq, if_rxd_info_t ri); 54 static void _ice_ift_rxd_refill(struct ice_rx_queue *rxq, uint32_t pidx, 56 static void _ice_ift_rxd_flush(struct ice_softc *sc, struct ice_rx_queue *rxq, 315 * @rxq: RX queue driver structure in ice_ift_rxd_pkt_get() 323 _ice_ift_rxd_available(struct ice_rx_queue *rxq, qidx_t pidx, qidx_t budget) in ice_ift_rxd_pkt_get() 329 nrxd = rxq->desc_count; in ice_ift_rxd_pkt_get() 332 rxd = &rxq->rx_base[i]; in ice_ift_rxd_pkt_get() 360 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid]; 362 return _ice_ift_rxd_available(rxq, pid 251 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid]; ice_ift_rxd_available() local 287 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[ri->iri_qsidx]; ice_ift_rxd_pkt_get() local 367 struct ice_rx_queue *rxq; ice_ift_rxd_refill() local 403 struct ice_rx_queue *rxq = &sc->pf_vsi.rx_queues[rxqid]; ice_ift_rxd_flush() local [all...] |
H A D | ice_common_txrx.h | 261 * @rxq: the Rx queue structure 272 ice_rx_checksum(struct ice_rx_queue *rxq, uint32_t *flags, uint32_t *data, in ice_rx_checksum() argument 326 rxq->stats.cso[ICE_CSO_STAT_RX_IP6_ERR]++; in ice_rx_checksum() 337 rxq->stats.cso[ICE_CSO_STAT_RX_IP4_ERR]++; in ice_rx_checksum() 341 rxq->stats.cso[ICE_CSO_STAT_RX_L3_ERR]++; in ice_rx_checksum() 352 rxq->stats.cso[ICE_CSO_STAT_RX_TCP_ERR]++; in ice_rx_checksum() 356 rxq->stats.cso[ICE_CSO_STAT_RX_UDP_ERR]++; in ice_rx_checksum() 360 rxq->stats.cso[ICE_CSO_STAT_RX_SCTP_ERR]++; in ice_rx_checksum() 368 rxq->stats.cso[ICE_CSO_STAT_RX_L4_ERR]++; in ice_rx_checksum()
|
H A D | if_ice_iflib.c | 1122 struct ice_rx_queue *rxq; in ice_if_rx_queues_alloc() local 1152 for (i = 0, rxq = vsi->rx_queues; i < nrxqsets; i++, rxq++) { in ice_if_rx_queues_alloc() 1153 rxq->me = i; in ice_if_rx_queues_alloc() 1154 rxq->vsi = vsi; in ice_if_rx_queues_alloc() 1157 rxq->desc_count = sc->scctx->isc_nrxd[0]; in ice_if_rx_queues_alloc() 1160 rxq->tail = QRX_TAIL(vsi->rx_qmap[i]); in ice_if_rx_queues_alloc() 1161 rxq->rx_base = (union ice_32b_rx_flex_desc *)vaddrs[i]; in ice_if_rx_queues_alloc() 1162 rxq->rx_paddr = paddrs[i]; in ice_if_rx_queues_alloc() 1164 ice_add_rxq_sysctls(rxq); in ice_if_rx_queues_alloc() 1243 struct ice_rx_queue __unused *rxq = (struct ice_rx_queue *)arg; in ice_msix_que() local [all …]
|
/freebsd/sys/dev/qlnx/qlnxe/ |
H A D | qlnx_os.c | 133 static int qlnx_alloc_rx_buffer(qlnx_host_t *ha, struct qlnx_rx_queue *rxq); 134 static void qlnx_reuse_rx_data(struct qlnx_rx_queue *rxq); 136 struct qlnx_rx_queue *rxq); 1462 if (fp->rxq->handle != NULL) { in qlnx_set_rx_coalesce() 1464 0, fp->rxq->handle); in qlnx_set_rx_coalesce() 3851 struct qlnx_rx_queue *rxq; in qlnx_rx_jumbo_chain() local 3854 rxq = fp->rxq; in qlnx_rx_jumbo_chain() 3858 rxq->sw_rx_cons = (rxq->sw_rx_cons + 1) & (RX_RING_SIZE - 1); in qlnx_rx_jumbo_chain() 3860 sw_rx_data = &rxq->sw_rx_ring[rxq->sw_rx_cons]; in qlnx_rx_jumbo_chain() 3866 rxq->sw_rx_cons = in qlnx_rx_jumbo_chain() [all …]
|
/freebsd/sys/dev/vmware/vmxnet3/ |
H A D | if_vmx.c | 470 struct vmxnet3_rxqueue *rxq; in vmxnet3_msix_intr_assign() local 481 rxq = &sc->vmx_rxq[i]; in vmxnet3_msix_intr_assign() 482 error = iflib_irq_alloc_generic(ctx, &rxq->vxrxq_irq, i + 1, in vmxnet3_msix_intr_assign() 483 IFLIB_INTR_RXTX, vmxnet3_rxq_intr, rxq, i, irq_name); in vmxnet3_msix_intr_assign() 527 struct vmxnet3_rxqueue *rxq; in vmxnet3_free_irqs() local 533 rxq = &sc->vmx_rxq[i]; in vmxnet3_free_irqs() 534 iflib_irq_free(sc->vmx_ctx, &rxq->vxrxq_irq); in vmxnet3_free_irqs() 689 struct vmxnet3_rxqueue *rxq; in vmxnet3_set_interrupt_idx() local 711 rxq = &sc->vmx_rxq[i]; in vmxnet3_set_interrupt_idx() 712 rxs = rxq->vxrxq_rs; in vmxnet3_set_interrupt_idx() [all …]
|
/freebsd/sys/net/ |
H A D | iflib.c | 141 static void iru_init(if_rxd_update_t iru, iflib_rxq_t rxq, uint8_t flid); 701 static int iflib_rxd_avail(if_ctx_t ctx, iflib_rxq_t rxq, qidx_t cidx, qidx_t budget); 768 static int netmap_fl_refill(iflib_rxq_t rxq, struct netmap_kring *kring, bool init); 842 iflib_rxq_t rxq = &ctx->ifc_rxqs[0]; in iflib_netmap_config() local 843 iflib_fl_t fl = &rxq->ifr_fl[0]; in iflib_netmap_config() 858 netmap_fl_refill(iflib_rxq_t rxq, struct netmap_kring *kring, bool init) in netmap_fl_refill() argument 865 if_ctx_t ctx = rxq->ifr_ctx; in netmap_fl_refill() 866 iflib_fl_t fl = &rxq->ifr_fl[0]; in netmap_fl_refill() 900 iru_init(&iru, rxq, 0 /* flid */); in netmap_fl_refill() 964 ctx->isc_rxd_flush(ctx->ifc_softc, rxq->ifr_id, fl->ifl_id, in netmap_fl_refill() [all …]
|
/freebsd/sys/dev/iavf/ |
H A D | iavf_vc_iflib.c | 102 vqpi->rxq.vsi_id = vqci->vsi_id; in iavf_configure_queues() 103 vqpi->rxq.queue_id = i; in iavf_configure_queues() 104 vqpi->rxq.ring_len = scctx->isc_nrxd[0]; in iavf_configure_queues() 105 vqpi->rxq.dma_ring_addr = rxr->rx_paddr; in iavf_configure_queues() 106 vqpi->rxq.max_pkt_size = scctx->isc_max_frame_size; in iavf_configure_queues() 107 vqpi->rxq.databuffer_size = rxr->mbuf_sz; in iavf_configure_queues() 108 vqpi->rxq.splithdr_enabled = 0; in iavf_configure_queues()
|
/freebsd/sys/dev/gve/ |
H A D | gve_sysctl.c | 50 struct sysctl_oid_list *child, struct gve_rx_ring *rxq) in gve_setup_rxq_sysctl() argument 57 snprintf(namebuf, sizeof(namebuf), "rxq%d", rxq->com.id); in gve_setup_rxq_sysctl() 62 stats = &rxq->stats; in gve_setup_rxq_sysctl() 104 &rxq->cnt, 0, "Number of descriptors completed"); in gve_setup_rxq_sysctl() 107 &rxq->fill_cnt, rxq->fill_cnt, in gve_setup_rxq_sysctl()
|
/freebsd/sys/dev/ral/ |
H A D | rt2560.c | 252 error = rt2560_alloc_rx_ring(sc, &sc->rxq, RT2560_RX_RING_COUNT); in rt2560_attach() 352 rt2560_free_rx_ring(sc, &sc->rxq); in rt2560_detach() 649 desc = &sc->rxq.desc[i]; in rt2560_alloc_rx_ring() 650 data = &sc->rxq.data[i]; in rt2560_alloc_rx_ring() 1097 hw = RAL_READ(sc, RT2560_SECCSR0) - sc->rxq.physaddr; in rt2560_decryption_intr() 1100 bus_dmamap_sync(sc->rxq.desc_dmat, sc->rxq.desc_map, in rt2560_decryption_intr() 1103 for (; sc->rxq.cur_decrypt != hw;) { in rt2560_decryption_intr() 1104 desc = &sc->rxq.desc[sc->rxq.cur_decrypt]; in rt2560_decryption_intr() 1105 data = &sc->rxq.data[sc->rxq.cur_decrypt]; in rt2560_decryption_intr() 1135 bus_dmamap_sync(sc->rxq.data_dmat, data->map, in rt2560_decryption_intr() [all …]
|
H A D | rt2661.c | 255 error = rt2661_alloc_rx_ring(sc, &sc->rxq, RT2661_RX_RING_COUNT); in rt2661_attach() 345 rt2661_free_rx_ring(sc, &sc->rxq); in rt2661_detach() 656 desc = &sc->rxq.desc[i]; in rt2661_alloc_rx_ring() 657 data = &sc->rxq.data[i]; in rt2661_alloc_rx_ring() 967 bus_dmamap_sync(sc->rxq.desc_dmat, sc->rxq.desc_map, in rt2661_rx_intr() 973 desc = &sc->rxq.desc[sc->rxq.cur]; in rt2661_rx_intr() 974 data = &sc->rxq.data[sc->rxq.cur]; in rt2661_rx_intr() 1009 bus_dmamap_sync(sc->rxq.data_dmat, data->map, in rt2661_rx_intr() 1011 bus_dmamap_unload(sc->rxq.data_dmat, data->map); in rt2661_rx_intr() 1013 error = bus_dmamap_load(sc->rxq.data_dmat, data->map, in rt2661_rx_intr() [all …]
|
/freebsd/sys/dev/ixl/ |
H A D | ixl_pf_iov.c | 605 struct i40e_hmc_obj_rxq rxq; in ixl_vf_config_rx_queue() local 611 bzero(&rxq, sizeof(rxq)); in ixl_vf_config_rx_queue() 627 rxq.hsplit_0 = info->rx_split_pos & in ixl_vf_config_rx_queue() 632 rxq.hbuff = info->hdr_size >> I40E_RXQ_CTX_HBUFF_SHIFT; in ixl_vf_config_rx_queue() 634 rxq.dtype = 2; in ixl_vf_config_rx_queue() 641 rxq.base = info->dma_ring_addr / IXL_RX_CTX_BASE_UNITS; in ixl_vf_config_rx_queue() 642 rxq.qlen = info->ring_len; in ixl_vf_config_rx_queue() 644 rxq.dbuff = info->databuffer_size >> I40E_RXQ_CTX_DBUFF_SHIFT; in ixl_vf_config_rx_queue() 646 rxq.dsize = 1; in ixl_vf_config_rx_queue() 647 rxq.crcstrip = 1; in ixl_vf_config_rx_queue() [all …]
|
/freebsd/sys/dev/nfe/ |
H A D | if_nfe.c | 577 if ((error = nfe_alloc_rx_ring(sc, &sc->rxq)) != 0) in nfe_attach() 762 nfe_free_rx_ring(sc, &sc->rxq); in nfe_detach() 1162 data = &sc->rxq.data[i]; in nfe_alloc_rx_ring() 1914 data = &sc->rxq.data[idx]; in nfe_discard_rxbuf() 1918 desc64 = &sc->rxq.desc64[idx]; in nfe_discard_rxbuf() 1925 desc32 = &sc->rxq.desc32[idx]; in nfe_discard_rxbuf() 1974 if (bus_dmamap_load_mbuf_sg(sc->rxq.rx_data_tag, sc->rxq.rx_spare_map, in nfe_newbuf() 1981 data = &sc->rxq.data[idx]; in nfe_newbuf() 1983 bus_dmamap_sync(sc->rxq.rx_data_tag, data->rx_data_map, in nfe_newbuf() 1985 bus_dmamap_unload(sc->rxq.rx_data_tag, data->rx_data_map); in nfe_newbuf() [all …]
|
/freebsd/sys/dev/cxgbe/ |
H A D | t4_sge.c | 1109 struct sge_rxq *rxq; in t4_setup_vi_queues() local 1166 for_each_rxq(vi, i, rxq) { in t4_setup_vi_queues() 1167 rc = alloc_rxq(vi, rxq, i, intr_idx, maxp); in t4_setup_vi_queues() 1217 struct sge_rxq *rxq; in t4_teardown_vi_queues() local 1260 for_each_rxq(vi, i, rxq) { in t4_teardown_vi_queues() 1261 free_rxq(vi, rxq); in t4_teardown_vi_queues() 1374 MPASS(irq->rxq != NULL); in t4_vi_intr() 1375 t4_intr(irq->rxq); in t4_vi_intr() 1572 struct sge_rxq *rxq = iq_to_rxq(iq); in service_iq_fl() local 1583 struct lro_ctrl *lro = &rxq->lro; in service_iq_fl() [all …]
|