Home
last modified time | relevance | path

Searched refs:rx_ring (Results 1 – 25 of 83) sorted by relevance

1234

/freebsd/sys/dev/rtwn/pci/
H A Drtwn_pci_attach.c125 struct rtwn_rx_ring *rx_ring = &pc->rx_ring; in rtwn_pci_alloc_rx_list() local
134 size, 1, size, 0, NULL, NULL, &rx_ring->desc_dmat); in rtwn_pci_alloc_rx_list()
140 error = bus_dmamem_alloc(rx_ring->desc_dmat, (void **)&rx_ring->desc, in rtwn_pci_alloc_rx_list()
142 &rx_ring->desc_map); in rtwn_pci_alloc_rx_list()
147 error = bus_dmamap_load(rx_ring->desc_dmat, rx_ring->desc_map, in rtwn_pci_alloc_rx_list()
148 rx_ring->desc, size, rtwn_pci_dma_map_addr, &rx_ring->paddr, 0); in rtwn_pci_alloc_rx_list()
153 bus_dmamap_sync(rx_ring->desc_dmat, rx_ring->desc_map, in rtwn_pci_alloc_rx_list()
159 MJUMPAGESIZE, 1, MJUMPAGESIZE, 0, NULL, NULL, &rx_ring->data_dmat); in rtwn_pci_alloc_rx_list()
167 rx_data = &rx_ring->rx_data[i]; in rtwn_pci_alloc_rx_list()
168 error = bus_dmamap_create(rx_ring->data_dmat, 0, &rx_data->map); in rtwn_pci_alloc_rx_list()
[all …]
/freebsd/sys/dev/ena/
H A Dena_datapath.c74 struct ena_ring *rx_ring; in ena_cleanup() local
81 rx_ring = que->rx_ring; in ena_cleanup()
99 atomic_store_8(&rx_ring->first_interrupt, 1); in ena_cleanup()
102 rxc = ena_rx_cleanup(rx_ring); in ena_cleanup()
346 ena_rx_hash_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx, in ena_rx_hash_mbuf() argument
349 struct ena_adapter *adapter = rx_ring->adapter; in ena_rx_hash_mbuf()
405 mbuf->m_pkthdr.flowid = rx_ring->qid; in ena_rx_hash_mbuf()
419 ena_rx_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_buf_info *ena_bufs, in ena_rx_mbuf() argument
430 adapter = rx_ring->adapter; in ena_rx_mbuf()
435 rx_info = &rx_ring->rx_buffer_info[req_id]; in ena_rx_mbuf()
[all …]
H A Dena.c400 rxr = &adapter->rx_ring[i]; in ena_init_io_rings_basic()
414 que->rx_ring = rxr; in ena_init_io_rings_basic()
432 rxr = &adapter->rx_ring[i]; in ena_init_io_rings_advanced()
475 struct ena_ring *rxr = &adapter->rx_ring[qid]; in ena_free_io_ring_resources()
873 struct ena_ring *rx_ring = que->rx_ring; in ena_setup_rx_resources() local
876 size = sizeof(struct ena_rx_buffer) * rx_ring->ring_size; in ena_setup_rx_resources()
880 rx_ring->initialized = false; in ena_setup_rx_resources()
889 rx_ring->rx_buffer_info = malloc(size, M_DEVBUF, M_WAITOK | M_ZERO); in ena_setup_rx_resources()
891 size = sizeof(uint16_t) * rx_ring->ring_size; in ena_setup_rx_resources()
892 rx_ring->free_rx_ids = malloc(size, M_DEVBUF, M_WAITOK); in ena_setup_rx_resources()
[all …]
H A Dena_netmap.h51 struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info);
53 struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info);
/freebsd/sys/ofed/drivers/infiniband/ulp/sdp/
H A Dsdp_rx.c109 int id = ring_head(ssk->rx_ring); in sdp_post_recv()
125 rx_req = ssk->rx_ring.buffer + (id & (SDP_RX_SIZE - 1)); in sdp_post_recv()
156 atomic_inc(&ssk->rx_ring.head); in sdp_post_recv()
286 ssk->recv_request_head = ring_head(ssk->rx_ring) + 1; in sdp_handle_resize_request()
288 ssk->recv_request_head = ring_tail(ssk->rx_ring); in sdp_handle_resize_request()
308 if (unlikely(id != ring_tail(ssk->rx_ring))) { in sdp_recv_completion()
310 id, ring_tail(ssk->rx_ring)); in sdp_recv_completion()
315 rx_req = &ssk->rx_ring.buffer[id & (SDP_RX_SIZE - 1)]; in sdp_recv_completion()
319 atomic_inc(&ssk->rx_ring.tail); in sdp_recv_completion()
534 struct ib_cq *cq = ssk->rx_ring.cq; in sdp_poll_rx_cq()
[all …]
H A Dsdp.h148 #define rx_ring_posted(ssk) ring_posted(ssk->rx_ring)
359 struct sdp_rx_ring rx_ring; member
423 static inline void rx_ring_unlock(struct sdp_rx_ring *rx_ring) in rx_ring_unlock() argument
425 rw_runlock(&rx_ring->destroyed_lock); in rx_ring_unlock()
428 static inline int rx_ring_trylock(struct sdp_rx_ring *rx_ring) in rx_ring_trylock() argument
430 rw_rlock(&rx_ring->destroyed_lock); in rx_ring_trylock()
431 if (rx_ring->destroyed) { in rx_ring_trylock()
432 rx_ring_unlock(rx_ring); in rx_ring_trylock()
438 static inline void rx_ring_destroy_lock(struct sdp_rx_ring *rx_ring) in rx_ring_destroy_lock() argument
440 rw_wlock(&rx_ring->destroyed_lock); in rx_ring_destroy_lock()
[all …]
/freebsd/sys/dev/al_eth/
H A Dal_eth.c1000 struct al_eth_ring *ring = &adapter->rx_ring[i]; in al_eth_init_rings()
1041 struct al_eth_ring *rx_ring, in al_eth_alloc_rx_buf() argument
1067 error = bus_dmamap_load_mbuf_sg(rx_ring->dma_buf_tag, rx_info->dma_map, in al_eth_alloc_rx_buf()
1070 device_printf(rx_ring->dev, "failed to map mbuf, error = %d\n", in al_eth_alloc_rx_buf()
1088 struct al_eth_ring *rx_ring = &adapter->rx_ring[qid]; in al_eth_refill_rx_bufs() local
1092 next_to_use = rx_ring->next_to_use; in al_eth_refill_rx_bufs()
1097 &rx_ring->rx_buffer_info[next_to_use]; in al_eth_refill_rx_bufs()
1100 rx_ring, rx_info) < 0)) { in al_eth_refill_rx_bufs()
1106 rc = al_eth_rx_buffer_add(rx_ring->dma_q, in al_eth_refill_rx_bufs()
1114 next_to_use = AL_ETH_RX_RING_IDX_NEXT(rx_ring, next_to_use); in al_eth_refill_rx_bufs()
[all …]
H A Dal_eth.h152 #define AL_ETH_RX_RING_IDX_NEXT(rx_ring, idx) (((idx) + 1) & (AL_ETH_DEFAULT_RX_DESCS - 1)) argument
153 #define AL_ETH_RX_RING_IDX_ADD(rx_ring, idx, n) (((idx) + (n)) & (AL_ETH_DEFAULT_RX_DESCS - 1)) argument
277 struct al_eth_ring rx_ring[AL_ETH_NUM_QUEUES]; member
/freebsd/sys/dev/qlxge/
H A Dqls_hw.c928 ha->rx_ring[0].cq_db_offset)); in qls_init_hw_if()
934 Q81_WR_LBQ_PROD_IDX(i, ha->rx_ring[i].lbq_in); in qls_init_hw_if()
935 Q81_WR_SBQ_PROD_IDX(i, ha->rx_ring[i].sbq_in); in qls_init_hw_if()
1049 rxr = &ha->rx_ring[cid]; in qls_init_comp_queue()
1538 if (ha->rx_ring[ridx].flags.cq_dma) { in qls_free_rx_ring_dma()
1539 qls_free_dmabuf(ha, &ha->rx_ring[ridx].cq_dma); in qls_free_rx_ring_dma()
1540 ha->rx_ring[ridx].flags.cq_dma = 0; in qls_free_rx_ring_dma()
1543 if (ha->rx_ring[ridx].flags.lbq_dma) { in qls_free_rx_ring_dma()
1544 qls_free_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma); in qls_free_rx_ring_dma()
1545 ha->rx_ring[ridx].flags.lbq_dma = 0; in qls_free_rx_ring_dma()
[all …]
H A Dqls_dbg.c274 qls_dump_buf32(ha, "cq_icb", ha->rx_ring[0].cq_icb_vaddr, in qls_dump_cq()
278 (void *)ha->rx_ring[0].lbq_addr_tbl_paddr); in qls_dump_cq()
280 qls_dump_buf32(ha, "lbq_addr_tbl", ha->rx_ring[0].lbq_addr_tbl_vaddr, in qls_dump_cq()
284 (void *)ha->rx_ring[0].lbq_paddr); in qls_dump_cq()
286 qls_dump_buf32(ha, "lbq", ha->rx_ring[0].lbq_vaddr, in qls_dump_cq()
290 (void *)ha->rx_ring[0].sbq_addr_tbl_paddr); in qls_dump_cq()
292 qls_dump_buf32(ha, "sbq_addr_tbl", ha->rx_ring[0].sbq_addr_tbl_vaddr, in qls_dump_cq()
296 (void *)ha->rx_ring[0].sbq_paddr); in qls_dump_cq()
298 qls_dump_buf32(ha, "sbq", ha->rx_ring[0].sbq_vaddr, in qls_dump_cq()
302 (void *)ha->rx_ring[0].lb_paddr); in qls_dump_cq()
H A Dqls_isr.c80 rxr = &ha->rx_ring[r_idx]; in qls_replenish_rx()
134 rxr = &ha->rx_ring[rxr_idx]; in qls_rx_comp()
235 struct lro_ctrl *lro = &ha->rx_ring[cq_idx].lro; in qls_cq_isr()
238 cq_b = ha->rx_ring[cq_idx].cq_base_vaddr; in qls_cq_isr()
240 cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr); in qls_cq_isr()
242 i = ha->rx_ring[cq_idx].cq_next; in qls_cq_isr()
263 __func__, i, (*(ha->rx_ring[cq_idx].cqi_vaddr)), in qls_cq_isr()
279 cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr); in qls_cq_isr()
292 ha->rx_ring[cq_idx].cq_next = cq_comp_idx; in qls_cq_isr()
295 Q81_WR_CQ_CONS_IDX(cq_idx, (ha->rx_ring[cq_idx].cq_next)); in qls_cq_isr()
H A Dqls_os.c203 (void *)ha->rx_ring[i].rx_int)); in qls_sysctl_get_drvr_stats()
208 (void *)ha->rx_ring[i].rss_int)); in qls_sysctl_get_drvr_stats()
213 ha->rx_ring[i].lbq_next); in qls_sysctl_get_drvr_stats()
218 ha->rx_ring[i].lbq_free); in qls_sysctl_get_drvr_stats()
223 ha->rx_ring[i].lbq_in); in qls_sysctl_get_drvr_stats()
228 ha->rx_ring[i].sbq_next); in qls_sysctl_get_drvr_stats()
233 ha->rx_ring[i].sbq_free); in qls_sysctl_get_drvr_stats()
238 ha->rx_ring[i].sbq_in); in qls_sysctl_get_drvr_stats()
1259 qla_rx_ring_t *rx_ring; in qls_alloc_rcv_mbufs() local
1262 rx_ring = &ha->rx_ring[r]; in qls_alloc_rcv_mbufs()
[all …]
H A Dqls_hw.h934 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG), idx)
937 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG))
940 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\
944 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\
948 (ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG), idx)
951 (ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG))
954 (ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG), idx)
957 (ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG))
/freebsd/sys/contrib/dev/rtw89/
H A Dpci.c154 struct rtw89_pci_rx_ring *rx_ring) in rtw89_pci_sync_skb_for_device()
156 struct rtw89_pci_dma_ring *bd_ring = &rx_ring->bd_ring; in rtw89_pci_sync_skb_for_device()
207 struct rtw89_pci_rx_ring *rx_ring,
218 if (rx_ring->target_rx_tag == 0) in rtw89_skb_put_rx_data()
221 target_rx_tag = rx_ring->target_rx_tag; in rtw89_skb_put_rx_data()
234 struct rtw89_pci_rx_ring *rx_ring, in rtw89_skb_put_rx_data()
245 ret = rtw89_pci_validate_rx_tag(rtwdev, rx_ring, skb); in rtw89_pci_rxbd_deliver_skbs()
251 rx_ring->target_rx_tag = rx_info->tag + 1; in rtw89_pci_rxbd_deliver_skbs()
329 struct rtw89_pci_rx_ring *rx_ring) in rtw89_pci_rxbd_deliver()
331 struct rtw89_pci_dma_ring *bd_ring = &rx_ring in rtw89_pci_rxbd_deliver()
128 rtw89_pci_rxbd_recalc(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_rxbd_recalc() argument
237 rtw89_pci_rxbd_deliver_skbs(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_rxbd_deliver_skbs() argument
322 rtw89_pci_rxbd_deliver(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 cnt) rtw89_pci_rxbd_deliver() argument
347 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_poll_rxq_dma() local
524 rtw89_pci_release_tx_skbs(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 max_cnt) rtw89_pci_release_tx_skbs() argument
575 rtw89_pci_release_tx(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 cnt) rtw89_pci_release_tx() argument
600 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_poll_rpq_dma() local
627 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_isr_rxd_unavail() local
982 struct rtw89_pci_rx_ring *rx_ring; __rtw89_pci_check_and_reclaim_tx_resource() local
1437 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_reset_trx_rings() local
1576 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_switch_bd_idx_addr() local
2888 rtw89_pci_free_rx_ring(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_free_rx_ring() argument
2922 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_free_rx_rings() local
2939 rtw89_pci_init_rx_bd(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring,struct sk_buff * skb,int buf_sz,u32 idx) rtw89_pci_init_rx_bd() argument
3106 rtw89_pci_alloc_rx_ring(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring,u32 desc_size,u32 len,u32 rxch) rtw89_pci_alloc_rx_ring() argument
3191 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_alloc_rx_rings() local
[all...]
/freebsd/sys/contrib/dev/athk/ath10k/
H A Dhtt_rx.c34 hash_for_each_possible(ar->htt.rx_ring.skb_table, rxcb, hlist, paddr) in ath10k_htt_rx_find_skb_paddr()
49 if (htt->rx_ring.in_ord_rx) { in ath10k_htt_rx_ring_free()
50 hash_for_each_safe(htt->rx_ring.skb_table, i, n, rxcb, hlist) { in ath10k_htt_rx_ring_free()
59 for (i = 0; i < htt->rx_ring.size; i++) { in ath10k_htt_rx_ring_free()
60 skb = htt->rx_ring.netbufs_ring[i]; in ath10k_htt_rx_ring_free()
72 htt->rx_ring.fill_cnt = 0; in ath10k_htt_rx_ring_free()
73 hash_init(htt->rx_ring.skb_table); in ath10k_htt_rx_ring_free()
74 memset(htt->rx_ring.netbufs_ring, 0, in ath10k_htt_rx_ring_free()
75 htt->rx_ring.size * sizeof(htt->rx_ring.netbufs_ring[0])); in ath10k_htt_rx_ring_free()
80 return htt->rx_ring.size * sizeof(htt->rx_ring.paddrs_ring_32); in ath10k_htt_get_rx_ring_size_32()
[all …]
/freebsd/sys/dev/et/
H A Dif_et.c860 struct et_rxdesc_ring *rx_ring; in et_dma_alloc() local
894 rx_ring = &sc->sc_rx_ring[0]; in et_dma_alloc()
896 &rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap, in et_dma_alloc()
897 &rx_ring->rr_paddr, "RX ring 0"); in et_dma_alloc()
898 rx_ring->rr_posreg = ET_RX_RING0_POS; in et_dma_alloc()
903 rx_ring = &sc->sc_rx_ring[1]; in et_dma_alloc()
905 &rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap, in et_dma_alloc()
906 &rx_ring->rr_paddr, "RX ring 1"); in et_dma_alloc()
907 rx_ring->rr_posreg = ET_RX_RING1_POS; in et_dma_alloc()
1039 struct et_rxdesc_ring *rx_ring; in et_dma_free() local
[all …]
/freebsd/sys/contrib/dev/athk/ath11k/
H A Ddp_rx.c344 struct dp_rxdma_ring *rx_ring, in ath11k_dp_rxbufs_replenish() argument
357 req_entries = min(req_entries, rx_ring->bufs_max); in ath11k_dp_rxbufs_replenish()
359 srng = &ab->hal.srng_list[rx_ring->refill_buf_ring.ring_id]; in ath11k_dp_rxbufs_replenish()
366 if (!req_entries && (num_free > (rx_ring->bufs_max * 3) / 4)) in ath11k_dp_rxbufs_replenish()
391 spin_lock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish()
392 buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 1, in ath11k_dp_rxbufs_replenish()
393 (rx_ring->bufs_max * 3) + 1, GFP_ATOMIC); in ath11k_dp_rxbufs_replenish()
394 spin_unlock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish()
419 spin_lock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish()
420 idr_remove(&rx_ring->bufs_idr, buf_id); in ath11k_dp_rxbufs_replenish()
[all …]
H A Dspectral.c833 ret = ath11k_dbring_srng_setup(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc()
840 ath11k_dbring_set_cfg(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc()
845 ret = ath11k_dbring_buf_setup(ar, &sp->rx_ring, db_cap); in ath11k_spectral_ring_alloc()
851 ret = ath11k_dbring_wmi_cfg_setup(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc()
861 ath11k_dbring_buf_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_alloc()
863 ath11k_dbring_srng_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_alloc()
871 ath11k_dbring_srng_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_free()
872 ath11k_dbring_buf_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_free()
1015 idr_init(&sp->rx_ring.bufs_idr); in ath11k_spectral_init()
1016 spin_lock_init(&sp->rx_ring.idr_lock); in ath11k_spectral_init()
[all …]
/freebsd/sys/contrib/dev/rtw88/
H A Dpci.c174 struct rtw_pci_rx_ring *rx_ring) in rtw_pci_free_rx_ring_skbs() argument
182 for (i = 0; i < rx_ring->r.len; i++) { in rtw_pci_free_rx_ring_skbs()
183 skb = rx_ring->buf[i]; in rtw_pci_free_rx_ring_skbs()
190 rx_ring->buf[i] = NULL; in rtw_pci_free_rx_ring_skbs()
195 struct rtw_pci_rx_ring *rx_ring) in rtw_pci_free_rx_ring() argument
198 u8 *head = rx_ring->r.head; in rtw_pci_free_rx_ring()
199 int ring_sz = rx_ring->r.desc_size * rx_ring->r.len; in rtw_pci_free_rx_ring()
201 rtw_pci_free_rx_ring_skbs(rtwdev, rx_ring); in rtw_pci_free_rx_ring()
203 dma_free_coherent(&pdev->dev, ring_sz, head, rx_ring in rtw_pci_free_rx_ring()
210 struct rtw_pci_rx_ring *rx_ring; rtw_pci_free_trx_ring() local
256 rtw_pci_reset_rx_desc(struct rtw_dev * rtwdev,struct sk_buff * skb,struct rtw_pci_rx_ring * rx_ring,u32 idx,u32 desc_sz) rtw_pci_reset_rx_desc() argument
282 rtw_pci_sync_rx_desc_device(struct rtw_dev * rtwdev,dma_addr_t dma,struct rtw_pci_rx_ring * rx_ring,u32 idx,u32 desc_sz) rtw_pci_sync_rx_desc_device() argument
299 rtw_pci_init_rx_ring(struct rtw_dev * rtwdev,struct rtw_pci_rx_ring * rx_ring,u8 desc_size,u32 len) rtw_pci_init_rx_ring() argument
365 struct rtw_pci_rx_ring *rx_ring; rtw_pci_init_trx_ring() local
734 rtw_pci_dma_check(struct rtw_dev * rtwdev,struct rtw_pci_rx_ring * rx_ring,u32 idx) rtw_pci_dma_check() argument
[all...]
/freebsd/sys/dev/axgbe/
H A Dif_axgbe_pci.c809 channel->rx_ring = NULL; in axgbe_alloc_channels()
1756 struct xgbe_ring *rx_ring; in axgbe_if_rx_queues_alloc() local
1774 rx_ring = (struct xgbe_ring*)malloc(nrxqs * in axgbe_if_rx_queues_alloc()
1777 if (rx_ring == NULL) { in axgbe_if_rx_queues_alloc()
1782 channel->rx_ring = rx_ring; in axgbe_if_rx_queues_alloc()
1784 for (j = 0; j < nrxqs; j++, rx_ring++) { in axgbe_if_rx_queues_alloc()
1785 rx_ring->rdata = in axgbe_if_rx_queues_alloc()
1790 rx_ring->rdesc = (struct xgbe_ring_desc *)va[i*nrxqs + j]; in axgbe_if_rx_queues_alloc()
1791 rx_ring->rdesc_paddr = pa[i*nrxqs + j]; in axgbe_if_rx_queues_alloc()
1792 rx_ring->rdesc_count = scctx->isc_nrxd[j]; in axgbe_if_rx_queues_alloc()
[all …]
/freebsd/sys/ofed/drivers/infiniband/ulp/ipoib/
H A Dipoib_cm.c130 rx_req = (struct ipoib_rx_buf *)&rx->rx_ring[id]; in ipoib_cm_post_receive_nonsrq()
143 m_freem(rx->rx_ring[id].mb); in ipoib_cm_post_receive_nonsrq()
144 rx->rx_ring[id].mb = NULL; in ipoib_cm_post_receive_nonsrq()
158 struct ipoib_cm_rx_buf *rx_ring) in ipoib_cm_free_rx_ring() argument
163 if (rx_ring[i].mb) { in ipoib_cm_free_rx_ring()
164 ipoib_cm_dma_unmap_rx(priv, &rx_ring[i]); in ipoib_cm_free_rx_ring()
165 m_freem(rx_ring[i].mb); in ipoib_cm_free_rx_ring()
168 kfree(rx_ring); in ipoib_cm_free_rx_ring()
310 rx->rx_ring = kzalloc(ipoib_recvq_size * sizeof *rx->rx_ring, GFP_KERNEL); in ipoib_cm_nonsrq_init_rx()
311 if (!rx->rx_ring) { in ipoib_cm_nonsrq_init_rx()
[all …]
H A Dipoib_ib.c155 rx_req = &priv->rx_ring[id]; in ipoib_ib_post_receive()
166 ipoib_dma_unmap_rx(priv, &priv->rx_ring[id]); in ipoib_ib_post_receive()
167 m_freem(priv->rx_ring[id].mb); in ipoib_ib_post_receive()
168 priv->rx_ring[id].mb = NULL; in ipoib_ib_post_receive()
177 return ipoib_alloc_map_mb(priv, &priv->rx_ring[id], in ipoib_alloc_rx_mb()
217 mb = priv->rx_ring[wr_id].mb; in ipoib_ib_handle_rx_wc()
227 ipoib_dma_unmap_rx(priv, &priv->rx_ring[wr_id]); in ipoib_ib_handle_rx_wc()
229 priv->rx_ring[wr_id].mb = NULL; in ipoib_ib_handle_rx_wc()
241 memcpy(&saverx, &priv->rx_ring[wr_id], sizeof(saverx)); in ipoib_ib_handle_rx_wc()
247 memcpy(&priv->rx_ring[wr_id], &saverx, sizeof(saverx)); in ipoib_ib_handle_rx_wc()
[all …]
/freebsd/sys/dev/bnxt/bnxt_en/
H A Dbnxt_txrx.c252 struct bnxt_ring *rx_ring; in bnxt_isc_rxd_refill() local
271 rx_ring = &softc->rx_rings[rxqid]; in bnxt_isc_rxd_refill()
275 rx_ring = &softc->ag_rings[rxqid]; in bnxt_isc_rxd_refill()
278 rxbd = (void *)rx_ring->vaddr; in bnxt_isc_rxd_refill()
287 if (++pidx == rx_ring->ring_size) in bnxt_isc_rxd_refill()
298 struct bnxt_ring *rx_ring; in bnxt_isc_rxd_flush() local
301 rx_ring = &softc->rx_rings[rxqid]; in bnxt_isc_rxd_flush()
303 rx_ring = &softc->ag_rings[rxqid]; in bnxt_isc_rxd_flush()
311 softc->db_ops.bnxt_db_rx(rx_ring, pidx); in bnxt_isc_rxd_flush()
/freebsd/sys/dev/neta/
H A Dif_mvnetavar.h213 KASSERT(mtx_owned(&(sc)->rx_ring[(q)].ring_mtx),\
289 struct mvneta_rx_ring rx_ring[MVNETA_RX_QNUM_MAX]; member
317 (&(sc)->rx_ring[(q)])
/freebsd/sys/contrib/dev/athk/ath12k/
H A Ddp_rx.c260 struct dp_rxdma_ring *rx_ring, in ath12k_dp_rx_bufs_replenish() argument
276 req_entries = min(req_entries, rx_ring->bufs_max); in ath12k_dp_rx_bufs_replenish()
278 srng = &ab->hal.srng_list[rx_ring->refill_buf_ring.ring_id]; in ath12k_dp_rx_bufs_replenish()
285 if (!req_entries && (num_free > (rx_ring->bufs_max * 3) / 4)) in ath12k_dp_rx_bufs_replenish()
334 spin_lock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish()
335 buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 0, in ath12k_dp_rx_bufs_replenish()
336 rx_ring->bufs_max * 3, GFP_ATOMIC); in ath12k_dp_rx_bufs_replenish()
337 spin_unlock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish()
371 spin_lock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish()
372 idr_remove(&rx_ring->bufs_idr, buf_id); in ath12k_dp_rx_bufs_replenish()
[all …]

1234