/freebsd/sys/dev/rtwn/pci/ |
H A D | rtwn_pci_attach.c | 125 struct rtwn_rx_ring *rx_ring = &pc->rx_ring; in rtwn_pci_alloc_rx_list() local 134 size, 1, size, 0, NULL, NULL, &rx_ring->desc_dmat); in rtwn_pci_alloc_rx_list() 140 error = bus_dmamem_alloc(rx_ring->desc_dmat, (void **)&rx_ring->desc, in rtwn_pci_alloc_rx_list() 142 &rx_ring->desc_map); in rtwn_pci_alloc_rx_list() 147 error = bus_dmamap_load(rx_ring->desc_dmat, rx_ring->desc_map, in rtwn_pci_alloc_rx_list() 148 rx_ring->desc, size, rtwn_pci_dma_map_addr, &rx_ring->paddr, 0); in rtwn_pci_alloc_rx_list() 153 bus_dmamap_sync(rx_ring->desc_dmat, rx_ring->desc_map, in rtwn_pci_alloc_rx_list() 159 MJUMPAGESIZE, 1, MJUMPAGESIZE, 0, NULL, NULL, &rx_ring->data_dmat); in rtwn_pci_alloc_rx_list() 167 rx_data = &rx_ring->rx_data[i]; in rtwn_pci_alloc_rx_list() 168 error = bus_dmamap_create(rx_ring->data_dmat, 0, &rx_data->map); in rtwn_pci_alloc_rx_list() [all …]
|
/freebsd/sys/dev/ena/ |
H A D | ena_datapath.c | 74 struct ena_ring *rx_ring; in ena_cleanup() local 81 rx_ring = que->rx_ring; in ena_cleanup() 99 atomic_store_8(&rx_ring->first_interrupt, 1); in ena_cleanup() 102 rxc = ena_rx_cleanup(rx_ring); in ena_cleanup() 346 ena_rx_hash_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_ctx *ena_rx_ctx, in ena_rx_hash_mbuf() argument 349 struct ena_adapter *adapter = rx_ring->adapter; in ena_rx_hash_mbuf() 405 mbuf->m_pkthdr.flowid = rx_ring->qid; in ena_rx_hash_mbuf() 419 ena_rx_mbuf(struct ena_ring *rx_ring, struct ena_com_rx_buf_info *ena_bufs, in ena_rx_mbuf() argument 430 adapter = rx_ring->adapter; in ena_rx_mbuf() 435 rx_info = &rx_ring->rx_buffer_info[req_id]; in ena_rx_mbuf() [all …]
|
H A D | ena.c | 400 rxr = &adapter->rx_ring[i]; in ena_init_io_rings_basic() 414 que->rx_ring = rxr; in ena_init_io_rings_basic() 432 rxr = &adapter->rx_ring[i]; in ena_init_io_rings_advanced() 475 struct ena_ring *rxr = &adapter->rx_ring[qid]; in ena_free_io_ring_resources() 873 struct ena_ring *rx_ring = que->rx_ring; in ena_setup_rx_resources() local 876 size = sizeof(struct ena_rx_buffer) * rx_ring->ring_size; in ena_setup_rx_resources() 880 rx_ring->initialized = false; in ena_setup_rx_resources() 889 rx_ring->rx_buffer_info = malloc(size, M_DEVBUF, M_WAITOK | M_ZERO); in ena_setup_rx_resources() 891 size = sizeof(uint16_t) * rx_ring->ring_size; in ena_setup_rx_resources() 892 rx_ring->free_rx_ids = malloc(size, M_DEVBUF, M_WAITOK); in ena_setup_rx_resources() [all …]
|
H A D | ena_netmap.h | 51 struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info); 53 struct ena_ring *rx_ring, struct ena_rx_buffer *rx_info);
|
/freebsd/sys/ofed/drivers/infiniband/ulp/sdp/ |
H A D | sdp_rx.c | 109 int id = ring_head(ssk->rx_ring); in sdp_post_recv() 125 rx_req = ssk->rx_ring.buffer + (id & (SDP_RX_SIZE - 1)); in sdp_post_recv() 156 atomic_inc(&ssk->rx_ring.head); in sdp_post_recv() 286 ssk->recv_request_head = ring_head(ssk->rx_ring) + 1; in sdp_handle_resize_request() 288 ssk->recv_request_head = ring_tail(ssk->rx_ring); in sdp_handle_resize_request() 308 if (unlikely(id != ring_tail(ssk->rx_ring))) { in sdp_recv_completion() 310 id, ring_tail(ssk->rx_ring)); in sdp_recv_completion() 315 rx_req = &ssk->rx_ring.buffer[id & (SDP_RX_SIZE - 1)]; in sdp_recv_completion() 319 atomic_inc(&ssk->rx_ring.tail); in sdp_recv_completion() 534 struct ib_cq *cq = ssk->rx_ring.cq; in sdp_poll_rx_cq() [all …]
|
H A D | sdp.h | 148 #define rx_ring_posted(ssk) ring_posted(ssk->rx_ring) 359 struct sdp_rx_ring rx_ring; member 423 static inline void rx_ring_unlock(struct sdp_rx_ring *rx_ring) in rx_ring_unlock() argument 425 rw_runlock(&rx_ring->destroyed_lock); in rx_ring_unlock() 428 static inline int rx_ring_trylock(struct sdp_rx_ring *rx_ring) in rx_ring_trylock() argument 430 rw_rlock(&rx_ring->destroyed_lock); in rx_ring_trylock() 431 if (rx_ring->destroyed) { in rx_ring_trylock() 432 rx_ring_unlock(rx_ring); in rx_ring_trylock() 438 static inline void rx_ring_destroy_lock(struct sdp_rx_ring *rx_ring) in rx_ring_destroy_lock() argument 440 rw_wlock(&rx_ring->destroyed_lock); in rx_ring_destroy_lock() [all …]
|
/freebsd/sys/dev/al_eth/ |
H A D | al_eth.c | 1000 struct al_eth_ring *ring = &adapter->rx_ring[i]; in al_eth_init_rings() 1041 struct al_eth_ring *rx_ring, in al_eth_alloc_rx_buf() argument 1067 error = bus_dmamap_load_mbuf_sg(rx_ring->dma_buf_tag, rx_info->dma_map, in al_eth_alloc_rx_buf() 1070 device_printf(rx_ring->dev, "failed to map mbuf, error = %d\n", in al_eth_alloc_rx_buf() 1088 struct al_eth_ring *rx_ring = &adapter->rx_ring[qid]; in al_eth_refill_rx_bufs() local 1092 next_to_use = rx_ring->next_to_use; in al_eth_refill_rx_bufs() 1097 &rx_ring->rx_buffer_info[next_to_use]; in al_eth_refill_rx_bufs() 1100 rx_ring, rx_info) < 0)) { in al_eth_refill_rx_bufs() 1106 rc = al_eth_rx_buffer_add(rx_ring->dma_q, in al_eth_refill_rx_bufs() 1114 next_to_use = AL_ETH_RX_RING_IDX_NEXT(rx_ring, next_to_use); in al_eth_refill_rx_bufs() [all …]
|
H A D | al_eth.h | 152 #define AL_ETH_RX_RING_IDX_NEXT(rx_ring, idx) (((idx) + 1) & (AL_ETH_DEFAULT_RX_DESCS - 1)) argument 153 #define AL_ETH_RX_RING_IDX_ADD(rx_ring, idx, n) (((idx) + (n)) & (AL_ETH_DEFAULT_RX_DESCS - 1)) argument 277 struct al_eth_ring rx_ring[AL_ETH_NUM_QUEUES]; member
|
/freebsd/sys/dev/qlxge/ |
H A D | qls_hw.c | 928 ha->rx_ring[0].cq_db_offset)); in qls_init_hw_if() 934 Q81_WR_LBQ_PROD_IDX(i, ha->rx_ring[i].lbq_in); in qls_init_hw_if() 935 Q81_WR_SBQ_PROD_IDX(i, ha->rx_ring[i].sbq_in); in qls_init_hw_if() 1049 rxr = &ha->rx_ring[cid]; in qls_init_comp_queue() 1538 if (ha->rx_ring[ridx].flags.cq_dma) { in qls_free_rx_ring_dma() 1539 qls_free_dmabuf(ha, &ha->rx_ring[ridx].cq_dma); in qls_free_rx_ring_dma() 1540 ha->rx_ring[ridx].flags.cq_dma = 0; in qls_free_rx_ring_dma() 1543 if (ha->rx_ring[ridx].flags.lbq_dma) { in qls_free_rx_ring_dma() 1544 qls_free_dmabuf(ha, &ha->rx_ring[ridx].lbq_dma); in qls_free_rx_ring_dma() 1545 ha->rx_ring[ridx].flags.lbq_dma = 0; in qls_free_rx_ring_dma() [all …]
|
H A D | qls_dbg.c | 274 qls_dump_buf32(ha, "cq_icb", ha->rx_ring[0].cq_icb_vaddr, in qls_dump_cq() 278 (void *)ha->rx_ring[0].lbq_addr_tbl_paddr); in qls_dump_cq() 280 qls_dump_buf32(ha, "lbq_addr_tbl", ha->rx_ring[0].lbq_addr_tbl_vaddr, in qls_dump_cq() 284 (void *)ha->rx_ring[0].lbq_paddr); in qls_dump_cq() 286 qls_dump_buf32(ha, "lbq", ha->rx_ring[0].lbq_vaddr, in qls_dump_cq() 290 (void *)ha->rx_ring[0].sbq_addr_tbl_paddr); in qls_dump_cq() 292 qls_dump_buf32(ha, "sbq_addr_tbl", ha->rx_ring[0].sbq_addr_tbl_vaddr, in qls_dump_cq() 296 (void *)ha->rx_ring[0].sbq_paddr); in qls_dump_cq() 298 qls_dump_buf32(ha, "sbq", ha->rx_ring[0].sbq_vaddr, in qls_dump_cq() 302 (void *)ha->rx_ring[0].lb_paddr); in qls_dump_cq()
|
H A D | qls_isr.c | 80 rxr = &ha->rx_ring[r_idx]; in qls_replenish_rx() 134 rxr = &ha->rx_ring[rxr_idx]; in qls_rx_comp() 235 struct lro_ctrl *lro = &ha->rx_ring[cq_idx].lro; in qls_cq_isr() 238 cq_b = ha->rx_ring[cq_idx].cq_base_vaddr; in qls_cq_isr() 240 cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr); in qls_cq_isr() 242 i = ha->rx_ring[cq_idx].cq_next; in qls_cq_isr() 263 __func__, i, (*(ha->rx_ring[cq_idx].cqi_vaddr)), in qls_cq_isr() 279 cq_comp_idx = *(ha->rx_ring[cq_idx].cqi_vaddr); in qls_cq_isr() 292 ha->rx_ring[cq_idx].cq_next = cq_comp_idx; in qls_cq_isr() 295 Q81_WR_CQ_CONS_IDX(cq_idx, (ha->rx_ring[cq_idx].cq_next)); in qls_cq_isr()
|
H A D | qls_os.c | 203 (void *)ha->rx_ring[i].rx_int)); in qls_sysctl_get_drvr_stats() 208 (void *)ha->rx_ring[i].rss_int)); in qls_sysctl_get_drvr_stats() 213 ha->rx_ring[i].lbq_next); in qls_sysctl_get_drvr_stats() 218 ha->rx_ring[i].lbq_free); in qls_sysctl_get_drvr_stats() 223 ha->rx_ring[i].lbq_in); in qls_sysctl_get_drvr_stats() 228 ha->rx_ring[i].sbq_next); in qls_sysctl_get_drvr_stats() 233 ha->rx_ring[i].sbq_free); in qls_sysctl_get_drvr_stats() 238 ha->rx_ring[i].sbq_in); in qls_sysctl_get_drvr_stats() 1259 qla_rx_ring_t *rx_ring; in qls_alloc_rcv_mbufs() local 1262 rx_ring = &ha->rx_ring[r]; in qls_alloc_rcv_mbufs() [all …]
|
H A D | qls_hw.h | 934 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG), idx) 937 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_INDEX_REG)) 940 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\ 944 (ha->rx_ring[cq_idx].cq_db_offset + Q81_COMPQ_VALID_REG),\ 948 (ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG), idx) 951 (ha->rx_ring[cq_idx].cq_db_offset + Q81_LRGBQ_INDEX_REG)) 954 (ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG), idx) 957 (ha->rx_ring[cq_idx].cq_db_offset + Q81_SMBQ_INDEX_REG))
|
/freebsd/sys/contrib/dev/rtw89/ |
H A D | pci.c | 154 struct rtw89_pci_rx_ring *rx_ring) in rtw89_pci_sync_skb_for_device() 156 struct rtw89_pci_dma_ring *bd_ring = &rx_ring->bd_ring; in rtw89_pci_sync_skb_for_device() 207 struct rtw89_pci_rx_ring *rx_ring, 218 if (rx_ring->target_rx_tag == 0) in rtw89_skb_put_rx_data() 221 target_rx_tag = rx_ring->target_rx_tag; in rtw89_skb_put_rx_data() 234 struct rtw89_pci_rx_ring *rx_ring, in rtw89_skb_put_rx_data() 245 ret = rtw89_pci_validate_rx_tag(rtwdev, rx_ring, skb); in rtw89_pci_rxbd_deliver_skbs() 251 rx_ring->target_rx_tag = rx_info->tag + 1; in rtw89_pci_rxbd_deliver_skbs() 329 struct rtw89_pci_rx_ring *rx_ring) in rtw89_pci_rxbd_deliver() 331 struct rtw89_pci_dma_ring *bd_ring = &rx_ring in rtw89_pci_rxbd_deliver() 128 rtw89_pci_rxbd_recalc(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_rxbd_recalc() argument 237 rtw89_pci_rxbd_deliver_skbs(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_rxbd_deliver_skbs() argument 322 rtw89_pci_rxbd_deliver(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 cnt) rtw89_pci_rxbd_deliver() argument 347 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_poll_rxq_dma() local 524 rtw89_pci_release_tx_skbs(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 max_cnt) rtw89_pci_release_tx_skbs() argument 575 rtw89_pci_release_tx(struct rtw89_dev * rtwdev,struct rtw89_pci_rx_ring * rx_ring,u32 cnt) rtw89_pci_release_tx() argument 600 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_poll_rpq_dma() local 627 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_isr_rxd_unavail() local 982 struct rtw89_pci_rx_ring *rx_ring; __rtw89_pci_check_and_reclaim_tx_resource() local 1437 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_reset_trx_rings() local 1576 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_switch_bd_idx_addr() local 2888 rtw89_pci_free_rx_ring(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring) rtw89_pci_free_rx_ring() argument 2922 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_free_rx_rings() local 2939 rtw89_pci_init_rx_bd(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring,struct sk_buff * skb,int buf_sz,u32 idx) rtw89_pci_init_rx_bd() argument 3106 rtw89_pci_alloc_rx_ring(struct rtw89_dev * rtwdev,struct pci_dev * pdev,struct rtw89_pci_rx_ring * rx_ring,u32 desc_size,u32 len,u32 rxch) rtw89_pci_alloc_rx_ring() argument 3191 struct rtw89_pci_rx_ring *rx_ring; rtw89_pci_alloc_rx_rings() local [all...] |
/freebsd/sys/contrib/dev/athk/ath10k/ |
H A D | htt_rx.c | 34 hash_for_each_possible(ar->htt.rx_ring.skb_table, rxcb, hlist, paddr) in ath10k_htt_rx_find_skb_paddr() 49 if (htt->rx_ring.in_ord_rx) { in ath10k_htt_rx_ring_free() 50 hash_for_each_safe(htt->rx_ring.skb_table, i, n, rxcb, hlist) { in ath10k_htt_rx_ring_free() 59 for (i = 0; i < htt->rx_ring.size; i++) { in ath10k_htt_rx_ring_free() 60 skb = htt->rx_ring.netbufs_ring[i]; in ath10k_htt_rx_ring_free() 72 htt->rx_ring.fill_cnt = 0; in ath10k_htt_rx_ring_free() 73 hash_init(htt->rx_ring.skb_table); in ath10k_htt_rx_ring_free() 74 memset(htt->rx_ring.netbufs_ring, 0, in ath10k_htt_rx_ring_free() 75 htt->rx_ring.size * sizeof(htt->rx_ring.netbufs_ring[0])); in ath10k_htt_rx_ring_free() 80 return htt->rx_ring.size * sizeof(htt->rx_ring.paddrs_ring_32); in ath10k_htt_get_rx_ring_size_32() [all …]
|
/freebsd/sys/dev/et/ |
H A D | if_et.c | 860 struct et_rxdesc_ring *rx_ring; in et_dma_alloc() local 894 rx_ring = &sc->sc_rx_ring[0]; in et_dma_alloc() 896 &rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap, in et_dma_alloc() 897 &rx_ring->rr_paddr, "RX ring 0"); in et_dma_alloc() 898 rx_ring->rr_posreg = ET_RX_RING0_POS; in et_dma_alloc() 903 rx_ring = &sc->sc_rx_ring[1]; in et_dma_alloc() 905 &rx_ring->rr_dtag, (uint8_t **)&rx_ring->rr_desc, &rx_ring->rr_dmap, in et_dma_alloc() 906 &rx_ring->rr_paddr, "RX ring 1"); in et_dma_alloc() 907 rx_ring->rr_posreg = ET_RX_RING1_POS; in et_dma_alloc() 1039 struct et_rxdesc_ring *rx_ring; in et_dma_free() local [all …]
|
/freebsd/sys/contrib/dev/athk/ath11k/ |
H A D | dp_rx.c | 344 struct dp_rxdma_ring *rx_ring, in ath11k_dp_rxbufs_replenish() argument 357 req_entries = min(req_entries, rx_ring->bufs_max); in ath11k_dp_rxbufs_replenish() 359 srng = &ab->hal.srng_list[rx_ring->refill_buf_ring.ring_id]; in ath11k_dp_rxbufs_replenish() 366 if (!req_entries && (num_free > (rx_ring->bufs_max * 3) / 4)) in ath11k_dp_rxbufs_replenish() 391 spin_lock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish() 392 buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 1, in ath11k_dp_rxbufs_replenish() 393 (rx_ring->bufs_max * 3) + 1, GFP_ATOMIC); in ath11k_dp_rxbufs_replenish() 394 spin_unlock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish() 419 spin_lock_bh(&rx_ring->idr_lock); in ath11k_dp_rxbufs_replenish() 420 idr_remove(&rx_ring->bufs_idr, buf_id); in ath11k_dp_rxbufs_replenish() [all …]
|
H A D | spectral.c | 833 ret = ath11k_dbring_srng_setup(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc() 840 ath11k_dbring_set_cfg(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc() 845 ret = ath11k_dbring_buf_setup(ar, &sp->rx_ring, db_cap); in ath11k_spectral_ring_alloc() 851 ret = ath11k_dbring_wmi_cfg_setup(ar, &sp->rx_ring, in ath11k_spectral_ring_alloc() 861 ath11k_dbring_buf_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_alloc() 863 ath11k_dbring_srng_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_alloc() 871 ath11k_dbring_srng_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_free() 872 ath11k_dbring_buf_cleanup(ar, &sp->rx_ring); in ath11k_spectral_ring_free() 1015 idr_init(&sp->rx_ring.bufs_idr); in ath11k_spectral_init() 1016 spin_lock_init(&sp->rx_ring.idr_lock); in ath11k_spectral_init() [all …]
|
/freebsd/sys/contrib/dev/rtw88/ |
H A D | pci.c | 174 struct rtw_pci_rx_ring *rx_ring) in rtw_pci_free_rx_ring_skbs() argument 182 for (i = 0; i < rx_ring->r.len; i++) { in rtw_pci_free_rx_ring_skbs() 183 skb = rx_ring->buf[i]; in rtw_pci_free_rx_ring_skbs() 190 rx_ring->buf[i] = NULL; in rtw_pci_free_rx_ring_skbs() 195 struct rtw_pci_rx_ring *rx_ring) in rtw_pci_free_rx_ring() argument 198 u8 *head = rx_ring->r.head; in rtw_pci_free_rx_ring() 199 int ring_sz = rx_ring->r.desc_size * rx_ring->r.len; in rtw_pci_free_rx_ring() 201 rtw_pci_free_rx_ring_skbs(rtwdev, rx_ring); in rtw_pci_free_rx_ring() 203 dma_free_coherent(&pdev->dev, ring_sz, head, rx_ring in rtw_pci_free_rx_ring() 210 struct rtw_pci_rx_ring *rx_ring; rtw_pci_free_trx_ring() local 256 rtw_pci_reset_rx_desc(struct rtw_dev * rtwdev,struct sk_buff * skb,struct rtw_pci_rx_ring * rx_ring,u32 idx,u32 desc_sz) rtw_pci_reset_rx_desc() argument 282 rtw_pci_sync_rx_desc_device(struct rtw_dev * rtwdev,dma_addr_t dma,struct rtw_pci_rx_ring * rx_ring,u32 idx,u32 desc_sz) rtw_pci_sync_rx_desc_device() argument 299 rtw_pci_init_rx_ring(struct rtw_dev * rtwdev,struct rtw_pci_rx_ring * rx_ring,u8 desc_size,u32 len) rtw_pci_init_rx_ring() argument 365 struct rtw_pci_rx_ring *rx_ring; rtw_pci_init_trx_ring() local 734 rtw_pci_dma_check(struct rtw_dev * rtwdev,struct rtw_pci_rx_ring * rx_ring,u32 idx) rtw_pci_dma_check() argument [all...] |
/freebsd/sys/dev/axgbe/ |
H A D | if_axgbe_pci.c | 809 channel->rx_ring = NULL; in axgbe_alloc_channels() 1756 struct xgbe_ring *rx_ring; in axgbe_if_rx_queues_alloc() local 1774 rx_ring = (struct xgbe_ring*)malloc(nrxqs * in axgbe_if_rx_queues_alloc() 1777 if (rx_ring == NULL) { in axgbe_if_rx_queues_alloc() 1782 channel->rx_ring = rx_ring; in axgbe_if_rx_queues_alloc() 1784 for (j = 0; j < nrxqs; j++, rx_ring++) { in axgbe_if_rx_queues_alloc() 1785 rx_ring->rdata = in axgbe_if_rx_queues_alloc() 1790 rx_ring->rdesc = (struct xgbe_ring_desc *)va[i*nrxqs + j]; in axgbe_if_rx_queues_alloc() 1791 rx_ring->rdesc_paddr = pa[i*nrxqs + j]; in axgbe_if_rx_queues_alloc() 1792 rx_ring->rdesc_count = scctx->isc_nrxd[j]; in axgbe_if_rx_queues_alloc() [all …]
|
/freebsd/sys/ofed/drivers/infiniband/ulp/ipoib/ |
H A D | ipoib_cm.c | 130 rx_req = (struct ipoib_rx_buf *)&rx->rx_ring[id]; in ipoib_cm_post_receive_nonsrq() 143 m_freem(rx->rx_ring[id].mb); in ipoib_cm_post_receive_nonsrq() 144 rx->rx_ring[id].mb = NULL; in ipoib_cm_post_receive_nonsrq() 158 struct ipoib_cm_rx_buf *rx_ring) in ipoib_cm_free_rx_ring() argument 163 if (rx_ring[i].mb) { in ipoib_cm_free_rx_ring() 164 ipoib_cm_dma_unmap_rx(priv, &rx_ring[i]); in ipoib_cm_free_rx_ring() 165 m_freem(rx_ring[i].mb); in ipoib_cm_free_rx_ring() 168 kfree(rx_ring); in ipoib_cm_free_rx_ring() 310 rx->rx_ring = kzalloc(ipoib_recvq_size * sizeof *rx->rx_ring, GFP_KERNEL); in ipoib_cm_nonsrq_init_rx() 311 if (!rx->rx_ring) { in ipoib_cm_nonsrq_init_rx() [all …]
|
H A D | ipoib_ib.c | 155 rx_req = &priv->rx_ring[id]; in ipoib_ib_post_receive() 166 ipoib_dma_unmap_rx(priv, &priv->rx_ring[id]); in ipoib_ib_post_receive() 167 m_freem(priv->rx_ring[id].mb); in ipoib_ib_post_receive() 168 priv->rx_ring[id].mb = NULL; in ipoib_ib_post_receive() 177 return ipoib_alloc_map_mb(priv, &priv->rx_ring[id], in ipoib_alloc_rx_mb() 217 mb = priv->rx_ring[wr_id].mb; in ipoib_ib_handle_rx_wc() 227 ipoib_dma_unmap_rx(priv, &priv->rx_ring[wr_id]); in ipoib_ib_handle_rx_wc() 229 priv->rx_ring[wr_id].mb = NULL; in ipoib_ib_handle_rx_wc() 241 memcpy(&saverx, &priv->rx_ring[wr_id], sizeof(saverx)); in ipoib_ib_handle_rx_wc() 247 memcpy(&priv->rx_ring[wr_id], &saverx, sizeof(saverx)); in ipoib_ib_handle_rx_wc() [all …]
|
/freebsd/sys/dev/bnxt/bnxt_en/ |
H A D | bnxt_txrx.c | 252 struct bnxt_ring *rx_ring; in bnxt_isc_rxd_refill() local 271 rx_ring = &softc->rx_rings[rxqid]; in bnxt_isc_rxd_refill() 275 rx_ring = &softc->ag_rings[rxqid]; in bnxt_isc_rxd_refill() 278 rxbd = (void *)rx_ring->vaddr; in bnxt_isc_rxd_refill() 287 if (++pidx == rx_ring->ring_size) in bnxt_isc_rxd_refill() 298 struct bnxt_ring *rx_ring; in bnxt_isc_rxd_flush() local 301 rx_ring = &softc->rx_rings[rxqid]; in bnxt_isc_rxd_flush() 303 rx_ring = &softc->ag_rings[rxqid]; in bnxt_isc_rxd_flush() 311 softc->db_ops.bnxt_db_rx(rx_ring, pidx); in bnxt_isc_rxd_flush()
|
/freebsd/sys/dev/neta/ |
H A D | if_mvnetavar.h | 213 KASSERT(mtx_owned(&(sc)->rx_ring[(q)].ring_mtx),\ 289 struct mvneta_rx_ring rx_ring[MVNETA_RX_QNUM_MAX]; member 317 (&(sc)->rx_ring[(q)])
|
/freebsd/sys/contrib/dev/athk/ath12k/ |
H A D | dp_rx.c | 260 struct dp_rxdma_ring *rx_ring, in ath12k_dp_rx_bufs_replenish() argument 276 req_entries = min(req_entries, rx_ring->bufs_max); in ath12k_dp_rx_bufs_replenish() 278 srng = &ab->hal.srng_list[rx_ring->refill_buf_ring.ring_id]; in ath12k_dp_rx_bufs_replenish() 285 if (!req_entries && (num_free > (rx_ring->bufs_max * 3) / 4)) in ath12k_dp_rx_bufs_replenish() 334 spin_lock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish() 335 buf_id = idr_alloc(&rx_ring->bufs_idr, skb, 0, in ath12k_dp_rx_bufs_replenish() 336 rx_ring->bufs_max * 3, GFP_ATOMIC); in ath12k_dp_rx_bufs_replenish() 337 spin_unlock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish() 371 spin_lock_bh(&rx_ring->idr_lock); in ath12k_dp_rx_bufs_replenish() 372 idr_remove(&rx_ring->bufs_idr, buf_id); in ath12k_dp_rx_bufs_replenish() [all …]
|