Searched refs:ssbdp (Results 1 – 6 of 6) sorted by relevance
/titanic_41/usr/src/uts/common/io/nge/ |
H A D | nge_tx.c | 58 sw_tx_sbd_t *ssbdp; in nge_tx_recycle_all() local 68 ssbdp = srp->sw_sbds + slot; in nge_tx_recycle_all() 70 DMA_ZERO(ssbdp->desc); in nge_tx_recycle_all() 72 if (ssbdp->mp != NULL) { in nge_tx_recycle_all() 74 for (dmah = ssbdp->mp_hndl.head; dmah != NULL; in nge_tx_recycle_all() 78 freemsg(ssbdp->mp); in nge_tx_recycle_all() 81 NGE_TXSWD_RECYCLE(ssbdp); in nge_tx_recycle_all() 160 sw_tx_sbd_t *ssbdp; in nge_tx_recycle() local 210 ssbdp = &srp->sw_sbds[slot]; in nge_tx_recycle() 211 hw_sbd_p = DMA_VPTR(ssbdp->desc); in nge_tx_recycle() [all …]
|
H A D | nge_main.c | 462 sw_tx_sbd_t *ssbdp; in nge_fini_send_ring() local 465 ssbdp = srp->sw_sbds; in nge_fini_send_ring() 483 kmem_free(ssbdp, srp->desc.nslots*sizeof (*ssbdp)); in nge_fini_send_ring() 501 sw_tx_sbd_t *ssbdp; in nge_init_send_ring() local 518 ssbdp = kmem_zalloc(nslots*sizeof (*ssbdp), KM_SLEEP); in nge_init_send_ring() 519 srp->sw_sbds = ssbdp; in nge_init_send_ring() 527 for (slot = 0; slot < nslots/ngep->nge_split; ++ssbdp, ++slot) { in nge_init_send_ring() 528 nge_slice_chunk(&ssbdp->desc, &desc, 1, in nge_init_send_ring() 530 nge_slice_chunk(&ssbdp->pbuf, &pbuf, 1, in nge_init_send_ring() 572 sw_tx_sbd_t *ssbdp; in nge_reinit_send_ring() local [all …]
|
/titanic_41/usr/src/uts/common/io/bge/ |
H A D | bge_send.c | 98 sw_sbd_t *ssbdp; in bge_recycle_ring() local 120 ssbdp = &srp->sw_sbds[slot]; in bge_recycle_ring() 121 ASSERT(ssbdp->pbuf != NULL); in bge_recycle_ring() 122 buf_item = ssbdp->pbuf; in bge_recycle_ring() 129 ssbdp->pbuf = NULL; in bge_recycle_ring() 370 sw_sbd_t *ssbdp; in bge_send_serial() local 432 ssbdp = &srp->sw_sbds[tx_next]; in bge_send_serial() 433 ASSERT(ssbdp->pbuf == NULL); in bge_send_serial() 434 ssbdp->pbuf = txbuf_item; in bge_send_serial() 439 hw_sbd_p = DMA_VPTR(ssbdp->desc); in bge_send_serial() [all …]
|
H A D | bge_main2.c | 215 sw_sbd_t *ssbdp; in bge_reinit_send_ring() local 270 ssbdp = srp->sw_sbds; in bge_reinit_send_ring() 271 for (slot = 0; slot < srp->desc.nslots; ++ssbdp, ++slot) in bge_reinit_send_ring() 272 ssbdp->pbuf = NULL; in bge_reinit_send_ring() 2371 sw_sbd_t *ssbdp; local 2419 ssbdp = kmem_zalloc(nslots*sizeof (*ssbdp), KM_SLEEP); 2424 srp->sw_sbds = ssbdp; 2438 for (slot = 0; slot < nslots; ++ssbdp, ++slot) { 2439 bge_slice_chunk(&ssbdp->desc, &desc, 1,
|
/titanic_41/usr/src/uts/common/io/rge/ |
H A D | rge_main.c | 308 sw_sbd_t *ssbdp; in rge_reinit_send_ring() local 316 ssbdp = rgep->sw_sbds; in rge_reinit_send_ring() 320 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress); in rge_reinit_send_ring() 322 RGE_BSWAP_32(ssbdp->pbuf.cookie.dmac_laddress >> 32); in rge_reinit_send_ring() 326 ssbdp++; in rge_reinit_send_ring() 396 sw_sbd_t *ssbdp; in rge_fini_send_ring() local 399 ssbdp = rgep->sw_sbds; in rge_fini_send_ring() 401 rge_free_dma_mem(&ssbdp->pbuf); in rge_fini_send_ring() 402 ssbdp++; in rge_fini_send_ring() 472 sw_sbd_t *ssbdp; in rge_init_send_ring() local [all …]
|
H A D | rge_rxtx.c | 483 sw_sbd_t *ssbdp; in rge_send_copy() local 506 ssbdp = &rgep->sw_sbds[slot]; in rge_send_copy() 517 txb = DMA_VPTR(ssbdp->pbuf); in rge_send_copy() 558 DMA_SYNC(ssbdp->pbuf, DDI_DMA_SYNC_FORDEV); in rge_send_copy() 564 ASSERT(hw_sbd_p == ssbdp->desc.mem_va); in rge_send_copy() 581 ethhdr = (struct ether_header *)(DMA_VPTR(ssbdp->pbuf)); in rge_send_copy() 587 ((uint8_t *)DMA_VPTR(ssbdp->pbuf) + in rge_send_copy()
|