/titanic_41/usr/src/uts/common/io/e1000g/ |
H A D | e1000g_tx.c | 93 packet->tx_buf->len = 0; in e1000g_free_tx_swpkt() 1263 dma_buffer_t *tx_buf; in e1000g_tx_copy() local 1269 tx_buf = packet->tx_buf; in e1000g_tx_copy() 1272 ASSERT((tx_buf->len + len) <= tx_buf->size); in e1000g_tx_copy() 1276 tx_buf->address + tx_buf->len, in e1000g_tx_copy() 1278 tx_buf->len += len; in e1000g_tx_copy() 1288 if ((tx_buf->len + len1) > tx_buf->size) in e1000g_tx_copy() 1300 (tx_buf->len > len)); in e1000g_tx_copy() 1309 ASSERT(tx_buf->len < ETHERMIN); in e1000g_tx_copy() 1311 bzero(tx_buf->address + tx_buf->len, in e1000g_tx_copy() [all …]
|
H A D | e1000g_alloc.c | 1129 dma_buffer_t *tx_buf; in e1000g_alloc_tx_packets() local 1194 tx_buf = packet->tx_buf; in e1000g_alloc_tx_packets() 1200 tx_buf, Adapter->tx_buffer_size); in e1000g_alloc_tx_packets() 1205 tx_buf, Adapter->tx_buffer_size, &dma_attr); in e1000g_alloc_tx_packets() 1471 dma_buffer_t *tx_buf; in e1000g_free_tx_packets() local 1506 tx_buf = packet->tx_buf; in e1000g_free_tx_packets() 1511 e1000g_free_dvma_buffer(tx_buf); in e1000g_free_tx_packets() 1515 e1000g_free_dma_buffer(tx_buf); in e1000g_free_tx_packets()
|
H A D | e1000g_sw.h | 569 dma_buffer_t tx_buf[1]; member
|
/titanic_41/usr/src/uts/common/io/igb/ |
H A D | igb_tx.c | 327 ASSERT((tcb->tx_buf.len + current_len) <= in igb_tx() 328 tcb->tx_buf.size); in igb_tx() 338 } else if ((tcb->tx_buf.len + current_len + next_len) > in igb_tx() 339 tcb->tx_buf.size) { in igb_tx() 485 dma_buffer_t *tx_buf; in igb_tx_copy() local 489 tx_buf = &tcb->tx_buf; in igb_tx_copy() 502 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in igb_tx_copy() 504 tx_buf->len += len; in igb_tx_copy() 521 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in igb_tx_copy() 530 igb_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in igb_tx_copy() [all …]
|
H A D | igb_buf.c | 666 dma_buffer_t *tx_buf; in igb_alloc_tcb_lists() local 748 tx_buf = &tcb->tx_buf; in igb_alloc_tcb_lists() 751 tx_buf, igb->tx_buf_size); in igb_alloc_tcb_lists() 800 igb_free_dma_buffer(&tcb->tx_buf); in igb_free_tcb_lists()
|
H A D | igb_sw.h | 390 dma_buffer_t tx_buf; member
|
/titanic_41/usr/src/uts/common/io/ixgbe/ |
H A D | ixgbe_tx.c | 315 ASSERT((tcb->tx_buf.len + current_len) <= in ixgbe_ring_tx() 316 tcb->tx_buf.size); in ixgbe_ring_tx() 326 } else if ((tcb->tx_buf.len + current_len + next_len) > in ixgbe_ring_tx() 327 tcb->tx_buf.size) { in ixgbe_ring_tx() 582 dma_buffer_t *tx_buf; in ixgbe_tx_copy() local 586 tx_buf = &tcb->tx_buf; in ixgbe_tx_copy() 599 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in ixgbe_tx_copy() 601 tx_buf->len += len; in ixgbe_tx_copy() 618 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in ixgbe_tx_copy() 627 ixgbe_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in ixgbe_tx_copy() [all …]
|
H A D | ixgbe_buf.c | 661 dma_buffer_t *tx_buf; in ixgbe_alloc_tcb_lists() local 743 tx_buf = &tcb->tx_buf; in ixgbe_alloc_tcb_lists() 746 tx_buf, ixgbe->tx_buf_size); in ixgbe_alloc_tcb_lists() 796 ixgbe_free_dma_buffer(&tcb->tx_buf); in ixgbe_free_tcb_lists()
|
H A D | ixgbe_sw.h | 434 dma_buffer_t tx_buf; member
|
/titanic_41/usr/src/uts/intel/io/amd8111s/ |
H A D | amd8111s_main.c | 1098 if (amd8111s_alloc_dma_ringbuf(pLayerPointers, &pOdl->tx_buf, in amd8111s_allocate_buffers() 1129 amd8111s_free_dma_ringbuf(&pLayerPointers->pOdl->tx_buf); in amd8111s_free_buffers() 1148 pLayerPointers->pOdl->tx_buf.free = in amd8111s_recycle_tx() 1149 NEXT(pLayerPointers->pOdl->tx_buf, free); in amd8111s_recycle_tx() 1178 if (pLayerPointers->pOdl->tx_buf.curr == in amd8111s_send_serial() 1179 pLayerPointers->pOdl->tx_buf.next) { in amd8111s_send_serial() 1195 ->pOdl->tx_buf.curr->msg_size; in amd8111s_send_serial() 1199 pLayerPointers->pOdl->tx_buf.curr->phy_addr; in amd8111s_send_serial() 1213 pLayerPointers->pOdl->tx_buf.curr = in amd8111s_send_serial() 1214 NEXT(pLayerPointers->pOdl->tx_buf, curr); in amd8111s_send_serial() [all …]
|
H A D | amd8111s_hw.c | 1721 pLayerPointers->pOdl->tx_buf.free = in milResetTxQ() 1722 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ() 1723 pLayerPointers->pOdl->tx_buf.next = in milResetTxQ() 1724 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ() 1725 pLayerPointers->pOdl->tx_buf.curr = in milResetTxQ() 1726 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ()
|
H A D | amd8111s_main.h | 326 struct amd8111s_dma_ringbuf tx_buf; member
|
/titanic_41/usr/src/uts/common/io/rtls/ |
H A D | rtls.c | 1128 mcopymsg(mp, rtlsp->tx_buf[cur_desc]); in rtls_send() 1131 if (*rtlsp->tx_buf[cur_desc] & 0x1) { in rtls_send() 1132 uint16_t *ptr = (void *)rtlsp->tx_buf[cur_desc]; in rtls_send() 1145 bzero(rtlsp->tx_buf[cur_desc] + totlen, ETHERMIN - totlen); in rtls_send() 1593 rtlsp->tx_buf[i] = (uint8_t *)rtlsp->dma_area_tx[i].mem_va; in rtls_alloc_bufs() 1624 rtlsp->tx_buf[i] = NULL; in rtls_free_bufs()
|
H A D | rtls.h | 180 uint8_t *tx_buf[RTLS_MAX_TX_DESC]; member
|
/titanic_41/usr/src/uts/common/io/sfe/ |
H A D | sfe_util.c | 108 &(dp)->tx_buf[SLOT((dp)->tx_slots_base + (sn), (dp)->gc.gc_tx_buf_size)] 701 for (i = 0, tbp = dp->tx_buf; in gem_alloc_memory() 754 (void) ddi_dma_unbind_handle(dp->tx_buf[i].txb_bdh); in gem_alloc_memory() 755 ddi_dma_mem_free(&dp->tx_buf[i].txb_bah); in gem_alloc_memory() 756 ddi_dma_free_handle(&dp->tx_buf[i].txb_bdh); in gem_alloc_memory() 788 for (i = dp->gc.gc_tx_buf_size, tbp = dp->tx_buf; i--; tbp++) { in gem_free_memory() 4850 dp->tx_buf = (void *)&dp->mc_list[GEM_MAXMC]; in gem_do_attach() 4853 dp->tx_buf[i].txb_next = in gem_do_attach() 4854 &dp->tx_buf[SLOT(i + 1, dp->gc.gc_tx_buf_size)]; in gem_do_attach()
|
H A D | sfe_util.h | 243 struct txbuf *tx_buf; member
|
/titanic_41/usr/src/uts/common/io/arn/ |
H A D | arn_xmit.c | 251 tid->tx_buf[cindex] = NULL; in arn_tx_update_baw() 253 while (tid->baw_head != tid->baw_tail && !tid->tx_buf[tid->baw_head]) { in arn_tx_update_baw() 272 ASSERT(tid->tx_buf[cindex] == NULL); in arn_tx_addto_baw() 273 tid->tx_buf[cindex] = bf; in arn_tx_addto_baw()
|
H A D | arn_core.h | 496 struct ath_buf *tx_buf[ATH_TID_MAX_BUFS]; /* active tx frames */ member
|