/titanic_50/usr/src/uts/common/io/e1000g/ |
H A D | e1000g_tx.c | 93 packet->tx_buf->len = 0; in e1000g_free_tx_swpkt() 1263 dma_buffer_t *tx_buf; in e1000g_tx_copy() local 1269 tx_buf = packet->tx_buf; in e1000g_tx_copy() 1272 ASSERT((tx_buf->len + len) <= tx_buf->size); in e1000g_tx_copy() 1276 tx_buf->address + tx_buf->len, in e1000g_tx_copy() 1278 tx_buf->len += len; in e1000g_tx_copy() 1288 if ((tx_buf->len + len1) > tx_buf->size) in e1000g_tx_copy() 1300 (tx_buf->len > len)); in e1000g_tx_copy() 1309 ASSERT(tx_buf->len < ETHERMIN); in e1000g_tx_copy() 1311 bzero(tx_buf->address + tx_buf->len, in e1000g_tx_copy() [all …]
|
H A D | e1000g_alloc.c | 1129 dma_buffer_t *tx_buf; in e1000g_alloc_tx_packets() local 1194 tx_buf = packet->tx_buf; in e1000g_alloc_tx_packets() 1200 tx_buf, Adapter->tx_buffer_size); in e1000g_alloc_tx_packets() 1205 tx_buf, Adapter->tx_buffer_size, &dma_attr); in e1000g_alloc_tx_packets() 1471 dma_buffer_t *tx_buf; in e1000g_free_tx_packets() local 1506 tx_buf = packet->tx_buf; in e1000g_free_tx_packets() 1511 e1000g_free_dvma_buffer(tx_buf); in e1000g_free_tx_packets() 1515 e1000g_free_dma_buffer(tx_buf); in e1000g_free_tx_packets()
|
H A D | e1000g_sw.h | 570 dma_buffer_t tx_buf[1]; member
|
/titanic_50/usr/src/uts/common/io/igb/ |
H A D | igb_tx.c | 327 ASSERT((tcb->tx_buf.len + current_len) <= in igb_tx() 328 tcb->tx_buf.size); in igb_tx() 338 } else if ((tcb->tx_buf.len + current_len + next_len) > in igb_tx() 339 tcb->tx_buf.size) { in igb_tx() 485 dma_buffer_t *tx_buf; in igb_tx_copy() local 489 tx_buf = &tcb->tx_buf; in igb_tx_copy() 502 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in igb_tx_copy() 504 tx_buf->len += len; in igb_tx_copy() 521 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in igb_tx_copy() 530 igb_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in igb_tx_copy() [all …]
|
H A D | igb_buf.c | 667 dma_buffer_t *tx_buf; in igb_alloc_tcb_lists() local 749 tx_buf = &tcb->tx_buf; in igb_alloc_tcb_lists() 752 tx_buf, igb->tx_buf_size); in igb_alloc_tcb_lists() 802 igb_free_dma_buffer(&tcb->tx_buf); in igb_free_tcb_lists()
|
H A D | igb_sw.h | 391 dma_buffer_t tx_buf; member
|
/titanic_50/usr/src/uts/common/io/ixgbe/ |
H A D | ixgbe_tx.c | 317 ASSERT((tcb->tx_buf.len + current_len) <= in ixgbe_ring_tx() 318 tcb->tx_buf.size); in ixgbe_ring_tx() 328 } else if ((tcb->tx_buf.len + current_len + next_len) > in ixgbe_ring_tx() 329 tcb->tx_buf.size) { in ixgbe_ring_tx() 584 dma_buffer_t *tx_buf; in ixgbe_tx_copy() local 588 tx_buf = &tcb->tx_buf; in ixgbe_tx_copy() 601 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in ixgbe_tx_copy() 603 tx_buf->len += len; in ixgbe_tx_copy() 620 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in ixgbe_tx_copy() 629 ixgbe_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in ixgbe_tx_copy() [all …]
|
H A D | ixgbe_buf.c | 661 dma_buffer_t *tx_buf; in ixgbe_alloc_tcb_lists() local 743 tx_buf = &tcb->tx_buf; in ixgbe_alloc_tcb_lists() 746 tx_buf, ixgbe->tx_buf_size); in ixgbe_alloc_tcb_lists() 796 ixgbe_free_dma_buffer(&tcb->tx_buf); in ixgbe_free_tcb_lists()
|
H A D | ixgbe_sw.h | 437 dma_buffer_t tx_buf; member
|
/titanic_50/usr/src/uts/intel/io/amd8111s/ |
H A D | amd8111s_main.c | 1098 if (amd8111s_alloc_dma_ringbuf(pLayerPointers, &pOdl->tx_buf, in amd8111s_allocate_buffers() 1129 amd8111s_free_dma_ringbuf(&pLayerPointers->pOdl->tx_buf); in amd8111s_free_buffers() 1148 pLayerPointers->pOdl->tx_buf.free = in amd8111s_recycle_tx() 1149 NEXT(pLayerPointers->pOdl->tx_buf, free); in amd8111s_recycle_tx() 1178 if (pLayerPointers->pOdl->tx_buf.curr == in amd8111s_send_serial() 1179 pLayerPointers->pOdl->tx_buf.next) { in amd8111s_send_serial() 1195 ->pOdl->tx_buf.curr->msg_size; in amd8111s_send_serial() 1199 pLayerPointers->pOdl->tx_buf.curr->phy_addr; in amd8111s_send_serial() 1213 pLayerPointers->pOdl->tx_buf.curr = in amd8111s_send_serial() 1214 NEXT(pLayerPointers->pOdl->tx_buf, curr); in amd8111s_send_serial() [all …]
|
H A D | amd8111s_hw.c | 1721 pLayerPointers->pOdl->tx_buf.free = in milResetTxQ() 1722 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ() 1723 pLayerPointers->pOdl->tx_buf.next = in milResetTxQ() 1724 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ() 1725 pLayerPointers->pOdl->tx_buf.curr = in milResetTxQ() 1726 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ()
|
H A D | amd8111s_main.h | 326 struct amd8111s_dma_ringbuf tx_buf; member
|
/titanic_50/usr/src/uts/common/io/rtls/ |
H A D | rtls.c | 1128 mcopymsg(mp, rtlsp->tx_buf[cur_desc]); in rtls_send() 1131 if (*rtlsp->tx_buf[cur_desc] & 0x1) { in rtls_send() 1132 uint16_t *ptr = (void *)rtlsp->tx_buf[cur_desc]; in rtls_send() 1145 bzero(rtlsp->tx_buf[cur_desc] + totlen, ETHERMIN - totlen); in rtls_send() 1593 rtlsp->tx_buf[i] = (uint8_t *)rtlsp->dma_area_tx[i].mem_va; in rtls_alloc_bufs() 1624 rtlsp->tx_buf[i] = NULL; in rtls_free_bufs()
|
H A D | rtls.h | 180 uint8_t *tx_buf[RTLS_MAX_TX_DESC]; member
|
/titanic_50/usr/src/uts/common/io/sfe/ |
H A D | sfe_util.c | 109 &(dp)->tx_buf[SLOT((dp)->tx_slots_base + (sn), (dp)->gc.gc_tx_buf_size)] 699 for (i = 0, tbp = dp->tx_buf; in gem_alloc_memory() 752 (void) ddi_dma_unbind_handle(dp->tx_buf[i].txb_bdh); in gem_alloc_memory() 753 ddi_dma_mem_free(&dp->tx_buf[i].txb_bah); in gem_alloc_memory() 754 ddi_dma_free_handle(&dp->tx_buf[i].txb_bdh); in gem_alloc_memory() 786 for (i = dp->gc.gc_tx_buf_size, tbp = dp->tx_buf; i--; tbp++) { in gem_free_memory() 4848 dp->tx_buf = (void *)&dp->mc_list[GEM_MAXMC]; in gem_do_attach() 4851 dp->tx_buf[i].txb_next = in gem_do_attach() 4852 &dp->tx_buf[SLOT(i + 1, dp->gc.gc_tx_buf_size)]; in gem_do_attach()
|
H A D | sfe_util.h | 243 struct txbuf *tx_buf; member
|
/titanic_50/usr/src/uts/common/io/arn/ |
H A D | arn_xmit.c | 251 tid->tx_buf[cindex] = NULL; in arn_tx_update_baw() 253 while (tid->baw_head != tid->baw_tail && !tid->tx_buf[tid->baw_head]) { in arn_tx_update_baw() 272 ASSERT(tid->tx_buf[cindex] == NULL); in arn_tx_addto_baw() 273 tid->tx_buf[cindex] = bf; in arn_tx_addto_baw()
|
H A D | arn_core.h | 496 struct ath_buf *tx_buf[ATH_TID_MAX_BUFS]; /* active tx frames */ member
|