Home
last modified time | relevance | path

Searched refs:tx_buf (Results 1 – 18 of 18) sorted by relevance

/titanic_50/usr/src/uts/common/io/e1000g/
H A De1000g_tx.c93 packet->tx_buf->len = 0; in e1000g_free_tx_swpkt()
1263 dma_buffer_t *tx_buf; in e1000g_tx_copy() local
1269 tx_buf = packet->tx_buf; in e1000g_tx_copy()
1272 ASSERT((tx_buf->len + len) <= tx_buf->size); in e1000g_tx_copy()
1276 tx_buf->address + tx_buf->len, in e1000g_tx_copy()
1278 tx_buf->len += len; in e1000g_tx_copy()
1288 if ((tx_buf->len + len1) > tx_buf->size) in e1000g_tx_copy()
1300 (tx_buf->len > len)); in e1000g_tx_copy()
1309 ASSERT(tx_buf->len < ETHERMIN); in e1000g_tx_copy()
1311 bzero(tx_buf->address + tx_buf->len, in e1000g_tx_copy()
[all …]
H A De1000g_alloc.c1129 dma_buffer_t *tx_buf; in e1000g_alloc_tx_packets() local
1194 tx_buf = packet->tx_buf; in e1000g_alloc_tx_packets()
1200 tx_buf, Adapter->tx_buffer_size); in e1000g_alloc_tx_packets()
1205 tx_buf, Adapter->tx_buffer_size, &dma_attr); in e1000g_alloc_tx_packets()
1471 dma_buffer_t *tx_buf; in e1000g_free_tx_packets() local
1506 tx_buf = packet->tx_buf; in e1000g_free_tx_packets()
1511 e1000g_free_dvma_buffer(tx_buf); in e1000g_free_tx_packets()
1515 e1000g_free_dma_buffer(tx_buf); in e1000g_free_tx_packets()
H A De1000g_sw.h570 dma_buffer_t tx_buf[1]; member
/titanic_50/usr/src/uts/common/io/igb/
H A Digb_tx.c327 ASSERT((tcb->tx_buf.len + current_len) <= in igb_tx()
328 tcb->tx_buf.size); in igb_tx()
338 } else if ((tcb->tx_buf.len + current_len + next_len) > in igb_tx()
339 tcb->tx_buf.size) { in igb_tx()
485 dma_buffer_t *tx_buf; in igb_tx_copy() local
489 tx_buf = &tcb->tx_buf; in igb_tx_copy()
502 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in igb_tx_copy()
504 tx_buf->len += len; in igb_tx_copy()
521 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in igb_tx_copy()
530 igb_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in igb_tx_copy()
[all …]
H A Digb_buf.c667 dma_buffer_t *tx_buf; in igb_alloc_tcb_lists() local
749 tx_buf = &tcb->tx_buf; in igb_alloc_tcb_lists()
752 tx_buf, igb->tx_buf_size); in igb_alloc_tcb_lists()
802 igb_free_dma_buffer(&tcb->tx_buf); in igb_free_tcb_lists()
H A Digb_sw.h391 dma_buffer_t tx_buf; member
/titanic_50/usr/src/uts/common/io/ixgbe/
H A Dixgbe_tx.c317 ASSERT((tcb->tx_buf.len + current_len) <= in ixgbe_ring_tx()
318 tcb->tx_buf.size); in ixgbe_ring_tx()
328 } else if ((tcb->tx_buf.len + current_len + next_len) > in ixgbe_ring_tx()
329 tcb->tx_buf.size) { in ixgbe_ring_tx()
584 dma_buffer_t *tx_buf; in ixgbe_tx_copy() local
588 tx_buf = &tcb->tx_buf; in ixgbe_tx_copy()
601 bcopy(mp->b_rptr, tx_buf->address + tx_buf->len, len); in ixgbe_tx_copy()
603 tx_buf->len += len; in ixgbe_tx_copy()
620 DMA_SYNC(tx_buf, DDI_DMA_SYNC_FORDEV); in ixgbe_tx_copy()
629 ixgbe_save_desc(tcb, tx_buf->dma_address, tx_buf->len); in ixgbe_tx_copy()
[all …]
H A Dixgbe_buf.c661 dma_buffer_t *tx_buf; in ixgbe_alloc_tcb_lists() local
743 tx_buf = &tcb->tx_buf; in ixgbe_alloc_tcb_lists()
746 tx_buf, ixgbe->tx_buf_size); in ixgbe_alloc_tcb_lists()
796 ixgbe_free_dma_buffer(&tcb->tx_buf); in ixgbe_free_tcb_lists()
H A Dixgbe_sw.h437 dma_buffer_t tx_buf; member
/titanic_50/usr/src/uts/intel/io/amd8111s/
H A Damd8111s_main.c1098 if (amd8111s_alloc_dma_ringbuf(pLayerPointers, &pOdl->tx_buf, in amd8111s_allocate_buffers()
1129 amd8111s_free_dma_ringbuf(&pLayerPointers->pOdl->tx_buf); in amd8111s_free_buffers()
1148 pLayerPointers->pOdl->tx_buf.free = in amd8111s_recycle_tx()
1149 NEXT(pLayerPointers->pOdl->tx_buf, free); in amd8111s_recycle_tx()
1178 if (pLayerPointers->pOdl->tx_buf.curr == in amd8111s_send_serial()
1179 pLayerPointers->pOdl->tx_buf.next) { in amd8111s_send_serial()
1195 ->pOdl->tx_buf.curr->msg_size; in amd8111s_send_serial()
1199 pLayerPointers->pOdl->tx_buf.curr->phy_addr; in amd8111s_send_serial()
1213 pLayerPointers->pOdl->tx_buf.curr = in amd8111s_send_serial()
1214 NEXT(pLayerPointers->pOdl->tx_buf, curr); in amd8111s_send_serial()
[all …]
H A Damd8111s_hw.c1721 pLayerPointers->pOdl->tx_buf.free = in milResetTxQ()
1722 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ()
1723 pLayerPointers->pOdl->tx_buf.next = in milResetTxQ()
1724 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ()
1725 pLayerPointers->pOdl->tx_buf.curr = in milResetTxQ()
1726 pLayerPointers->pOdl->tx_buf.msg_buf; in milResetTxQ()
H A Damd8111s_main.h326 struct amd8111s_dma_ringbuf tx_buf; member
/titanic_50/usr/src/uts/common/io/rtls/
H A Drtls.c1128 mcopymsg(mp, rtlsp->tx_buf[cur_desc]); in rtls_send()
1131 if (*rtlsp->tx_buf[cur_desc] & 0x1) { in rtls_send()
1132 uint16_t *ptr = (void *)rtlsp->tx_buf[cur_desc]; in rtls_send()
1145 bzero(rtlsp->tx_buf[cur_desc] + totlen, ETHERMIN - totlen); in rtls_send()
1593 rtlsp->tx_buf[i] = (uint8_t *)rtlsp->dma_area_tx[i].mem_va; in rtls_alloc_bufs()
1624 rtlsp->tx_buf[i] = NULL; in rtls_free_bufs()
H A Drtls.h180 uint8_t *tx_buf[RTLS_MAX_TX_DESC]; member
/titanic_50/usr/src/uts/common/io/sfe/
H A Dsfe_util.c109 &(dp)->tx_buf[SLOT((dp)->tx_slots_base + (sn), (dp)->gc.gc_tx_buf_size)]
699 for (i = 0, tbp = dp->tx_buf; in gem_alloc_memory()
752 (void) ddi_dma_unbind_handle(dp->tx_buf[i].txb_bdh); in gem_alloc_memory()
753 ddi_dma_mem_free(&dp->tx_buf[i].txb_bah); in gem_alloc_memory()
754 ddi_dma_free_handle(&dp->tx_buf[i].txb_bdh); in gem_alloc_memory()
786 for (i = dp->gc.gc_tx_buf_size, tbp = dp->tx_buf; i--; tbp++) { in gem_free_memory()
4848 dp->tx_buf = (void *)&dp->mc_list[GEM_MAXMC]; in gem_do_attach()
4851 dp->tx_buf[i].txb_next = in gem_do_attach()
4852 &dp->tx_buf[SLOT(i + 1, dp->gc.gc_tx_buf_size)]; in gem_do_attach()
H A Dsfe_util.h243 struct txbuf *tx_buf; member
/titanic_50/usr/src/uts/common/io/arn/
H A Darn_xmit.c251 tid->tx_buf[cindex] = NULL; in arn_tx_update_baw()
253 while (tid->baw_head != tid->baw_tail && !tid->tx_buf[tid->baw_head]) { in arn_tx_update_baw()
272 ASSERT(tid->tx_buf[cindex] == NULL); in arn_tx_addto_baw()
273 tid->tx_buf[cindex] = bf; in arn_tx_addto_baw()
H A Darn_core.h496 struct ath_buf *tx_buf[ATH_TID_MAX_BUFS]; /* active tx frames */ member