Home
last modified time | relevance | path

Searched refs:tx_ring (Results 1 – 25 of 42) sorted by relevance

12

/titanic_44/usr/src/uts/common/io/ixgbe/
H A Dixgbe_tx.c87 ixgbe_tx_ring_t *tx_ring = (ixgbe_tx_ring_t *)arg; in ixgbe_ring_tx() local
88 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_ring_tx()
156 if (tx_ring->tbd_free < ixgbe->tx_recycle_thresh) { in ixgbe_ring_tx()
157 tx_ring->tx_recycle(tx_ring); in ixgbe_ring_tx()
165 if (tx_ring->tbd_free < ixgbe->tx_overload_thresh) { in ixgbe_ring_tx()
166 tx_ring->reschedule = B_TRUE; in ixgbe_ring_tx()
167 tx_ring->stat_overload++; in ixgbe_ring_tx()
217 tx_ring->stat_lso_header_fail++; in ixgbe_ring_tx()
296 tcb = ixgbe_get_free_list(tx_ring); in ixgbe_ring_tx()
299 tx_ring->stat_fail_no_tcb++; in ixgbe_ring_tx()
[all …]
H A Dixgbe_buf.c127 ixgbe_tx_ring_t *tx_ring; in ixgbe_alloc_dma() local
148 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_alloc_dma()
150 if (ixgbe_alloc_tbd_ring(tx_ring) != IXGBE_SUCCESS) in ixgbe_alloc_dma()
153 if (ixgbe_alloc_tcb_lists(tx_ring) != IXGBE_SUCCESS) in ixgbe_alloc_dma()
173 ixgbe_tx_ring_t *tx_ring; in ixgbe_free_dma() local
191 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_free_dma()
192 ixgbe_free_tbd_ring(tx_ring); in ixgbe_free_dma()
193 ixgbe_free_tcb_lists(tx_ring); in ixgbe_free_dma()
309 ixgbe_alloc_tbd_ring(ixgbe_tx_ring_t *tx_ring) in ixgbe_alloc_tbd_ring() argument
317 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_alloc_tbd_ring()
[all …]
H A Dixgbe_main.c1100 ixgbe_tx_ring_t *tx_ring; in ixgbe_init_driver_settings() local
1155 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_init_driver_settings()
1156 tx_ring->index = i; in ixgbe_init_driver_settings()
1157 tx_ring->ixgbe = ixgbe; in ixgbe_init_driver_settings()
1159 tx_ring->tx_recycle = ixgbe_tx_recycle_head_wb; in ixgbe_init_driver_settings()
1161 tx_ring->tx_recycle = ixgbe_tx_recycle_legacy; in ixgbe_init_driver_settings()
1163 tx_ring->ring_size = ixgbe->tx_ring_size; in ixgbe_init_driver_settings()
1164 tx_ring->free_list_size = ixgbe->tx_ring_size + in ixgbe_init_driver_settings()
1189 ixgbe_tx_ring_t *tx_ring; in ixgbe_init_locks() local
1199 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_init_locks()
[all …]
H A Dixgbe_stat.c864 ixgbe_tx_ring_t *tx_ring = (ixgbe_tx_ring_t *)rh; in ixgbe_tx_ring_stat() local
865 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_tx_ring_stat()
873 *val = tx_ring->stat_obytes; in ixgbe_tx_ring_stat()
877 *val = tx_ring->stat_opackets; in ixgbe_tx_ring_stat()
/titanic_44/usr/src/uts/common/io/igb/
H A Digb_tx.c56 igb_tx_ring_t *tx_ring = (igb_tx_ring_t *)arg; in igb_tx_ring_send() local
59 ASSERT(tx_ring != NULL); in igb_tx_ring_send()
61 igb = tx_ring->igb; in igb_tx_ring_send()
71 return ((igb_tx(tx_ring, mp)) ? NULL : mp); in igb_tx_ring_send()
104 igb_tx(igb_tx_ring_t *tx_ring, mblk_t *mp) in igb_tx() argument
106 igb_t *igb = tx_ring->igb; in igb_tx()
165 if (tx_ring->tbd_free < igb->tx_recycle_thresh) in igb_tx()
166 tx_ring->tx_recycle(tx_ring); in igb_tx()
173 if (tx_ring->tbd_free < igb->tx_overload_thresh) { in igb_tx()
174 tx_ring->reschedule = B_TRUE; in igb_tx()
[all …]
H A Digb_buf.c129 igb_tx_ring_t *tx_ring; in igb_alloc_dma() local
150 tx_ring = &igb->tx_rings[i]; in igb_alloc_dma()
152 if (igb_alloc_tbd_ring(tx_ring) != IGB_SUCCESS) in igb_alloc_dma()
155 if (igb_alloc_tcb_lists(tx_ring) != IGB_SUCCESS) in igb_alloc_dma()
176 igb_tx_ring_t *tx_ring; in igb_free_dma() local
194 tx_ring = &igb->tx_rings[i]; in igb_free_dma()
195 igb_free_tbd_ring(tx_ring); in igb_free_dma()
196 igb_free_tcb_lists(tx_ring); in igb_free_dma()
204 igb_alloc_tbd_ring(igb_tx_ring_t *tx_ring) in igb_alloc_tbd_ring() argument
212 igb_t *igb = tx_ring->igb; in igb_alloc_tbd_ring()
[all …]
H A Digb_main.c999 igb_tx_ring_t *tx_ring; in igb_init_driver_settings() local
1050 tx_ring = &igb->tx_rings[i]; in igb_init_driver_settings()
1051 tx_ring->index = i; in igb_init_driver_settings()
1052 tx_ring->igb = igb; in igb_init_driver_settings()
1054 tx_ring->tx_recycle = igb_tx_recycle_head_wb; in igb_init_driver_settings()
1056 tx_ring->tx_recycle = igb_tx_recycle_legacy; in igb_init_driver_settings()
1058 tx_ring->ring_size = igb->tx_ring_size; in igb_init_driver_settings()
1059 tx_ring->free_list_size = igb->tx_ring_size + in igb_init_driver_settings()
1084 igb_tx_ring_t *tx_ring; in igb_init_locks() local
1094 tx_ring = &igb->tx_rings[i]; in igb_init_locks()
[all …]
H A Digb_stat.c304 igb_tx_ring_t *tx_ring = (igb_tx_ring_t *)rh; in igb_tx_ring_stat() local
308 *val = tx_ring->tx_bytes; in igb_tx_ring_stat()
312 *val = tx_ring->tx_pkts; in igb_tx_ring_stat()
/titanic_44/usr/src/uts/common/io/e1000g/
H A De1000g_tx.c175 e1000g_tx_ring_t *tx_ring; in e1000g_send() local
178 tx_ring = Adapter->tx_ring; in e1000g_send()
207 E1000G_STAT(tx_ring->stat_over_size); in e1000g_send()
220 if (tx_ring->tbd_avail < DEFAULT_TX_NO_RESOURCE) { in e1000g_send()
221 (void) e1000g_recycle(tx_ring); in e1000g_send()
222 E1000G_DEBUG_STAT(tx_ring->stat_recycle); in e1000g_send()
224 if (tx_ring->tbd_avail < DEFAULT_TX_NO_RESOURCE) { in e1000g_send()
225 E1000G_DEBUG_STAT(tx_ring->stat_lack_desc); in e1000g_send()
235 E1000G_DEBUG_STAT(tx_ring->stat_under_size); in e1000g_send()
296 E1000G_DEBUG_STAT(tx_ring->stat_lso_header_fail); in e1000g_send()
[all …]
H A De1000g_alloc.c221 e1000g_tx_ring_t *tx_ring; in e1000g_alloc_descriptors() local
235 tx_ring = Adapter->tx_ring; in e1000g_alloc_descriptors()
237 result = e1000g_alloc_tx_descriptors(tx_ring); in e1000g_alloc_descriptors()
245 e1000g_free_tx_descriptors(tx_ring); in e1000g_alloc_descriptors()
255 e1000g_tx_ring_t *tx_ring; in e1000g_free_descriptors() local
258 tx_ring = Adapter->tx_ring; in e1000g_free_descriptors()
261 e1000g_free_tx_descriptors(tx_ring); in e1000g_free_descriptors()
266 e1000g_alloc_tx_descriptors(e1000g_tx_ring_t *tx_ring) in e1000g_alloc_tx_descriptors() argument
279 Adapter = tx_ring->adapter; in e1000g_alloc_tx_descriptors()
312 &tx_ring->tbd_dma_handle); in e1000g_alloc_tx_descriptors()
[all …]
H A De1000g_stat.c154 e1000g_tx_ring_t *tx_ring; in e1000g_update_stats() local
171 tx_ring = Adapter->tx_ring; in e1000g_update_stats()
185 e1000g_ksp->tx_no_swpkt.value.ul = tx_ring->stat_no_swpkt; in e1000g_update_stats()
186 e1000g_ksp->tx_no_desc.value.ul = tx_ring->stat_no_desc; in e1000g_update_stats()
187 e1000g_ksp->tx_send_fail.value.ul = tx_ring->stat_send_fail; in e1000g_update_stats()
188 e1000g_ksp->tx_reschedule.value.ul = tx_ring->stat_reschedule; in e1000g_update_stats()
189 e1000g_ksp->tx_over_size.value.ul = tx_ring->stat_over_size; in e1000g_update_stats()
198 e1000g_ksp->tx_under_size.value.ul = tx_ring->stat_under_size; in e1000g_update_stats()
199 e1000g_ksp->tx_exceed_frags.value.ul = tx_ring->stat_exceed_frags; in e1000g_update_stats()
200 e1000g_ksp->tx_empty_frags.value.ul = tx_ring->stat_empty_frags; in e1000g_update_stats()
[all …]
H A De1000g_main.c436 Adapter->tx_ring->adapter = Adapter; in e1000g_attach()
1250 e1000g_tx_ring_t *tx_ring; in e1000g_init_locks() local
1260 tx_ring = Adapter->tx_ring; in e1000g_init_locks()
1262 mutex_init(&tx_ring->tx_lock, NULL, in e1000g_init_locks()
1264 mutex_init(&tx_ring->usedlist_lock, NULL, in e1000g_init_locks()
1266 mutex_init(&tx_ring->freelist_lock, NULL, in e1000g_init_locks()
1281 e1000g_tx_ring_t *tx_ring; in e1000g_destroy_locks() local
1286 tx_ring = Adapter->tx_ring; in e1000g_destroy_locks()
1287 mutex_destroy(&tx_ring->tx_lock); in e1000g_destroy_locks()
1288 mutex_destroy(&tx_ring->usedlist_lock); in e1000g_destroy_locks()
[all …]
/titanic_44/usr/src/uts/common/io/fibre-channel/fca/qlge/
H A Dqlge_gld.c223 struct tx_ring *tx_ring; in ql_m_getstat() local
254 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat()
255 cur_stats->multixmt += tx_ring->multixmt; in ql_m_getstat()
263 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat()
264 cur_stats->brdcstxmt += tx_ring->brdcstxmt; in ql_m_getstat()
280 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat()
281 cur_stats->obytes += tx_ring->obytes; in ql_m_getstat()
289 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat()
290 cur_stats->opackets += tx_ring->opackets; in ql_m_getstat()
298 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat()
[all …]
H A Dqlge.c1786 struct tx_ring *tx_ring = (struct tx_ring *)arg; in ql_ring_tx() local
1787 qlge_t *qlge = tx_ring->qlge; in ql_ring_tx()
1805 mutex_enter(&tx_ring->tx_lock); in ql_ring_tx()
1807 if (tx_ring->mac_flags != QL_MAC_STARTED) { in ql_ring_tx()
1810 __func__, qlge->instance, tx_ring->mac_flags); in ql_ring_tx()
1811 mutex_exit(&tx_ring->tx_lock); in ql_ring_tx()
1821 if (tx_ring->tx_free_count <= TX_STOP_THRESHOLD) { in ql_ring_tx()
1822 tx_ring->queue_stopped = 1; in ql_ring_tx()
1828 tx_ring->defer++; in ql_ring_tx()
1839 rval = ql_send_common(tx_ring, mp); in ql_ring_tx()
[all …]
/titanic_44/usr/src/grub/grub-0.97/netboot/
H A Depic100.c87 static struct epic_tx_desc tx_ring[TX_RING_SIZE] variable
251 outl(virt_to_le32desc(&tx_ring), ptcdar); in epic100_open()
283 tx_ring[i].status = 0x0000; /* Owned by CPU */ in epic100_init_ring()
284 tx_ring[i].buflength = 0x0000 | cpu_to_le32(TD_STDFLAGS << 16); in epic100_init_ring()
285 tx_ring[i].bufaddr = virt_to_bus(&tx_packet[i * PKT_BUF_SZ]); in epic100_init_ring()
286 tx_ring[i].next = virt_to_le32desc(&tx_ring[i + 1]); in epic100_init_ring()
288 tx_ring[i-1].next = virt_to_le32desc(&tx_ring[0]); in epic100_init_ring()
311 if ((tx_ring[entry].status & TRING_OWN) == TRING_OWN) { in epic100_transmit()
313 tx_ring[entry].status); in epic100_transmit()
337 tx_ring[entry].buflength |= cpu_to_le32(len); in epic100_transmit()
[all …]
H A Dpcnet32.c223 u32 tx_ring; member
237 static struct pcnet32_tx_head tx_ring[TX_RING_SIZE] variable
420 tx_ring[i].base = 0; in pcnet32_init_ring()
421 tx_ring[i].status = 0; in pcnet32_init_ring()
430 lp->init_block.tx_ring = (u32) virt_to_le32desc(&tx_ring[0]); in pcnet32_init_ring()
608 tx_ring[entry].length = le16_to_cpu(-s); in pcnet32_transmit()
609 tx_ring[entry].misc = 0x00000000; in pcnet32_transmit()
610 tx_ring[entry].base = (u32) virt_to_le32desc(ptxb); in pcnet32_transmit()
613 tx_ring[entry].status = le16_to_cpu(status); in pcnet32_transmit()
623 ((short) le16_to_cpu(tx_ring[entry].status) < 0)); in pcnet32_transmit()
[all …]
H A Dw89c840.c232 struct w840_tx_desc tx_ring[TX_RING_SIZE]; member
314 writel(virt_to_bus(w840private.tx_ring), ioaddr + TxRingPtr); in w89c840_reset()
513 w840private.tx_ring[entry].buffer1 = virt_to_le32desc(tx_packet); in w89c840_transmit()
515 w840private.tx_ring[entry].length = (DescWholePkt | (u32) s); in w89c840_transmit()
517 w840private.tx_ring[entry].length |= (DescIntr | DescEndRing); in w89c840_transmit()
518 w840private.tx_ring[entry].status = (DescOwn); in w89c840_transmit()
539 transmit_status = w840private.tx_ring[entry].status; in w89c840_transmit()
555 transmit_status = w840private.tx_ring[entry].status; in w89c840_transmit()
566 w840private.tx_ring[entry].status); in w89c840_transmit()
574 printf("winbond-840 : transmission TIMEOUT : status %X\n", w840private.tx_ring[entry].status); in w89c840_transmit()
[all …]
H A Dsundance.c254 static struct netdev_desc tx_ring[TX_RING_SIZE]; variable
371 tx_ring[0].status = 0x00000000; in init_ring()
372 tx_ring[0].addr = virt_to_bus(&txb[0]); in init_ring()
373 tx_ring[0].next_desc = 0; /* virt_to_bus(&tx_ring[1]); */ in init_ring()
376 tx_ring[1].status = 0x00000000; in init_ring()
377 tx_ring[1].addr = 0; /*virt_to_bus(&txb[0]); */ in init_ring()
378 tx_ring[1].next_desc = 0; in init_ring()
382 tx_ring[1].length = cpu_to_le32(LastFrag | PKT_BUF_SZ); in init_ring()
538 tx_ring[0].length = cpu_to_le32(s | LastFrag); in sundance_transmit()
539 tx_ring[0].status = cpu_to_le32(0x00000001); in sundance_transmit()
[all …]
H A Dvia-rhine.c632 struct rhine_tx_desc *tx_ring; member
701 tp->tx_ring[i].tx_status.lw = 0; in rhine_init_ring()
702 tp->tx_ring[i].tx_ctrl.lw = 0x00e08000; in rhine_init_ring()
703 tp->tx_ring[i].buf_addr_1 = virt_to_bus (tp->tx_buffs[i]); in rhine_init_ring()
704 tp->tx_ring[i].buf_addr_2 = virt_to_bus (&tp->tx_ring[i + 1]); in rhine_init_ring()
708 tp->tx_ring[i - 1].buf_addr_2 = virt_to_bus (&tp->tx_ring[0]); in rhine_init_ring()
1125 tp->tx_ring = (struct rhine_tx_desc *) bus_to_virt (j); in rhine_reset()
1161 outl (virt_to_bus (tp->tx_ring), dwCurrentTxDescAddr); in rhine_reset()
1273 tp->tx_ring[entry].tx_ctrl.bits.tx_buf_size = s; in rhine_transmit()
1275 tp->tx_ring[entry].tx_status.bits.own_bit = 1; in rhine_transmit()
[all …]
H A Dns83820.c374 static struct ring_desc tx_ring[NR_TX_DESC] variable
576 tx_ring[i].link = 0; in ns83820_reset()
577 tx_ring[i].bufptr = 0; in ns83820_reset()
578 tx_ring[i].cmdsts = cpu_to_le32(0); in ns83820_reset()
579 tx_ring[i].extsts = cpu_to_le32(0); in ns83820_reset()
745 tx_ring[cur_tx].bufptr = virt_to_le32desc(&txb); in ns83820_transmit()
746 tx_ring[cur_tx].extsts = cpu_to_le32(extsts); in ns83820_transmit()
750 tx_ring[cur_tx].cmdsts = cpu_to_le32(cmdsts); in ns83820_transmit()
752 writel(virt_to_le32desc(&tx_ring[0]), ns->base + TXDP); in ns83820_transmit()
H A Dtulip.c398 static struct tulip_tx_desc tx_ring[TX_RING_SIZE] __attribute__ ((aligned(4))); variable
920 tx_ring[0].status = 0x00000000; in tulip_init_ring()
921 tx_ring[0].buffer1 = virt_to_le32desc(&txb[0]); in tulip_init_ring()
922 tx_ring[0].buffer2 = virt_to_le32desc(&tx_ring[1]); in tulip_init_ring()
926 tx_ring[1].status = 0x00000000; in tulip_init_ring()
927 tx_ring[1].buffer1 = virt_to_le32desc(&txb[0]); in tulip_init_ring()
928 tx_ring[1].buffer2 = virt_to_le32desc(&tx_ring[0]); in tulip_init_ring()
931 tx_ring[1].length = cpu_to_le32(DESC_RING_WRAP | BUFLEN); in tulip_init_ring()
1028 tx_ring[0].length = cpu_to_le32(tx_flags); in tulip_reset()
1029 tx_ring[0].buffer1 = virt_to_le32desc(&txb[0]); in tulip_reset()
[all …]
H A Dforcedeth.c294 static struct ring_desc tx_ring[TX_RING]; variable
334 struct ring_desc *tx_ring; member
589 tx_ring[i].Flags = 0; in init_ring()
744 writel((u32) virt_to_le32desc(&tx_ring[0]), in forcedeth_reset()
863 tx_ring[nr].PacketBuffer = (u32) virt_to_le32desc(ptxb); in forcedeth_transmit()
864 tx_ring[nr].Length = cpu_to_le16(s - 1); in forcedeth_transmit()
867 tx_ring[nr].Flags = np->tx_flags; in forcedeth_transmit()
871 tx_ring[nr].Flags = np->tx_flags; in forcedeth_transmit()
/titanic_44/usr/src/uts/common/io/xge/drv/
H A Dxgell.c1476 xgell_tx_ring_t *tx_ring; in xgell_fill_ring() local
1482 tx_ring = lldev->tx_ring + index; in xgell_fill_ring()
1483 tx_ring->ring_handle = rh; in xgell_fill_ring()
1485 infop->mri_driver = (mac_ring_driver_t)tx_ring; in xgell_fill_ring()
1751 xgell_tx_ring_open(xgell_tx_ring_t *tx_ring) in xgell_tx_ring_open() argument
1755 xgelldev_t *lldev = tx_ring->lldev; in xgell_tx_ring_open()
1757 if (tx_ring->live) in xgell_tx_ring_open()
1760 attr.post_qid = tx_ring->index; in xgell_tx_ring_open()
1768 attr.userdata = tx_ring; in xgell_tx_ring_open()
1770 status = xge_hal_channel_open(lldev->devh, &attr, &tx_ring->channelh, in xgell_tx_ring_open()
[all …]
/titanic_44/usr/src/uts/common/sys/fibre-channel/fca/qlge/
H A Dqlge.h190 #define RESUME_TX(tx_ring) mac_tx_update(tx_ring->qlge->mh); argument
444 struct tx_ring { struct
800 struct tx_ring tx_ring[MAX_TX_RINGS]; member
865 extern int ql_send_common(struct tx_ring *, mblk_t *);
/titanic_44/usr/src/uts/common/io/aggr/
H A Daggr_grp.c1941 aggr_pseudo_tx_ring_t *tx_ring = (aggr_pseudo_tx_ring_t *)rdriver; in aggr_tx_ring_stat() local
1943 if (tx_ring->atr_hw_rh != NULL) { in aggr_tx_ring_stat()
1944 *val = mac_pseudo_tx_ring_stat_get(tx_ring->atr_hw_rh, stat); in aggr_tx_ring_stat()
1946 aggr_port_t *port = tx_ring->atr_port; in aggr_tx_ring_stat()
2282 aggr_pseudo_tx_ring_t *tx_ring; in aggr_fill_ring() local
2287 tx_ring = &tx_group->atg_rings[index]; in aggr_fill_ring()
2288 tx_ring->atr_rh = rh; in aggr_fill_ring()
2290 infop->mri_driver = (mac_ring_driver_t)tx_ring; in aggr_fill_ring()
2300 if (tx_ring->atr_hw_rh != NULL) { in aggr_fill_ring()
2302 mac_hwring_getinfo(tx_ring->atr_hw_rh); in aggr_fill_ring()

12