/titanic_44/usr/src/uts/common/io/ixgbe/ |
H A D | ixgbe_tx.c | 87 ixgbe_tx_ring_t *tx_ring = (ixgbe_tx_ring_t *)arg; in ixgbe_ring_tx() local 88 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_ring_tx() 156 if (tx_ring->tbd_free < ixgbe->tx_recycle_thresh) { in ixgbe_ring_tx() 157 tx_ring->tx_recycle(tx_ring); in ixgbe_ring_tx() 165 if (tx_ring->tbd_free < ixgbe->tx_overload_thresh) { in ixgbe_ring_tx() 166 tx_ring->reschedule = B_TRUE; in ixgbe_ring_tx() 167 tx_ring->stat_overload++; in ixgbe_ring_tx() 217 tx_ring->stat_lso_header_fail++; in ixgbe_ring_tx() 296 tcb = ixgbe_get_free_list(tx_ring); in ixgbe_ring_tx() 299 tx_ring->stat_fail_no_tcb++; in ixgbe_ring_tx() [all …]
|
H A D | ixgbe_buf.c | 127 ixgbe_tx_ring_t *tx_ring; in ixgbe_alloc_dma() local 148 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_alloc_dma() 150 if (ixgbe_alloc_tbd_ring(tx_ring) != IXGBE_SUCCESS) in ixgbe_alloc_dma() 153 if (ixgbe_alloc_tcb_lists(tx_ring) != IXGBE_SUCCESS) in ixgbe_alloc_dma() 173 ixgbe_tx_ring_t *tx_ring; in ixgbe_free_dma() local 191 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_free_dma() 192 ixgbe_free_tbd_ring(tx_ring); in ixgbe_free_dma() 193 ixgbe_free_tcb_lists(tx_ring); in ixgbe_free_dma() 309 ixgbe_alloc_tbd_ring(ixgbe_tx_ring_t *tx_ring) in ixgbe_alloc_tbd_ring() argument 317 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_alloc_tbd_ring() [all …]
|
H A D | ixgbe_main.c | 1100 ixgbe_tx_ring_t *tx_ring; in ixgbe_init_driver_settings() local 1155 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_init_driver_settings() 1156 tx_ring->index = i; in ixgbe_init_driver_settings() 1157 tx_ring->ixgbe = ixgbe; in ixgbe_init_driver_settings() 1159 tx_ring->tx_recycle = ixgbe_tx_recycle_head_wb; in ixgbe_init_driver_settings() 1161 tx_ring->tx_recycle = ixgbe_tx_recycle_legacy; in ixgbe_init_driver_settings() 1163 tx_ring->ring_size = ixgbe->tx_ring_size; in ixgbe_init_driver_settings() 1164 tx_ring->free_list_size = ixgbe->tx_ring_size + in ixgbe_init_driver_settings() 1189 ixgbe_tx_ring_t *tx_ring; in ixgbe_init_locks() local 1199 tx_ring = &ixgbe->tx_rings[i]; in ixgbe_init_locks() [all …]
|
H A D | ixgbe_stat.c | 864 ixgbe_tx_ring_t *tx_ring = (ixgbe_tx_ring_t *)rh; in ixgbe_tx_ring_stat() local 865 ixgbe_t *ixgbe = tx_ring->ixgbe; in ixgbe_tx_ring_stat() 873 *val = tx_ring->stat_obytes; in ixgbe_tx_ring_stat() 877 *val = tx_ring->stat_opackets; in ixgbe_tx_ring_stat()
|
/titanic_44/usr/src/uts/common/io/igb/ |
H A D | igb_tx.c | 56 igb_tx_ring_t *tx_ring = (igb_tx_ring_t *)arg; in igb_tx_ring_send() local 59 ASSERT(tx_ring != NULL); in igb_tx_ring_send() 61 igb = tx_ring->igb; in igb_tx_ring_send() 71 return ((igb_tx(tx_ring, mp)) ? NULL : mp); in igb_tx_ring_send() 104 igb_tx(igb_tx_ring_t *tx_ring, mblk_t *mp) in igb_tx() argument 106 igb_t *igb = tx_ring->igb; in igb_tx() 165 if (tx_ring->tbd_free < igb->tx_recycle_thresh) in igb_tx() 166 tx_ring->tx_recycle(tx_ring); in igb_tx() 173 if (tx_ring->tbd_free < igb->tx_overload_thresh) { in igb_tx() 174 tx_ring->reschedule = B_TRUE; in igb_tx() [all …]
|
H A D | igb_buf.c | 129 igb_tx_ring_t *tx_ring; in igb_alloc_dma() local 150 tx_ring = &igb->tx_rings[i]; in igb_alloc_dma() 152 if (igb_alloc_tbd_ring(tx_ring) != IGB_SUCCESS) in igb_alloc_dma() 155 if (igb_alloc_tcb_lists(tx_ring) != IGB_SUCCESS) in igb_alloc_dma() 176 igb_tx_ring_t *tx_ring; in igb_free_dma() local 194 tx_ring = &igb->tx_rings[i]; in igb_free_dma() 195 igb_free_tbd_ring(tx_ring); in igb_free_dma() 196 igb_free_tcb_lists(tx_ring); in igb_free_dma() 204 igb_alloc_tbd_ring(igb_tx_ring_t *tx_ring) in igb_alloc_tbd_ring() argument 212 igb_t *igb = tx_ring->igb; in igb_alloc_tbd_ring() [all …]
|
H A D | igb_main.c | 999 igb_tx_ring_t *tx_ring; in igb_init_driver_settings() local 1050 tx_ring = &igb->tx_rings[i]; in igb_init_driver_settings() 1051 tx_ring->index = i; in igb_init_driver_settings() 1052 tx_ring->igb = igb; in igb_init_driver_settings() 1054 tx_ring->tx_recycle = igb_tx_recycle_head_wb; in igb_init_driver_settings() 1056 tx_ring->tx_recycle = igb_tx_recycle_legacy; in igb_init_driver_settings() 1058 tx_ring->ring_size = igb->tx_ring_size; in igb_init_driver_settings() 1059 tx_ring->free_list_size = igb->tx_ring_size + in igb_init_driver_settings() 1084 igb_tx_ring_t *tx_ring; in igb_init_locks() local 1094 tx_ring = &igb->tx_rings[i]; in igb_init_locks() [all …]
|
H A D | igb_stat.c | 304 igb_tx_ring_t *tx_ring = (igb_tx_ring_t *)rh; in igb_tx_ring_stat() local 308 *val = tx_ring->tx_bytes; in igb_tx_ring_stat() 312 *val = tx_ring->tx_pkts; in igb_tx_ring_stat()
|
/titanic_44/usr/src/uts/common/io/e1000g/ |
H A D | e1000g_tx.c | 175 e1000g_tx_ring_t *tx_ring; in e1000g_send() local 178 tx_ring = Adapter->tx_ring; in e1000g_send() 207 E1000G_STAT(tx_ring->stat_over_size); in e1000g_send() 220 if (tx_ring->tbd_avail < DEFAULT_TX_NO_RESOURCE) { in e1000g_send() 221 (void) e1000g_recycle(tx_ring); in e1000g_send() 222 E1000G_DEBUG_STAT(tx_ring->stat_recycle); in e1000g_send() 224 if (tx_ring->tbd_avail < DEFAULT_TX_NO_RESOURCE) { in e1000g_send() 225 E1000G_DEBUG_STAT(tx_ring->stat_lack_desc); in e1000g_send() 235 E1000G_DEBUG_STAT(tx_ring->stat_under_size); in e1000g_send() 296 E1000G_DEBUG_STAT(tx_ring->stat_lso_header_fail); in e1000g_send() [all …]
|
H A D | e1000g_alloc.c | 221 e1000g_tx_ring_t *tx_ring; in e1000g_alloc_descriptors() local 235 tx_ring = Adapter->tx_ring; in e1000g_alloc_descriptors() 237 result = e1000g_alloc_tx_descriptors(tx_ring); in e1000g_alloc_descriptors() 245 e1000g_free_tx_descriptors(tx_ring); in e1000g_alloc_descriptors() 255 e1000g_tx_ring_t *tx_ring; in e1000g_free_descriptors() local 258 tx_ring = Adapter->tx_ring; in e1000g_free_descriptors() 261 e1000g_free_tx_descriptors(tx_ring); in e1000g_free_descriptors() 266 e1000g_alloc_tx_descriptors(e1000g_tx_ring_t *tx_ring) in e1000g_alloc_tx_descriptors() argument 279 Adapter = tx_ring->adapter; in e1000g_alloc_tx_descriptors() 312 &tx_ring->tbd_dma_handle); in e1000g_alloc_tx_descriptors() [all …]
|
H A D | e1000g_stat.c | 154 e1000g_tx_ring_t *tx_ring; in e1000g_update_stats() local 171 tx_ring = Adapter->tx_ring; in e1000g_update_stats() 185 e1000g_ksp->tx_no_swpkt.value.ul = tx_ring->stat_no_swpkt; in e1000g_update_stats() 186 e1000g_ksp->tx_no_desc.value.ul = tx_ring->stat_no_desc; in e1000g_update_stats() 187 e1000g_ksp->tx_send_fail.value.ul = tx_ring->stat_send_fail; in e1000g_update_stats() 188 e1000g_ksp->tx_reschedule.value.ul = tx_ring->stat_reschedule; in e1000g_update_stats() 189 e1000g_ksp->tx_over_size.value.ul = tx_ring->stat_over_size; in e1000g_update_stats() 198 e1000g_ksp->tx_under_size.value.ul = tx_ring->stat_under_size; in e1000g_update_stats() 199 e1000g_ksp->tx_exceed_frags.value.ul = tx_ring->stat_exceed_frags; in e1000g_update_stats() 200 e1000g_ksp->tx_empty_frags.value.ul = tx_ring->stat_empty_frags; in e1000g_update_stats() [all …]
|
H A D | e1000g_main.c | 436 Adapter->tx_ring->adapter = Adapter; in e1000g_attach() 1250 e1000g_tx_ring_t *tx_ring; in e1000g_init_locks() local 1260 tx_ring = Adapter->tx_ring; in e1000g_init_locks() 1262 mutex_init(&tx_ring->tx_lock, NULL, in e1000g_init_locks() 1264 mutex_init(&tx_ring->usedlist_lock, NULL, in e1000g_init_locks() 1266 mutex_init(&tx_ring->freelist_lock, NULL, in e1000g_init_locks() 1281 e1000g_tx_ring_t *tx_ring; in e1000g_destroy_locks() local 1286 tx_ring = Adapter->tx_ring; in e1000g_destroy_locks() 1287 mutex_destroy(&tx_ring->tx_lock); in e1000g_destroy_locks() 1288 mutex_destroy(&tx_ring->usedlist_lock); in e1000g_destroy_locks() [all …]
|
/titanic_44/usr/src/uts/common/io/fibre-channel/fca/qlge/ |
H A D | qlge_gld.c | 223 struct tx_ring *tx_ring; in ql_m_getstat() local 254 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat() 255 cur_stats->multixmt += tx_ring->multixmt; in ql_m_getstat() 263 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat() 264 cur_stats->brdcstxmt += tx_ring->brdcstxmt; in ql_m_getstat() 280 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat() 281 cur_stats->obytes += tx_ring->obytes; in ql_m_getstat() 289 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat() 290 cur_stats->opackets += tx_ring->opackets; in ql_m_getstat() 298 tx_ring = &qlge->tx_ring[i]; in ql_m_getstat() [all …]
|
H A D | qlge.c | 1786 struct tx_ring *tx_ring = (struct tx_ring *)arg; in ql_ring_tx() local 1787 qlge_t *qlge = tx_ring->qlge; in ql_ring_tx() 1805 mutex_enter(&tx_ring->tx_lock); in ql_ring_tx() 1807 if (tx_ring->mac_flags != QL_MAC_STARTED) { in ql_ring_tx() 1810 __func__, qlge->instance, tx_ring->mac_flags); in ql_ring_tx() 1811 mutex_exit(&tx_ring->tx_lock); in ql_ring_tx() 1821 if (tx_ring->tx_free_count <= TX_STOP_THRESHOLD) { in ql_ring_tx() 1822 tx_ring->queue_stopped = 1; in ql_ring_tx() 1828 tx_ring->defer++; in ql_ring_tx() 1839 rval = ql_send_common(tx_ring, mp); in ql_ring_tx() [all …]
|
/titanic_44/usr/src/grub/grub-0.97/netboot/ |
H A D | epic100.c | 87 static struct epic_tx_desc tx_ring[TX_RING_SIZE] variable 251 outl(virt_to_le32desc(&tx_ring), ptcdar); in epic100_open() 283 tx_ring[i].status = 0x0000; /* Owned by CPU */ in epic100_init_ring() 284 tx_ring[i].buflength = 0x0000 | cpu_to_le32(TD_STDFLAGS << 16); in epic100_init_ring() 285 tx_ring[i].bufaddr = virt_to_bus(&tx_packet[i * PKT_BUF_SZ]); in epic100_init_ring() 286 tx_ring[i].next = virt_to_le32desc(&tx_ring[i + 1]); in epic100_init_ring() 288 tx_ring[i-1].next = virt_to_le32desc(&tx_ring[0]); in epic100_init_ring() 311 if ((tx_ring[entry].status & TRING_OWN) == TRING_OWN) { in epic100_transmit() 313 tx_ring[entry].status); in epic100_transmit() 337 tx_ring[entry].buflength |= cpu_to_le32(len); in epic100_transmit() [all …]
|
H A D | pcnet32.c | 223 u32 tx_ring; member 237 static struct pcnet32_tx_head tx_ring[TX_RING_SIZE] variable 420 tx_ring[i].base = 0; in pcnet32_init_ring() 421 tx_ring[i].status = 0; in pcnet32_init_ring() 430 lp->init_block.tx_ring = (u32) virt_to_le32desc(&tx_ring[0]); in pcnet32_init_ring() 608 tx_ring[entry].length = le16_to_cpu(-s); in pcnet32_transmit() 609 tx_ring[entry].misc = 0x00000000; in pcnet32_transmit() 610 tx_ring[entry].base = (u32) virt_to_le32desc(ptxb); in pcnet32_transmit() 613 tx_ring[entry].status = le16_to_cpu(status); in pcnet32_transmit() 623 ((short) le16_to_cpu(tx_ring[entry].status) < 0)); in pcnet32_transmit() [all …]
|
H A D | w89c840.c | 232 struct w840_tx_desc tx_ring[TX_RING_SIZE]; member 314 writel(virt_to_bus(w840private.tx_ring), ioaddr + TxRingPtr); in w89c840_reset() 513 w840private.tx_ring[entry].buffer1 = virt_to_le32desc(tx_packet); in w89c840_transmit() 515 w840private.tx_ring[entry].length = (DescWholePkt | (u32) s); in w89c840_transmit() 517 w840private.tx_ring[entry].length |= (DescIntr | DescEndRing); in w89c840_transmit() 518 w840private.tx_ring[entry].status = (DescOwn); in w89c840_transmit() 539 transmit_status = w840private.tx_ring[entry].status; in w89c840_transmit() 555 transmit_status = w840private.tx_ring[entry].status; in w89c840_transmit() 566 w840private.tx_ring[entry].status); in w89c840_transmit() 574 printf("winbond-840 : transmission TIMEOUT : status %X\n", w840private.tx_ring[entry].status); in w89c840_transmit() [all …]
|
H A D | sundance.c | 254 static struct netdev_desc tx_ring[TX_RING_SIZE]; variable 371 tx_ring[0].status = 0x00000000; in init_ring() 372 tx_ring[0].addr = virt_to_bus(&txb[0]); in init_ring() 373 tx_ring[0].next_desc = 0; /* virt_to_bus(&tx_ring[1]); */ in init_ring() 376 tx_ring[1].status = 0x00000000; in init_ring() 377 tx_ring[1].addr = 0; /*virt_to_bus(&txb[0]); */ in init_ring() 378 tx_ring[1].next_desc = 0; in init_ring() 382 tx_ring[1].length = cpu_to_le32(LastFrag | PKT_BUF_SZ); in init_ring() 538 tx_ring[0].length = cpu_to_le32(s | LastFrag); in sundance_transmit() 539 tx_ring[0].status = cpu_to_le32(0x00000001); in sundance_transmit() [all …]
|
H A D | via-rhine.c | 632 struct rhine_tx_desc *tx_ring; member 701 tp->tx_ring[i].tx_status.lw = 0; in rhine_init_ring() 702 tp->tx_ring[i].tx_ctrl.lw = 0x00e08000; in rhine_init_ring() 703 tp->tx_ring[i].buf_addr_1 = virt_to_bus (tp->tx_buffs[i]); in rhine_init_ring() 704 tp->tx_ring[i].buf_addr_2 = virt_to_bus (&tp->tx_ring[i + 1]); in rhine_init_ring() 708 tp->tx_ring[i - 1].buf_addr_2 = virt_to_bus (&tp->tx_ring[0]); in rhine_init_ring() 1125 tp->tx_ring = (struct rhine_tx_desc *) bus_to_virt (j); in rhine_reset() 1161 outl (virt_to_bus (tp->tx_ring), dwCurrentTxDescAddr); in rhine_reset() 1273 tp->tx_ring[entry].tx_ctrl.bits.tx_buf_size = s; in rhine_transmit() 1275 tp->tx_ring[entry].tx_status.bits.own_bit = 1; in rhine_transmit() [all …]
|
H A D | ns83820.c | 374 static struct ring_desc tx_ring[NR_TX_DESC] variable 576 tx_ring[i].link = 0; in ns83820_reset() 577 tx_ring[i].bufptr = 0; in ns83820_reset() 578 tx_ring[i].cmdsts = cpu_to_le32(0); in ns83820_reset() 579 tx_ring[i].extsts = cpu_to_le32(0); in ns83820_reset() 745 tx_ring[cur_tx].bufptr = virt_to_le32desc(&txb); in ns83820_transmit() 746 tx_ring[cur_tx].extsts = cpu_to_le32(extsts); in ns83820_transmit() 750 tx_ring[cur_tx].cmdsts = cpu_to_le32(cmdsts); in ns83820_transmit() 752 writel(virt_to_le32desc(&tx_ring[0]), ns->base + TXDP); in ns83820_transmit()
|
H A D | tulip.c | 398 static struct tulip_tx_desc tx_ring[TX_RING_SIZE] __attribute__ ((aligned(4))); variable 920 tx_ring[0].status = 0x00000000; in tulip_init_ring() 921 tx_ring[0].buffer1 = virt_to_le32desc(&txb[0]); in tulip_init_ring() 922 tx_ring[0].buffer2 = virt_to_le32desc(&tx_ring[1]); in tulip_init_ring() 926 tx_ring[1].status = 0x00000000; in tulip_init_ring() 927 tx_ring[1].buffer1 = virt_to_le32desc(&txb[0]); in tulip_init_ring() 928 tx_ring[1].buffer2 = virt_to_le32desc(&tx_ring[0]); in tulip_init_ring() 931 tx_ring[1].length = cpu_to_le32(DESC_RING_WRAP | BUFLEN); in tulip_init_ring() 1028 tx_ring[0].length = cpu_to_le32(tx_flags); in tulip_reset() 1029 tx_ring[0].buffer1 = virt_to_le32desc(&txb[0]); in tulip_reset() [all …]
|
H A D | forcedeth.c | 294 static struct ring_desc tx_ring[TX_RING]; variable 334 struct ring_desc *tx_ring; member 589 tx_ring[i].Flags = 0; in init_ring() 744 writel((u32) virt_to_le32desc(&tx_ring[0]), in forcedeth_reset() 863 tx_ring[nr].PacketBuffer = (u32) virt_to_le32desc(ptxb); in forcedeth_transmit() 864 tx_ring[nr].Length = cpu_to_le16(s - 1); in forcedeth_transmit() 867 tx_ring[nr].Flags = np->tx_flags; in forcedeth_transmit() 871 tx_ring[nr].Flags = np->tx_flags; in forcedeth_transmit()
|
/titanic_44/usr/src/uts/common/io/xge/drv/ |
H A D | xgell.c | 1476 xgell_tx_ring_t *tx_ring; in xgell_fill_ring() local 1482 tx_ring = lldev->tx_ring + index; in xgell_fill_ring() 1483 tx_ring->ring_handle = rh; in xgell_fill_ring() 1485 infop->mri_driver = (mac_ring_driver_t)tx_ring; in xgell_fill_ring() 1751 xgell_tx_ring_open(xgell_tx_ring_t *tx_ring) in xgell_tx_ring_open() argument 1755 xgelldev_t *lldev = tx_ring->lldev; in xgell_tx_ring_open() 1757 if (tx_ring->live) in xgell_tx_ring_open() 1760 attr.post_qid = tx_ring->index; in xgell_tx_ring_open() 1768 attr.userdata = tx_ring; in xgell_tx_ring_open() 1770 status = xge_hal_channel_open(lldev->devh, &attr, &tx_ring->channelh, in xgell_tx_ring_open() [all …]
|
/titanic_44/usr/src/uts/common/sys/fibre-channel/fca/qlge/ |
H A D | qlge.h | 190 #define RESUME_TX(tx_ring) mac_tx_update(tx_ring->qlge->mh); argument 444 struct tx_ring { struct 800 struct tx_ring tx_ring[MAX_TX_RINGS]; member 865 extern int ql_send_common(struct tx_ring *, mblk_t *);
|
/titanic_44/usr/src/uts/common/io/aggr/ |
H A D | aggr_grp.c | 1941 aggr_pseudo_tx_ring_t *tx_ring = (aggr_pseudo_tx_ring_t *)rdriver; in aggr_tx_ring_stat() local 1943 if (tx_ring->atr_hw_rh != NULL) { in aggr_tx_ring_stat() 1944 *val = mac_pseudo_tx_ring_stat_get(tx_ring->atr_hw_rh, stat); in aggr_tx_ring_stat() 1946 aggr_port_t *port = tx_ring->atr_port; in aggr_tx_ring_stat() 2282 aggr_pseudo_tx_ring_t *tx_ring; in aggr_fill_ring() local 2287 tx_ring = &tx_group->atg_rings[index]; in aggr_fill_ring() 2288 tx_ring->atr_rh = rh; in aggr_fill_ring() 2290 infop->mri_driver = (mac_ring_driver_t)tx_ring; in aggr_fill_ring() 2300 if (tx_ring->atr_hw_rh != NULL) { in aggr_fill_ring() 2302 mac_hwring_getinfo(tx_ring->atr_hw_rh); in aggr_fill_ring()
|