| /linux/drivers/net/ethernet/amd/ |
| H A D | ariadne.c | 92 int cur_tx, cur_rx; /* The next free ring entry */ member 125 priv->cur_rx = priv->cur_tx = 0; in ariadne_init_ring() 308 while (dirty_tx < priv->cur_tx) { in ariadne_interrupt() 345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { in ariadne_interrupt() 347 dirty_tx, priv->cur_tx, in ariadne_interrupt() 354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { in ariadne_interrupt() 576 entry = priv->cur_tx % TX_RING_SIZE; in ariadne_start_xmit() 596 priv->cur_tx++; in ariadne_start_xmit() 597 if ((priv->cur_tx >= TX_RING_SIZE) && in ariadne_start_xmit() 601 priv->cur_tx, priv->dirty_tx); in ariadne_start_xmit() [all …]
|
| H A D | atarilance.c | 223 int cur_rx, cur_tx; /* The next free ring entry */ member 689 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring() 748 lp->dirty_tx, lp->cur_tx, in lance_tx_timeout() 815 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit() 829 lp->cur_tx++; in lance_start_xmit() 830 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { in lance_start_xmit() 831 lp->cur_tx -= TX_RING_SIZE; in lance_start_xmit() 885 while( dirty_tx < lp->cur_tx) { in lance_interrupt() 921 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt() 924 dirty_tx, lp->cur_tx, lp->tx_full )); in lance_interrupt() [all …]
|
| H A D | lance.c | 249 int cur_rx, cur_tx; /* The next free ring entry */ member 872 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring() 936 lp->dirty_tx, lp->cur_tx, netif_queue_stopped(dev) ? " (full)" : "", in lance_tx_timeout() 976 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit() 1011 lp->cur_tx++; in lance_start_xmit() 1017 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) in lance_start_xmit() 1056 while (dirty_tx < lp->cur_tx) { in lance_interrupt() 1100 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt() 1102 dirty_tx, lp->cur_tx, in lance_interrupt() 1110 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) in lance_interrupt()
|
| /linux/drivers/net/ethernet/packetengines/ |
| H A D | yellowfin.c | 326 unsigned int cur_tx, dirty_tx; member 693 yp->cur_tx, yp->dirty_tx, in yellowfin_tx_timeout() 718 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_tx_timeout() 732 yp->cur_rx = yp->cur_tx = 0; in yellowfin_init_ring() 826 entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit() 857 yp->cur_tx++; in yellowfin_start_xmit() 866 yp->cur_tx++; in yellowfin_start_xmit() 868 unsigned next_entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit() 883 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_start_xmit() 890 yp->cur_tx, entry); in yellowfin_start_xmit() [all …]
|
| H A D | hamachi.c | 495 unsigned int cur_tx, dirty_tx; member 1001 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++) { in hamachi_tx() 1116 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_tx_timeout() 1172 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_init_ring() 1229 printk(KERN_WARNING "%s: Hamachi transmit queue full at slot %d.\n",dev->name, hmp->cur_tx); in hamachi_start_xmit() 1243 entry = hmp->cur_tx % TX_RING_SIZE; in hamachi_start_xmit() 1267 hmp->cur_tx++; in hamachi_start_xmit() 1286 if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4)) in hamachi_start_xmit() 1295 dev->name, hmp->cur_tx, entry); in hamachi_start_xmit() 1340 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++){ in hamachi_interrupt() [all …]
|
| /linux/drivers/net/arcnet/ |
| H A D | arcnet.c | 567 lp->cur_tx = lp->next_tx = -1; in arcnet_open() 709 lp->hw.status(dev), lp->cur_tx, lp->next_tx, skb->len, skb->protocol); in arcnet_send_packet() 790 lp->hw.status(dev), lp->intmask, lp->next_tx, lp->cur_tx); in go_tx() 792 if (lp->cur_tx != -1 || lp->next_tx == -1) in go_tx() 798 lp->cur_tx = lp->next_tx; in go_tx() 802 lp->hw.command(dev, TXcmd | (lp->cur_tx << 3)); in go_tx() 828 lp->hw.command(dev, NOTXcmd | (lp->cur_tx << 3)); in arcnet_timeout() 845 if (lp->cur_tx == -1) in arcnet_timeout() 965 if (lp->cur_tx != -1 && !lp->timed_out) { in arcnet_interrupt() 990 if (lp->cur_tx != -1) in arcnet_interrupt() [all …]
|
| /linux/drivers/net/can/spi/mcp251xfd/ |
| H A D | mcp251xfd-ethtool.c | 44 layout.cur_tx != priv->tx->obj_num) && in mcp251xfd_ring_set_ringparam() 50 priv->tx->obj_num = layout.cur_tx; in mcp251xfd_ring_set_ringparam() 112 priv->tx->obj_num = layout.cur_tx; in mcp251xfd_ring_set_coalesce()
|
| H A D | mcp251xfd-ram.c | 153 layout->cur_tx = num_tx; in can_ram_get_layout() 158 layout->cur_tx = layout->default_tx; in can_ram_get_layout()
|
| H A D | mcp251xfd-ram.h | 50 u8 cur_tx; member
|
| /linux/drivers/net/ethernet/dec/tulip/ |
| H A D | tulip_core.c | 323 tp->cur_rx = tp->cur_tx = 0; in tulip_up() 356 tp->tx_buffers[tp->cur_tx].skb = NULL; in tulip_up() 357 tp->tx_buffers[tp->cur_tx].mapping = mapping; in tulip_up() 360 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up() 361 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up() 362 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up() 364 tp->cur_tx++; in tulip_up() 664 entry = tp->cur_tx % TX_RING_SIZE; in tulip_start_xmit() 672 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ in tulip_start_xmit() 674 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { in tulip_start_xmit() [all …]
|
| H A D | winbond-840.c | 304 unsigned int cur_tx, dirty_tx; member 821 np->tx_q_bytes = np->dirty_tx = np->cur_tx = 0; in init_rxtx_rings() 932 np->cur_tx, np->dirty_tx, np->tx_full, np->tx_q_bytes); in tx_timeout() 992 entry = np->cur_tx % TX_RING_SIZE; in start_tx() 1021 np->cur_tx++; in start_tx() 1030 if (np->cur_tx - np->dirty_tx > TX_QUEUE_LEN || in start_tx() 1040 np->cur_tx, entry); in start_tx() 1048 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { in netdev_tx_done() 1085 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN_RESTART && in netdev_tx_done() 1126 np->cur_tx != np->dirty_tx) { in intr_handler() [all …]
|
| H A D | interrupt.c | 595 for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0; in tulip_interrupt() 652 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { in tulip_interrupt() 655 dirty_tx, tp->cur_tx); in tulip_interrupt() 660 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) in tulip_interrupt()
|
| /linux/drivers/net/ethernet/smsc/ |
| H A D | epic100.c | 267 unsigned int cur_tx, dirty_tx; member 790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); in epic_restart() 881 ep->dirty_tx, ep->cur_tx); in epic_tx_timeout() 905 ep->dirty_tx = ep->cur_tx = 0; in epic_init_ring() 962 free_count = ep->cur_tx - ep->dirty_tx; in epic_start_xmit() 963 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit() 985 ep->cur_tx++; in epic_start_xmit() 1024 unsigned int dirty_tx, cur_tx; in epic_tx() local 1030 cur_tx = ep->cur_tx; in epic_tx() 1031 for (dirty_tx = ep->dirty_tx; cur_tx - dirty_tx > 0; dirty_tx++) { in epic_tx() [all …]
|
| /linux/drivers/net/ethernet/dlink/ |
| H A D | sundance.c | 391 unsigned int cur_tx, dirty_tx; member 990 np->cur_tx, np->cur_tx % TX_RING_SIZE, in tx_timeout() 1005 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { in tx_timeout() 1019 np->cur_rx = np->cur_tx = 0; in init_ring() 1067 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; in tx_poll() 1070 for (; np->cur_tx - np->cur_task > 0; np->cur_task++) { in tx_poll() 1096 entry = np->cur_tx % TX_RING_SIZE; in start_tx() 1111 np->cur_tx++; in start_tx() 1117 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 1 && in start_tx() 1126 dev->name, np->cur_tx, entry); in start_tx() [all …]
|
| H A D | dl2k.c | 479 np->cur_tx = 0; in rio_reset_ring() 739 entry = np->cur_tx % TX_RING_SIZE; in start_xmit() 775 np->cur_tx = (np->cur_tx + 1) % TX_RING_SIZE; in start_xmit() 776 if ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in start_xmit() 828 if (np->cur_tx != np->old_tx) in rio_interrupt() 846 while (entry != np->cur_tx) { in rio_free_tx() 873 ((np->cur_tx - np->old_tx + TX_RING_SIZE) % TX_RING_SIZE in rio_free_tx()
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | stmmac_main.c | 344 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail() 345 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail() 347 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail() 379 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_eee_tx_busy() 2608 unsigned int entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc() 2702 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_xdp_xmit_zc() 2703 entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc() 2763 while ((entry != tx_q->cur_tx) && count < priv->dma_conf.dma_tx_size) { in stmmac_tx_clean() 2915 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_tx_clean() 4237 p = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_vlan_insert() [all …]
|
| /linux/drivers/net/ethernet/adaptec/ |
| H A D | starfire.c | 544 unsigned int cur_tx, dirty_tx, reap_tx; member 1124 np->cur_rx = np->cur_tx = np->reap_tx = 0; in init_ring() 1182 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { in start_tx() 1194 prev_tx = np->cur_tx; in start_tx() 1195 entry = np->cur_tx % TX_RING_SIZE; in start_tx() 1239 dev->name, np->cur_tx, np->dirty_tx, in start_tx() 1243 np->cur_tx += np->tx_info[entry].used_slots; in start_tx() 1247 np->cur_tx += np->tx_info[entry].used_slots; in start_tx() 1251 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) in start_tx() 1264 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) in start_tx() [all …]
|
| /linux/drivers/net/ethernet/samsung/sxgbe/ |
| H A D | sxgbe_main.c | 194 return queue->dirty_tx + tx_qsize - queue->cur_tx - 1; in sxgbe_tx_avail() 427 tx_ring->cur_tx = 0; in init_tx_ring() 747 while (tqueue->dirty_tx != tqueue->cur_tx) { in sxgbe_tx_queue_clean() 760 __func__, tqueue->cur_tx, tqueue->dirty_tx); in sxgbe_tx_queue_clean() 835 tx_ring->cur_tx = 0; in sxgbe_restart_tx_queue() 1315 entry = tqueue->cur_tx % tx_rsize; in sxgbe_xmit() 1340 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit() 1362 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit() 1402 tqueue->cur_tx++; in sxgbe_xmit() 1406 __func__, tqueue->cur_tx % tx_rsize, in sxgbe_xmit()
|
| /linux/drivers/net/ethernet/3com/ |
| H A D | 3c59x.c | 604 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member 1679 vp->cur_tx = vp->dirty_tx = 0; in vortex_up() 1914 if (vp->cur_tx - vp->dirty_tx > 0 && ioread32(ioaddr + DownListPtr) == 0) in vortex_tx_timeout() 1917 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { in vortex_tx_timeout() 2116 int entry = vp->cur_tx % TX_RING_SIZE; in boomerang_start_xmit() 2118 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit() 2125 dev->name, vp->cur_tx); in boomerang_start_xmit() 2137 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { in boomerang_start_xmit() 2222 vp->cur_tx++; in boomerang_start_xmit() 2225 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) { in boomerang_start_xmit() [all …]
|
| /linux/drivers/net/ethernet/via/ |
| H A D | via-rhine.c | 469 unsigned int cur_tx, dirty_tx; member 1307 rp->dirty_tx = rp->cur_tx = 0; in alloc_tbufs() 1772 return (rp->cur_tx - rp->dirty_tx) >= TX_QUEUE_LEN; in rhine_tx_queue_full() 1787 entry = rp->cur_tx % TX_RING_SIZE; in rhine_start_tx() 1849 rp->cur_tx++; in rhine_start_tx() 1878 rp->cur_tx - 1, entry); in rhine_start_tx() 1924 unsigned int cur_tx; in rhine_tx() local 1934 cur_tx = rp->cur_tx; in rhine_tx() 1936 while (dirty_tx != cur_tx) { in rhine_tx()
|
| /linux/drivers/net/ethernet/realtek/ |
| H A D | 8139too.c | 593 unsigned long cur_tx; member 1454 tp->cur_tx = 0; in rtl8139_init_ring() 1649 tp->cur_tx = 0; in rtl8139_tx_clear() 1673 tp->cur_tx, tp->dirty_tx); in rtl8139_tx_timeout_task() 1727 entry = tp->cur_tx % NUM_TX_DESC; in rtl8139_start_xmit() 1751 tp->cur_tx++; in rtl8139_start_xmit() 1753 if ((tp->cur_tx - NUM_TX_DESC) == tp->dirty_tx) in rtl8139_start_xmit() 1774 tx_left = tp->cur_tx - dirty_tx; in rtl8139_tx_interrupt() 1819 if (tp->cur_tx - dirty_tx > NUM_TX_DESC) { in rtl8139_tx_interrupt() 1821 dirty_tx, tp->cur_tx); in rtl8139_tx_interrupt()
|
| /linux/drivers/net/ethernet/freescale/fs_enet/ |
| H A D | fs_enet-main.c | 349 fep->cur_tx = fep->tx_bd_base; in fs_init_bds() 490 bdp = fep->cur_tx; in fs_enet_start_xmit() 569 fep->cur_tx = bdp; in fs_enet_start_xmit() 609 !(CBDR_SC(fep->cur_tx) & BD_ENET_TX_READY); in fs_timeout_work()
|
| /linux/drivers/net/ethernet/sis/ |
| H A D | sis190.c | 272 u32 cur_tx; member 698 pending = tp->cur_tx - dirty_tx; in sis190_tx_interrupt() 811 tp->dirty_tx = tp->dirty_rx = tp->cur_tx = tp->cur_rx = 0; in sis190_init_ring_indexes() 1127 tp->cur_tx = tp->dirty_tx = 0; in sis190_tx_clear() 1199 entry = tp->cur_tx % NUM_TX_DESC; in sis190_start_xmit() 1236 tp->cur_tx++; in sis190_start_xmit() 1243 if ((tp->cur_tx - NUM_TX_DESC) == dirty_tx) { in sis190_start_xmit()
|
| H A D | sis900.c | 180 unsigned int cur_tx, dirty_tx; member 1138 sis_priv->dirty_tx = sis_priv->cur_tx = 0; in sis900_init_tx_ring() 1569 sis_priv->dirty_tx = sis_priv->cur_tx = 0; in sis900_tx_timeout() 1621 entry = sis_priv->cur_tx % NUM_TX_DESC; in sis900_start_xmit() 1639 sis_priv->cur_tx ++; in sis900_start_xmit() 1640 index_cur_tx = sis_priv->cur_tx; in sis900_start_xmit() 1906 for (; sis_priv->dirty_tx != sis_priv->cur_tx; sis_priv->dirty_tx++) { in sis900_finish_xmit() 1954 sis_priv->cur_tx - sis_priv->dirty_tx < NUM_TX_DESC - 4) { in sis900_finish_xmit()
|
| /linux/drivers/net/wireless/admtek/ |
| H A D | adm8211.c | 309 for (dirty_tx = priv->dirty_tx; priv->cur_tx - dirty_tx; dirty_tx++) { in adm8211_interrupt_tci() 342 if (priv->cur_tx - dirty_tx < priv->tx_ring_size - 2) in adm8211_interrupt_tci() 1475 priv->cur_rx = priv->cur_tx = priv->dirty_tx = 0; in adm8211_init_rings() 1638 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size / 2) in adm8211_tx_raw() 1643 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size - 2) in adm8211_tx_raw() 1646 entry = priv->cur_tx % priv->tx_ring_size; in adm8211_tx_raw() 1661 priv->cur_tx++; in adm8211_tx_raw()
|