Home
last modified time | relevance | path

Searched refs:cur_tx (Results 1 – 25 of 56) sorted by relevance

123

/linux/drivers/net/ethernet/amd/
H A Dariadne.c92 int cur_tx, cur_rx; /* The next free ring entry */ member
125 priv->cur_rx = priv->cur_tx = 0; in ariadne_init_ring()
308 while (dirty_tx < priv->cur_tx) { in ariadne_interrupt()
345 if (priv->cur_tx - dirty_tx >= TX_RING_SIZE) { in ariadne_interrupt()
347 dirty_tx, priv->cur_tx, in ariadne_interrupt()
354 dirty_tx > priv->cur_tx - TX_RING_SIZE + 2) { in ariadne_interrupt()
576 entry = priv->cur_tx % TX_RING_SIZE; in ariadne_start_xmit()
596 priv->cur_tx++; in ariadne_start_xmit()
597 if ((priv->cur_tx >= TX_RING_SIZE) && in ariadne_start_xmit()
601 priv->cur_tx, priv->dirty_tx); in ariadne_start_xmit()
[all …]
H A Datarilance.c223 int cur_rx, cur_tx; /* The next free ring entry */ member
689 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring()
748 lp->dirty_tx, lp->cur_tx, in lance_tx_timeout()
815 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit()
829 lp->cur_tx++; in lance_start_xmit()
830 while( lp->cur_tx >= TX_RING_SIZE && lp->dirty_tx >= TX_RING_SIZE ) { in lance_start_xmit()
831 lp->cur_tx -= TX_RING_SIZE; in lance_start_xmit()
885 while( dirty_tx < lp->cur_tx) { in lance_interrupt()
921 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
924 dirty_tx, lp->cur_tx, lp->tx_full )); in lance_interrupt()
[all …]
H A Dlance.c249 int cur_rx, cur_tx; /* The next free ring entry */ member
872 lp->cur_rx = lp->cur_tx = 0; in lance_init_ring()
936 lp->dirty_tx, lp->cur_tx, netif_queue_stopped(dev) ? " (full)" : "", in lance_tx_timeout()
976 entry = lp->cur_tx & TX_RING_MOD_MASK; in lance_start_xmit()
1011 lp->cur_tx++; in lance_start_xmit()
1017 if ((lp->cur_tx - lp->dirty_tx) >= TX_RING_SIZE) in lance_start_xmit()
1056 while (dirty_tx < lp->cur_tx) { in lance_interrupt()
1100 if (lp->cur_tx - dirty_tx >= TX_RING_SIZE) { in lance_interrupt()
1102 dirty_tx, lp->cur_tx, in lance_interrupt()
1110 dirty_tx > lp->cur_tx - TX_RING_SIZE + 2) in lance_interrupt()
/linux/drivers/net/ethernet/packetengines/
H A Dyellowfin.c326 unsigned int cur_tx, dirty_tx; member
693 yp->cur_tx, yp->dirty_tx, in yellowfin_tx_timeout()
718 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_tx_timeout()
732 yp->cur_rx = yp->cur_tx = 0; in yellowfin_init_ring()
826 entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit()
857 yp->cur_tx++; in yellowfin_start_xmit()
866 yp->cur_tx++; in yellowfin_start_xmit()
868 unsigned next_entry = yp->cur_tx % TX_RING_SIZE; in yellowfin_start_xmit()
883 if (yp->cur_tx - yp->dirty_tx < TX_QUEUE_SIZE) in yellowfin_start_xmit()
890 yp->cur_tx, entry); in yellowfin_start_xmit()
[all …]
H A Dhamachi.c495 unsigned int cur_tx, dirty_tx; member
1001 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++) { in hamachi_tx()
1116 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_tx_timeout()
1172 hmp->cur_rx = hmp->cur_tx = 0; in hamachi_init_ring()
1229 printk(KERN_WARNING "%s: Hamachi transmit queue full at slot %d.\n",dev->name, hmp->cur_tx); in hamachi_start_xmit()
1243 entry = hmp->cur_tx % TX_RING_SIZE; in hamachi_start_xmit()
1267 hmp->cur_tx++; in hamachi_start_xmit()
1286 if ((hmp->cur_tx - hmp->dirty_tx) < (TX_RING_SIZE - 4)) in hamachi_start_xmit()
1295 dev->name, hmp->cur_tx, entry); in hamachi_start_xmit()
1340 for (; hmp->cur_tx - hmp->dirty_tx > 0; hmp->dirty_tx++){ in hamachi_interrupt()
[all …]
/linux/drivers/net/arcnet/
H A Darcnet.c567 lp->cur_tx = lp->next_tx = -1; in arcnet_open()
709 lp->hw.status(dev), lp->cur_tx, lp->next_tx, skb->len, skb->protocol); in arcnet_send_packet()
790 lp->hw.status(dev), lp->intmask, lp->next_tx, lp->cur_tx); in go_tx()
792 if (lp->cur_tx != -1 || lp->next_tx == -1) in go_tx()
798 lp->cur_tx = lp->next_tx; in go_tx()
802 lp->hw.command(dev, TXcmd | (lp->cur_tx << 3)); in go_tx()
828 lp->hw.command(dev, NOTXcmd | (lp->cur_tx << 3)); in arcnet_timeout()
845 if (lp->cur_tx == -1) in arcnet_timeout()
965 if (lp->cur_tx != -1 && !lp->timed_out) { in arcnet_interrupt()
990 if (lp->cur_tx != -1) in arcnet_interrupt()
[all …]
/linux/drivers/net/can/spi/mcp251xfd/
H A Dmcp251xfd-ethtool.c44 layout.cur_tx != priv->tx->obj_num) && in mcp251xfd_ring_set_ringparam()
50 priv->tx->obj_num = layout.cur_tx; in mcp251xfd_ring_set_ringparam()
112 priv->tx->obj_num = layout.cur_tx; in mcp251xfd_ring_set_coalesce()
H A Dmcp251xfd-ram.c153 layout->cur_tx = num_tx; in can_ram_get_layout()
158 layout->cur_tx = layout->default_tx; in can_ram_get_layout()
H A Dmcp251xfd-ram.h50 u8 cur_tx; member
/linux/drivers/net/ethernet/
H A Dfealnx.c396 struct fealnx_desc *cur_tx; member
1146 iowrite32(np->tx_ring_dma + ((char*)np->cur_tx - (char*)np->tx_ring), in enable_rxtx()
1271 np->cur_tx = &np->tx_ring[0]; in init_ring()
1379 np->cur_tx = &np->tx_ring[0]; in reset_tx_descriptors()
1488 long tx_status = np->cur_tx->status; in intr_handler()
1489 long tx_control = np->cur_tx->control; in intr_handler()
1494 next = np->cur_tx->next_desc_logical; in intr_handler()
1532 np->cur_tx->buffer, in intr_handler()
1533 np->cur_tx->skbuff->len, in intr_handler()
1535 dev_consume_skb_irq(np->cur_tx->skbuff); in intr_handler()
[all …]
H A Dethoc.c205 unsigned int cur_tx; member
308 dev->cur_tx = 0; in ethoc_init_ring()
525 if (bd.stat & TX_BD_READY || (priv->dty_tx == priv->cur_tx)) { in ethoc_tx()
536 (priv->dty_tx == priv->cur_tx)) in ethoc_tx()
544 if ((priv->cur_tx - priv->dty_tx) <= (priv->num_tx / 2)) in ethoc_tx()
897 entry = priv->cur_tx % priv->num_tx; in ethoc_start_xmit()
899 priv->cur_tx++; in ethoc_start_xmit()
917 if (priv->cur_tx == (priv->dty_tx + priv->num_tx)) { in ethoc_start_xmit()
/linux/drivers/net/ethernet/dec/tulip/
H A Dtulip_core.c323 tp->cur_rx = tp->cur_tx = 0; in tulip_up()
356 tp->tx_buffers[tp->cur_tx].skb = NULL; in tulip_up()
357 tp->tx_buffers[tp->cur_tx].mapping = mapping; in tulip_up()
360 tp->tx_ring[tp->cur_tx].length = cpu_to_le32(0x08000000 | 192); in tulip_up()
361 tp->tx_ring[tp->cur_tx].buffer1 = cpu_to_le32(mapping); in tulip_up()
362 tp->tx_ring[tp->cur_tx].status = cpu_to_le32(DescOwned); in tulip_up()
364 tp->cur_tx++; in tulip_up()
664 entry = tp->cur_tx % TX_RING_SIZE; in tulip_start_xmit()
672 if (tp->cur_tx - tp->dirty_tx < TX_RING_SIZE/2) {/* Typical path */ in tulip_start_xmit()
674 } else if (tp->cur_tx - tp->dirty_tx == TX_RING_SIZE/2) { in tulip_start_xmit()
[all …]
H A Dwinbond-840.c304 unsigned int cur_tx, dirty_tx; member
821 np->tx_q_bytes = np->dirty_tx = np->cur_tx = 0; in init_rxtx_rings()
932 np->cur_tx, np->dirty_tx, np->tx_full, np->tx_q_bytes); in tx_timeout()
992 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1021 np->cur_tx++; in start_tx()
1030 if (np->cur_tx - np->dirty_tx > TX_QUEUE_LEN || in start_tx()
1040 np->cur_tx, entry); in start_tx()
1048 for (; np->cur_tx - np->dirty_tx > 0; np->dirty_tx++) { in netdev_tx_done()
1085 np->cur_tx - np->dirty_tx < TX_QUEUE_LEN_RESTART && in netdev_tx_done()
1126 np->cur_tx != np->dirty_tx) { in intr_handler()
[all …]
H A Dinterrupt.c595 for (dirty_tx = tp->dirty_tx; tp->cur_tx - dirty_tx > 0; in tulip_interrupt()
652 if (tp->cur_tx - dirty_tx > TX_RING_SIZE) { in tulip_interrupt()
655 dirty_tx, tp->cur_tx); in tulip_interrupt()
660 if (tp->cur_tx - dirty_tx < TX_RING_SIZE - 2) in tulip_interrupt()
/linux/drivers/net/ethernet/smsc/
H A Depic100.c267 unsigned int cur_tx, dirty_tx; member
790 ep->cur_rx, ep->dirty_rx, ep->dirty_tx, ep->cur_tx); in epic_restart()
881 ep->dirty_tx, ep->cur_tx); in epic_tx_timeout()
905 ep->dirty_tx = ep->cur_tx = 0; in epic_init_ring()
962 free_count = ep->cur_tx - ep->dirty_tx; in epic_start_xmit()
963 entry = ep->cur_tx % TX_RING_SIZE; in epic_start_xmit()
985 ep->cur_tx++; in epic_start_xmit()
1024 unsigned int dirty_tx, cur_tx; in epic_tx() local
1030 cur_tx = ep->cur_tx; in epic_tx()
1031 for (dirty_tx = ep->dirty_tx; cur_tx - dirty_tx > 0; dirty_tx++) { in epic_tx()
[all …]
/linux/drivers/net/ethernet/dlink/
H A Dsundance.c391 unsigned int cur_tx, dirty_tx; member
990 np->cur_tx, np->cur_tx % TX_RING_SIZE, in tx_timeout()
1005 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 4) { in tx_timeout()
1019 np->cur_rx = np->cur_tx = 0; in init_ring()
1066 &np->tx_ring[(np->cur_tx - 1) % TX_RING_SIZE]; in tx_poll()
1069 for (; np->cur_tx - np->cur_task > 0; np->cur_task++) { in tx_poll()
1094 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1108 np->cur_tx++; in start_tx()
1114 if (np->cur_tx - np->dirty_tx < TX_QUEUE_LEN - 1 && in start_tx()
1123 dev->name, np->cur_tx, entry); in start_tx()
[all …]
/linux/drivers/net/ethernet/3com/
H A D3c515.c303 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member
829 vp->cur_tx = vp->dirty_tx = 0; in corkscrew_open()
967 vp->cur_tx); in corkscrew_timeout()
1000 int entry = vp->cur_tx % TX_RING_SIZE; in corkscrew_start_xmit()
1007 if (vp->cur_tx != 0) in corkscrew_start_xmit()
1008 prev_entry = &vp->tx_ring[(vp->cur_tx - 1) % TX_RING_SIZE]; in corkscrew_start_xmit()
1013 dev->name, vp->cur_tx); in corkscrew_start_xmit()
1037 vp->cur_tx++; in corkscrew_start_xmit()
1038 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) in corkscrew_start_xmit()
1162 while (lp->cur_tx - dirty_tx > 0) { in corkscrew_interrupt()
[all …]
H A D3c59x.c604 unsigned int cur_rx, cur_tx; /* The next free ring entry */ member
1679 vp->cur_tx = vp->dirty_tx = 0; in vortex_up()
1914 if (vp->cur_tx - vp->dirty_tx > 0 && ioread32(ioaddr + DownListPtr) == 0) in vortex_tx_timeout()
1917 if (vp->cur_tx - vp->dirty_tx < TX_RING_SIZE) { in vortex_tx_timeout()
2116 int entry = vp->cur_tx % TX_RING_SIZE; in boomerang_start_xmit()
2118 struct boom_tx_desc *prev_entry = &vp->tx_ring[(vp->cur_tx-1) % TX_RING_SIZE]; in boomerang_start_xmit()
2125 dev->name, vp->cur_tx); in boomerang_start_xmit()
2137 if (vp->cur_tx - vp->dirty_tx >= TX_RING_SIZE) { in boomerang_start_xmit()
2222 vp->cur_tx++; in boomerang_start_xmit()
2225 if (vp->cur_tx - vp->dirty_tx > TX_RING_SIZE - 1) { in boomerang_start_xmit()
[all …]
/linux/drivers/net/ethernet/stmicro/stmmac/
H A Dstmmac_main.c367 if (tx_q->dirty_tx > tx_q->cur_tx) in stmmac_tx_avail()
368 avail = tx_q->dirty_tx - tx_q->cur_tx - 1; in stmmac_tx_avail()
370 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
418 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_enable_eee_mode()
2484 unsigned int entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc()
2576 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_xdp_xmit_zc()
2577 entry = tx_q->cur_tx; in stmmac_xdp_xmit_zc()
2637 while ((entry != tx_q->cur_tx) && count < priv->dma_conf.dma_tx_size) { in stmmac_tx_clean()
2792 if (tx_q->dirty_tx != tx_q->cur_tx) in stmmac_tx_clean()
4098 p = &tx_q->dma_entx[tx_q->cur_tx].basic; in stmmac_vlan_insert()
[all …]
H A Dchain_mode.c22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
81 tx_q->cur_tx = entry; in jumbo_frm()
H A Dring_mode.c22 unsigned int entry = tx_q->cur_tx; in jumbo_frm()
89 tx_q->cur_tx = entry; in jumbo_frm()
/linux/drivers/net/ethernet/adaptec/
H A Dstarfire.c544 unsigned int cur_tx, dirty_tx, reap_tx; member
1124 np->cur_rx = np->cur_tx = np->reap_tx = 0; in init_ring()
1182 if ((np->cur_tx - np->dirty_tx) + skb_num_frags(skb) * 2 > TX_RING_SIZE) { in start_tx()
1194 prev_tx = np->cur_tx; in start_tx()
1195 entry = np->cur_tx % TX_RING_SIZE; in start_tx()
1239 dev->name, np->cur_tx, np->dirty_tx, in start_tx()
1243 np->cur_tx += np->tx_info[entry].used_slots; in start_tx()
1247 np->cur_tx += np->tx_info[entry].used_slots; in start_tx()
1251 if (np->cur_tx % (TX_RING_SIZE / 2) == 0) in start_tx()
1264 if ((np->cur_tx - np->dirty_tx) + 4 > TX_RING_SIZE) in start_tx()
[all …]
/linux/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_main.c193 return queue->dirty_tx + tx_qsize - queue->cur_tx - 1; in sxgbe_tx_avail()
426 tx_ring->cur_tx = 0; in init_tx_ring()
746 while (tqueue->dirty_tx != tqueue->cur_tx) { in sxgbe_tx_queue_clean()
759 __func__, tqueue->cur_tx, tqueue->dirty_tx); in sxgbe_tx_queue_clean()
834 tx_ring->cur_tx = 0; in sxgbe_restart_tx_queue()
1314 entry = tqueue->cur_tx % tx_rsize; in sxgbe_xmit()
1339 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit()
1361 entry = (++tqueue->cur_tx) % tx_rsize; in sxgbe_xmit()
1401 tqueue->cur_tx++; in sxgbe_xmit()
1405 __func__, tqueue->cur_tx % tx_rsize, in sxgbe_xmit()
/linux/tools/testing/selftests/net/
H A Dveth.sh89 local cur_tx=`ip netns exec $BASE$target ethtool -l $dev |\
95 if [ "$cur_rx" = "$rx" -a "$cur_tx" = "$tx" -a "$cur_combined" = "n/a" ]; then
98 echo " fail rx:$rx:$cur_rx tx:$tx:$cur_tx combined:n/a:$cur_combined"
/linux/drivers/net/ethernet/via/
H A Dvia-rhine.c469 unsigned int cur_tx, dirty_tx; member
1307 rp->dirty_tx = rp->cur_tx = 0; in alloc_tbufs()
1766 return (rp->cur_tx - rp->dirty_tx) >= TX_QUEUE_LEN; in rhine_tx_queue_full()
1781 entry = rp->cur_tx % TX_RING_SIZE; in rhine_start_tx()
1843 rp->cur_tx++; in rhine_start_tx()
1872 rp->cur_tx - 1, entry); in rhine_start_tx()
1918 unsigned int cur_tx; in rhine_tx() local
1928 cur_tx = rp->cur_tx; in rhine_tx()
1930 while (dirty_tx != cur_tx) { in rhine_tx()

123