/linux/drivers/net/ethernet/marvell/ |
H A D | mv643xx_eth.c | 181 (addr < txq->tso_hdrs_dma + txq->tx_ring_size * TSO_HEADER_SIZE)) 343 int tx_ring_size; member 404 int tx_ring_size; member 736 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_data_tso() 814 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_put_hdr_tso() 829 if ((txq->tx_desc_count + tso_count_descs(skb)) >= txq->tx_ring_size) { in txq_submit_tso() 904 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_submit_frag_skb() 943 if (txq->tx_ring_size - txq->tx_desc_count < MAX_SKB_FRAGS + 1) { in txq_submit_skb() 955 if (txq->tx_curr_desc == txq->tx_ring_size) in txq_submit_skb() 1085 if (txq->tx_used_desc == txq->tx_ring_size) in txq_reclaim() [all …]
|
H A D | pxa168_eth.c | 225 int tx_ring_size; member 722 pep->tx_used_desc_q = (tx_index + 1) % pep->tx_ring_size; in txq_reclaim() 1084 int tx_desc_num = pep->tx_ring_size; in txq_init() 1092 size = pep->tx_ring_size * sizeof(struct tx_desc); in txq_init() 1217 pep->tx_curr_desc_q = (tx_desc_curr + 1) % pep->tx_ring_size; in eth_alloc_tx_desc_index() 1238 && pep->tx_ring_size - pep->tx_desc_count > 1) { in pxa168_rx_poll() 1278 if (pep->tx_ring_size - pep->tx_desc_count <= 1) { in pxa168_eth_start_xmit() 1449 pep->tx_ring_size = NUM_TX_DESCS; in pxa168_eth_probe() 1457 pep->tx_ring_size = pep->pd->tx_queue_size; in pxa168_eth_probe()
|
H A D | sky2.c | 1105 *slot = RING_NEXT(*slot, sky2->tx_ring_size); in get_tx_le() 1599 sky2->tx_ring_size * sizeof(struct sky2_tx_le), in sky2_alloc_buffers() 1604 sky2->tx_ring = kcalloc(sky2->tx_ring_size, sizeof(struct tx_ring_info), in sky2_alloc_buffers() 1637 sky2->tx_ring_size * sizeof(struct sky2_tx_le), in sky2_free_buffers() 1704 sky2->tx_ring_size - 1); in sky2_hw_up() 1780 return (sky2->tx_prod - sky2->tx_cons) & (sky2->tx_ring_size - 1); in tx_inuse() 1978 for (i = sky2->tx_prod; i != slot; i = RING_NEXT(i, sky2->tx_ring_size)) { in sky2_xmit_frame() 2007 BUG_ON(done >= sky2->tx_ring_size); in sky2_tx_complete() 2010 idx = RING_NEXT(idx, sky2->tx_ring_size)) { in sky2_tx_complete() 2026 sky2->tx_next = RING_NEXT(idx, sky2->tx_ring_size); in sky2_tx_complete() [all …]
|
H A D | mvneta.c | 532 u16 tx_ring_size; member 3530 txq->size = pp->tx_ring_size; in mvneta_txq_sw_init() 4783 ring->tx_pending = pp->tx_ring_size; in mvneta_ethtool_get_ringparam() 4799 pp->tx_ring_size = clamp_t(u16, ring->tx_pending, in mvneta_ethtool_set_ringparam() 4801 if (pp->tx_ring_size != ring->tx_pending) in mvneta_ethtool_set_ringparam() 4803 pp->tx_ring_size, ring->tx_pending); in mvneta_ethtool_set_ringparam() 5382 txq->size = pp->tx_ring_size; in mvneta_init() 5661 pp->tx_ring_size = MVNETA_MAX_TXD; in mvneta_probe()
|
/linux/drivers/net/wireless/admtek/ |
H A D | adm8211.c | 33 static unsigned int tx_ring_size __read_mostly = 16; 36 module_param(tx_ring_size, uint, 0); 310 unsigned int entry = dirty_tx % priv->tx_ring_size; in adm8211_interrupt_tci() 342 if (priv->cur_tx - dirty_tx < priv->tx_ring_size - 2) in adm8211_interrupt_tci() 1465 for (i = 0; i < priv->tx_ring_size; i++) { in adm8211_init_rings() 1498 for (i = 0; i < priv->tx_ring_size; i++) { in adm8211_free_rings() 1638 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size / 2) in adm8211_tx_raw() 1643 if (priv->cur_tx - priv->dirty_tx == priv->tx_ring_size - 2) in adm8211_tx_raw() 1646 entry = priv->cur_tx % priv->tx_ring_size; in adm8211_tx_raw() 1653 if (entry == priv->tx_ring_size - 1) in adm8211_tx_raw() [all …]
|
H A D | adm8211.h | 547 unsigned int rx_ring_size, tx_ring_size; member
|
/linux/drivers/net/ethernet/freescale/ |
H A D | gianfar.c | 1072 for (i = 0; i < tx_queue->tx_ring_size; i++) { in free_skb_tx_queue() 1297 tx_queue->num_txbdfree = tx_queue->tx_ring_size; in gfar_init_bds() 1305 for (j = 0; j < tx_queue->tx_ring_size; j++) { in gfar_init_bds() 1347 priv->total_tx_ring_size += priv->tx_queue[i]->tx_ring_size; in gfar_alloc_skb_resources() 1369 addr += sizeof(struct txbd8) * tx_queue->tx_ring_size; in gfar_alloc_skb_resources() 1370 vaddr += sizeof(struct txbd8) * tx_queue->tx_ring_size; in gfar_alloc_skb_resources() 1388 kmalloc_array(tx_queue->tx_ring_size, in gfar_alloc_skb_resources() 1394 for (j = 0; j < tx_queue->tx_ring_size; j++) in gfar_alloc_skb_resources() 1884 tx_queue->tx_ring_size); in gfar_start_xmit() 1898 txbdp = next_txbd(txbdp, base, tx_queue->tx_ring_size); in gfar_start_xmit() [all …]
|
H A D | gianfar_ethtool.c | 401 rvals->tx_pending = tx_queue->tx_ring_size; in gfar_gringparam() 442 priv->tx_queue[i]->tx_ring_size = rvals->tx_pending; in gfar_sringparam()
|
/linux/drivers/infiniband/hw/hfi1/ |
H A D | ipoib_tx.c | 686 u32 tx_ring_size, tx_item_size; in hfi1_ipoib_txreq_init() local 694 tx_ring_size = roundup_pow_of_two(dev->tx_queue_len + 1); in hfi1_ipoib_txreq_init() 732 kvzalloc_node(array_size(tx_ring_size, tx_item_size), in hfi1_ipoib_txreq_init() 737 txq->tx_ring.max_items = tx_ring_size; in hfi1_ipoib_txreq_init() 741 for (j = 0; j < tx_ring_size; j++) { in hfi1_ipoib_txreq_init() 760 for (j = 0; j < tx_ring_size; j++) in hfi1_ipoib_txreq_init()
|
/linux/drivers/net/ethernet/oki-semi/pch_gbe/ |
H A D | pch_gbe_ethtool.c | 309 int tx_ring_size, rx_ring_size; in pch_gbe_set_ringparam() local 314 tx_ring_size = (int)sizeof(struct pch_gbe_tx_ring); in pch_gbe_set_ringparam() 322 txdr = kzalloc(tx_ring_size, GFP_KERNEL); in pch_gbe_set_ringparam()
|
/linux/drivers/net/ethernet/broadcom/bnx2x/ |
H A D | bnx2x_cmn.h | 748 WARN_ON(used > txdata->tx_ring_size); in bnx2x_tx_avail() 749 WARN_ON((txdata->tx_ring_size - used) > MAX_TX_AVAIL); in bnx2x_tx_avail() 752 return (s16)(txdata->tx_ring_size) - used; in bnx2x_tx_avail() 1170 txdata->tx_ring_size = IS_FCOE_FP(fp) ? MAX_TX_AVAIL : bp->tx_ring_size; in bnx2x_init_txdata()
|
/linux/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 832 int tx_ring_size; member 1027 np->tx_ring_size), in free_rings() 1034 np->tx_ring_size), in free_rings() 1940 np->last_tx.orig = &np->tx_ring.orig[np->tx_ring_size-1]; in nv_init_tx() 1942 np->last_tx.ex = &np->tx_ring.ex[np->tx_ring_size-1]; in nv_init_tx() 1945 np->last_tx_ctx = &np->tx_skb[np->tx_ring_size-1]; in nv_init_tx() 1952 for (i = 0; i < np->tx_ring_size; i++) { in nv_init_tx() 2015 for (i = 0; i < np->tx_ring_size; i++) { in nv_drain_tx() 2076 …return (u32)(np->tx_ring_size - ((np->tx_ring_size + (np->put_tx_ctx - np->get_tx_ctx)) % np->tx_r… in nv_get_empty_tx_slots() 2754 for (i = 0; i < np->tx_ring_size; i += 4) { in nv_tx_timeout() [all …]
|
/linux/drivers/net/vmxnet3/ |
H A D | vmxnet3_int.h | 420 u32 tx_ring_size; member 555 u32 tx_ring_size, u32 rx_ring_size, u32 rx_ring2_size,
|
H A D | vmxnet3_ethtool.c | 680 param->tx_pending = adapter->tx_ring_size; in vmxnet3_get_ringparam() 766 if (new_tx_ring_size == adapter->tx_ring_size && in vmxnet3_set_ringparam() 822 adapter->tx_ring_size = new_tx_ring_size; in vmxnet3_set_ringparam()
|
/linux/drivers/net/wan/ |
H A D | fsl_ucc_hdlc.h | 99 unsigned short tx_ring_size; member
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
H A D | en_main.c | 171 params->prof[i].tx_ring_size = MLX4_EN_MIN_TX_SIZE; in mlx4_en_get_profile() 174 params->prof[i].tx_ring_size = MLX4_EN_DEF_TX_RING_SIZE; in mlx4_en_get_profile()
|
H A D | mlx4_en.h | 390 u32 tx_ring_size; member
|
/linux/drivers/net/ethernet/ |
H A D | jme.c | 562 TX_RING_ALLOC_SIZE(jme->tx_ring_size), in jme_setup_tx_resources() 577 atomic_set(&txring->nr_free, jme->tx_ring_size); in jme_setup_tx_resources() 579 txring->bufinf = kcalloc(jme->tx_ring_size, in jme_setup_tx_resources() 589 TX_RING_ALLOC_SIZE(jme->tx_ring_size), in jme_setup_tx_resources() 611 for (i = 0 ; i < jme->tx_ring_size ; ++i) { in jme_free_tx_resources() 626 TX_RING_ALLOC_SIZE(jme->tx_ring_size), in jme_free_tx_resources() 660 jwrite32(jme, JME_TXQDC, jme->tx_ring_size); in jme_enable_tx_engine() 1430 max = jme->tx_ring_size - atomic_read(&txring->nr_free); in jme_tx_clean_tasklet() 2988 jme->tx_ring_size = 1 << 10; in jme_init_one() 2989 jme->tx_ring_mask = jme->tx_ring_size - 1; in jme_init_one()
|
H A D | jme.h | 423 u32 tx_ring_size; member
|
/linux/drivers/net/ethernet/broadcom/ |
H A D | bcm63xx_enet.h | 257 int tx_ring_size; member
|
/linux/Documentation/networking/ |
H A D | driver.rst | 54 return dr->tx_ring_size - (used & bp->tx_ring_mask);
|
/linux/drivers/net/ethernet/cadence/ |
H A D | macb_main.c | 64 * (bp)->tx_ring_size) 67 #define MACB_TX_WAKEUP_THRESH(bp) (3 * (bp)->tx_ring_size / 4) 184 return index & (bp->tx_ring_size - 1); in macb_tx_ring_wrap() 1291 bp->tx_ring_size) <= MACB_TX_WAKEUP_THRESH(bp)) in macb_tx_complete() 2157 if (unlikely(entry == (bp->tx_ring_size - 1))) in macb_tx_map() 2374 bp->tx_ring_size) < desc_cnt) { in macb_start_xmit() 2396 if (CIRC_SPACE(queue->tx_head, queue->tx_tail, bp->tx_ring_size) < 1) in macb_start_xmit() 2556 size = bp->tx_ring_size * sizeof(struct macb_tx_skb); in macb_alloc_consistent() 2611 for (i = 0; i < bp->tx_ring_size; i++) { in gem_init_rings() 2636 for (i = 0; i < bp->tx_ring_size; i++) { in macb_init_rings() [all …]
|
/linux/drivers/net/ethernet/adaptec/ |
H A D | starfire.c | 861 size_t tx_done_q_size, rx_done_q_size, tx_ring_size, rx_ring_size; in netdev_open() local 880 …tx_ring_size = ((sizeof(starfire_tx_desc) * TX_RING_SIZE + QUEUE_ALIGN - 1) / QUEUE_ALIGN) * QUEUE… in netdev_open() 882 np->queue_mem_size = tx_done_q_size + rx_done_q_size + tx_ring_size + rx_ring_size; in netdev_open() 897 np->rx_ring = (void *) np->tx_ring + tx_ring_size; in netdev_open() 898 np->rx_ring_dma = np->tx_ring_dma + tx_ring_size; in netdev_open()
|
/linux/drivers/net/ethernet/altera/ |
H A D | altera_tse.h | 430 u32 tx_ring_size; member
|
/linux/drivers/net/ethernet/broadcom/bnxt/ |
H A D | bnxt_xdp.c | 255 if (tx_avail != bp->tx_ring_size) in bnxt_rx_xdp()
|