Lines Matching refs:dma_conf

78 #define STMMAC_TX_THRESH(x)	((x)->dma_conf.dma_tx_size / 4)
79 #define STMMAC_RX_THRESH(x) ((x)->dma_conf.dma_rx_size / 4)
238 rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_disable_all_queues()
364 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tx_avail()
370 avail = priv->dma_conf.dma_tx_size - tx_q->cur_tx + tx_q->dirty_tx - 1; in stmmac_tx_avail()
382 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rx_dirty()
388 dirty = priv->dma_conf.dma_rx_size - rx_q->dirty_rx + rx_q->cur_rx; in stmmac_rx_dirty()
416 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_enable_eee_mode()
1268 struct stmmac_dma_conf *dma_conf) in stmmac_display_rx_rings() argument
1277 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in stmmac_display_rx_rings()
1290 stmmac_display_ring(priv, head_rx, dma_conf->dma_rx_size, true, in stmmac_display_rx_rings()
1296 struct stmmac_dma_conf *dma_conf) in stmmac_display_tx_rings() argument
1305 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in stmmac_display_tx_rings()
1320 stmmac_display_ring(priv, head_tx, dma_conf->dma_tx_size, false, in stmmac_display_tx_rings()
1326 struct stmmac_dma_conf *dma_conf) in stmmac_display_rings() argument
1329 stmmac_display_rx_rings(priv, dma_conf); in stmmac_display_rings()
1332 stmmac_display_tx_rings(priv, dma_conf); in stmmac_display_rings()
1362 struct stmmac_dma_conf *dma_conf, in stmmac_clear_rx_descriptors() argument
1365 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in stmmac_clear_rx_descriptors()
1369 for (i = 0; i < dma_conf->dma_rx_size; i++) in stmmac_clear_rx_descriptors()
1373 (i == dma_conf->dma_rx_size - 1), in stmmac_clear_rx_descriptors()
1374 dma_conf->dma_buf_sz); in stmmac_clear_rx_descriptors()
1378 (i == dma_conf->dma_rx_size - 1), in stmmac_clear_rx_descriptors()
1379 dma_conf->dma_buf_sz); in stmmac_clear_rx_descriptors()
1391 struct stmmac_dma_conf *dma_conf, in stmmac_clear_tx_descriptors() argument
1394 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in stmmac_clear_tx_descriptors()
1398 for (i = 0; i < dma_conf->dma_tx_size; i++) { in stmmac_clear_tx_descriptors()
1399 int last = (i == (dma_conf->dma_tx_size - 1)); in stmmac_clear_tx_descriptors()
1421 struct stmmac_dma_conf *dma_conf) in stmmac_clear_descriptors() argument
1429 stmmac_clear_rx_descriptors(priv, dma_conf, queue); in stmmac_clear_descriptors()
1433 stmmac_clear_tx_descriptors(priv, dma_conf, queue); in stmmac_clear_descriptors()
1448 struct stmmac_dma_conf *dma_conf, in stmmac_init_rx_buffers() argument
1452 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in stmmac_init_rx_buffers()
1481 if (dma_conf->dma_buf_sz == BUF_SIZE_16KiB) in stmmac_init_rx_buffers()
1516 struct stmmac_dma_conf *dma_conf, in stmmac_free_tx_buffer() argument
1519 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in stmmac_free_tx_buffer()
1562 struct stmmac_dma_conf *dma_conf, in dma_free_rx_skbufs() argument
1565 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in dma_free_rx_skbufs()
1568 for (i = 0; i < dma_conf->dma_rx_size; i++) in dma_free_rx_skbufs()
1573 struct stmmac_dma_conf *dma_conf, in stmmac_alloc_rx_buffers() argument
1576 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in stmmac_alloc_rx_buffers()
1579 for (i = 0; i < dma_conf->dma_rx_size; i++) { in stmmac_alloc_rx_buffers()
1588 ret = stmmac_init_rx_buffers(priv, dma_conf, p, i, flags, in stmmac_alloc_rx_buffers()
1606 struct stmmac_dma_conf *dma_conf, in dma_free_rx_xskbufs() argument
1609 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in dma_free_rx_xskbufs()
1612 for (i = 0; i < dma_conf->dma_rx_size; i++) { in dma_free_rx_xskbufs()
1624 struct stmmac_dma_conf *dma_conf, in stmmac_alloc_rx_buffers_zc() argument
1627 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in stmmac_alloc_rx_buffers_zc()
1636 for (i = 0; i < dma_conf->dma_rx_size; i++) { in stmmac_alloc_rx_buffers_zc()
1679 struct stmmac_dma_conf *dma_conf, in __init_dma_rx_desc_rings() argument
1682 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in __init_dma_rx_desc_rings()
1689 stmmac_clear_rx_descriptors(priv, dma_conf, queue); in __init_dma_rx_desc_rings()
1716 stmmac_alloc_rx_buffers_zc(priv, dma_conf, queue); in __init_dma_rx_desc_rings()
1718 ret = stmmac_alloc_rx_buffers(priv, dma_conf, queue, flags); in __init_dma_rx_desc_rings()
1728 dma_conf->dma_rx_size, 1); in __init_dma_rx_desc_rings()
1732 dma_conf->dma_rx_size, 0); in __init_dma_rx_desc_rings()
1739 struct stmmac_dma_conf *dma_conf, in init_dma_rx_desc_rings() argument
1752 ret = __init_dma_rx_desc_rings(priv, dma_conf, queue, flags); in init_dma_rx_desc_rings()
1761 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in init_dma_rx_desc_rings()
1764 dma_free_rx_xskbufs(priv, dma_conf, queue); in init_dma_rx_desc_rings()
1766 dma_free_rx_skbufs(priv, dma_conf, queue); in init_dma_rx_desc_rings()
1787 struct stmmac_dma_conf *dma_conf, in __init_dma_tx_desc_rings() argument
1790 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in __init_dma_tx_desc_rings()
1802 dma_conf->dma_tx_size, 1); in __init_dma_tx_desc_rings()
1806 dma_conf->dma_tx_size, 0); in __init_dma_tx_desc_rings()
1811 for (i = 0; i < dma_conf->dma_tx_size; i++) { in __init_dma_tx_desc_rings()
1834 struct stmmac_dma_conf *dma_conf) in init_dma_tx_desc_rings() argument
1843 __init_dma_tx_desc_rings(priv, dma_conf, queue); in init_dma_tx_desc_rings()
1858 struct stmmac_dma_conf *dma_conf, in init_dma_desc_rings() argument
1864 ret = init_dma_rx_desc_rings(dev, dma_conf, flags); in init_dma_desc_rings()
1868 ret = init_dma_tx_desc_rings(dev, dma_conf); in init_dma_desc_rings()
1870 stmmac_clear_descriptors(priv, dma_conf); in init_dma_desc_rings()
1873 stmmac_display_rings(priv, dma_conf); in init_dma_desc_rings()
1885 struct stmmac_dma_conf *dma_conf, in dma_free_tx_skbufs() argument
1888 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in dma_free_tx_skbufs()
1893 for (i = 0; i < dma_conf->dma_tx_size; i++) in dma_free_tx_skbufs()
1894 stmmac_free_tx_buffer(priv, dma_conf, queue, i); in dma_free_tx_skbufs()
1913 dma_free_tx_skbufs(priv, &priv->dma_conf, queue); in stmmac_free_tx_skbufs()
1923 struct stmmac_dma_conf *dma_conf, in __free_dma_rx_desc_resources() argument
1926 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in __free_dma_rx_desc_resources()
1930 dma_free_rx_xskbufs(priv, dma_conf, queue); in __free_dma_rx_desc_resources()
1932 dma_free_rx_skbufs(priv, dma_conf, queue); in __free_dma_rx_desc_resources()
1939 dma_free_coherent(priv->device, dma_conf->dma_rx_size * in __free_dma_rx_desc_resources()
1943 dma_free_coherent(priv->device, dma_conf->dma_rx_size * in __free_dma_rx_desc_resources()
1956 struct stmmac_dma_conf *dma_conf) in free_dma_rx_desc_resources() argument
1963 __free_dma_rx_desc_resources(priv, dma_conf, queue); in free_dma_rx_desc_resources()
1973 struct stmmac_dma_conf *dma_conf, in __free_dma_tx_desc_resources() argument
1976 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in __free_dma_tx_desc_resources()
1981 dma_free_tx_skbufs(priv, dma_conf, queue); in __free_dma_tx_desc_resources()
1994 size *= dma_conf->dma_tx_size; in __free_dma_tx_desc_resources()
2003 struct stmmac_dma_conf *dma_conf) in free_dma_tx_desc_resources() argument
2010 __free_dma_tx_desc_resources(priv, dma_conf, queue); in free_dma_tx_desc_resources()
2024 struct stmmac_dma_conf *dma_conf, in __alloc_dma_rx_desc_resources() argument
2027 struct stmmac_rx_queue *rx_q = &dma_conf->rx_queue[queue]; in __alloc_dma_rx_desc_resources()
2039 pp_params.pool_size = dma_conf->dma_rx_size; in __alloc_dma_rx_desc_resources()
2040 num_pages = DIV_ROUND_UP(dma_conf->dma_buf_sz, PAGE_SIZE); in __alloc_dma_rx_desc_resources()
2055 rx_q->buf_pool = kcalloc(dma_conf->dma_rx_size, in __alloc_dma_rx_desc_resources()
2063 dma_conf->dma_rx_size * in __alloc_dma_rx_desc_resources()
2072 dma_conf->dma_rx_size * in __alloc_dma_rx_desc_resources()
2098 struct stmmac_dma_conf *dma_conf) in alloc_dma_rx_desc_resources() argument
2106 ret = __alloc_dma_rx_desc_resources(priv, dma_conf, queue); in alloc_dma_rx_desc_resources()
2114 free_dma_rx_desc_resources(priv, dma_conf); in alloc_dma_rx_desc_resources()
2130 struct stmmac_dma_conf *dma_conf, in __alloc_dma_tx_desc_resources() argument
2133 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[queue]; in __alloc_dma_tx_desc_resources()
2140 tx_q->tx_skbuff_dma = kcalloc(dma_conf->dma_tx_size, in __alloc_dma_tx_desc_resources()
2146 tx_q->tx_skbuff = kcalloc(dma_conf->dma_tx_size, in __alloc_dma_tx_desc_resources()
2159 size *= dma_conf->dma_tx_size; in __alloc_dma_tx_desc_resources()
2177 struct stmmac_dma_conf *dma_conf) in alloc_dma_tx_desc_resources() argument
2185 ret = __alloc_dma_tx_desc_resources(priv, dma_conf, queue); in alloc_dma_tx_desc_resources()
2193 free_dma_tx_desc_resources(priv, dma_conf); in alloc_dma_tx_desc_resources()
2207 struct stmmac_dma_conf *dma_conf) in alloc_dma_desc_resources() argument
2210 int ret = alloc_dma_rx_desc_resources(priv, dma_conf); in alloc_dma_desc_resources()
2215 ret = alloc_dma_tx_desc_resources(priv, dma_conf); in alloc_dma_desc_resources()
2226 struct stmmac_dma_conf *dma_conf) in free_dma_desc_resources() argument
2229 free_dma_tx_desc_resources(priv, dma_conf); in free_dma_desc_resources()
2234 free_dma_rx_desc_resources(priv, dma_conf); in free_dma_desc_resources()
2410 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[chan]; in stmmac_dma_operation_mode()
2425 priv->dma_conf.dma_buf_sz, in stmmac_dma_operation_mode()
2481 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_xdp_xmit_zc()
2576 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_xdp_xmit_zc()
2624 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tx_clean()
2637 while ((entry != tx_q->cur_tx) && count < priv->dma_conf.dma_tx_size) { in stmmac_tx_clean()
2747 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); in stmmac_tx_clean()
2818 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_tx_err()
2823 dma_free_tx_skbufs(priv, &priv->dma_conf, chan); in stmmac_tx_err()
2824 stmmac_clear_tx_descriptors(priv, &priv->dma_conf, chan); in stmmac_tx_err()
2885 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[chan]; in stmmac_napi_check()
2886 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_napi_check()
3051 rx_q = &priv->dma_conf.rx_queue[chan]; in stmmac_init_dma_engine()
3065 tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_init_dma_engine()
3080 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tx_timer_arm()
3146 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_init_coalesce()
3168 (priv->dma_conf.dma_tx_size - 1), chan); in stmmac_set_rings_length()
3173 (priv->dma_conf.dma_rx_size - 1), chan); in stmmac_set_rings_length()
3491 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_hw_setup()
3513 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_hw_setup()
3552 free_irq(priv->tx_irq[j], &priv->dma_conf.tx_queue[j]); in stmmac_free_irq()
3561 free_irq(priv->rx_irq[j], &priv->dma_conf.rx_queue[j]); in stmmac_free_irq()
3718 0, int_name, &priv->dma_conf.rx_queue[i]); in stmmac_request_irq_multi_msi()
3743 0, int_name, &priv->dma_conf.tx_queue[i]); in stmmac_request_irq_multi_msi()
3857 struct stmmac_dma_conf *dma_conf; in stmmac_setup_dma_desc() local
3860 dma_conf = kzalloc(sizeof(*dma_conf), GFP_KERNEL); in stmmac_setup_dma_desc()
3861 if (!dma_conf) { in stmmac_setup_dma_desc()
3874 dma_conf->dma_buf_sz = bfsize; in stmmac_setup_dma_desc()
3878 dma_conf->dma_tx_size = priv->dma_conf.dma_tx_size; in stmmac_setup_dma_desc()
3879 dma_conf->dma_rx_size = priv->dma_conf.dma_rx_size; in stmmac_setup_dma_desc()
3881 if (!dma_conf->dma_tx_size) in stmmac_setup_dma_desc()
3882 dma_conf->dma_tx_size = DMA_DEFAULT_TX_SIZE; in stmmac_setup_dma_desc()
3883 if (!dma_conf->dma_rx_size) in stmmac_setup_dma_desc()
3884 dma_conf->dma_rx_size = DMA_DEFAULT_RX_SIZE; in stmmac_setup_dma_desc()
3888 struct stmmac_tx_queue *tx_q = &dma_conf->tx_queue[chan]; in stmmac_setup_dma_desc()
3895 ret = alloc_dma_desc_resources(priv, dma_conf); in stmmac_setup_dma_desc()
3902 ret = init_dma_desc_rings(priv->dev, dma_conf, GFP_KERNEL); in stmmac_setup_dma_desc()
3909 return dma_conf; in stmmac_setup_dma_desc()
3912 free_dma_desc_resources(priv, dma_conf); in stmmac_setup_dma_desc()
3914 kfree(dma_conf); in stmmac_setup_dma_desc()
3929 struct stmmac_dma_conf *dma_conf) in __stmmac_open() argument
3953 buf_sz = dma_conf->dma_buf_sz; in __stmmac_open()
3955 if (priv->dma_conf.tx_queue[i].tbs & STMMAC_TBS_EN) in __stmmac_open()
3956 dma_conf->tx_queue[i].tbs = priv->dma_conf.tx_queue[i].tbs; in __stmmac_open()
3957 memcpy(&priv->dma_conf, dma_conf, sizeof(*dma_conf)); in __stmmac_open()
3997 hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); in __stmmac_open()
4010 struct stmmac_dma_conf *dma_conf; in stmmac_open() local
4013 dma_conf = stmmac_setup_dma_desc(priv, dev->mtu); in stmmac_open()
4014 if (IS_ERR(dma_conf)) in stmmac_open()
4015 return PTR_ERR(dma_conf); in stmmac_open()
4017 ret = __stmmac_open(dev, dma_conf); in stmmac_open()
4019 free_dma_desc_resources(priv, dma_conf); in stmmac_open()
4021 kfree(dma_conf); in stmmac_open()
4045 hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); in stmmac_release()
4061 free_dma_desc_resources(priv, &priv->dma_conf); in stmmac_release()
4107 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_vlan_insert()
4125 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tso_allocator()
4136 priv->dma_conf.dma_tx_size); in stmmac_tso_allocator()
4164 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_flush_tx_descriptors()
4241 tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_tso_xmit()
4282 priv->dma_conf.dma_tx_size); in stmmac_tso_xmit()
4399 tx_q->cur_tx = STMMAC_GET_ENTRY(tx_q->cur_tx, priv->dma_conf.dma_tx_size); in stmmac_tso_xmit()
4514 tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_xmit()
4599 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); in stmmac_xmit()
4669 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); in stmmac_xmit()
4786 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rx_refill()
4840 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_rx_size); in stmmac_rx_refill()
4868 return priv->dma_conf.dma_buf_sz; in stmmac_rx_buf1_len()
4873 return min_t(unsigned int, priv->dma_conf.dma_buf_sz, plen); in stmmac_rx_buf1_len()
4889 return priv->dma_conf.dma_buf_sz; in stmmac_rx_buf2_len()
4901 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_xdp_xmit_xdpf()
4973 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_tx_size); in stmmac_xdp_xmit_xdpf()
5154 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rx_refill_zc()
5197 entry = STMMAC_GET_ENTRY(entry, priv->dma_conf.dma_rx_size); in stmmac_rx_refill_zc()
5223 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rx_zc()
5246 stmmac_display_ring(priv, rx_head, priv->dma_conf.dma_rx_size, true, in stmmac_rx_zc()
5293 priv->dma_conf.dma_rx_size); in stmmac_rx_zc()
5416 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rx()
5429 buf_sz = DIV_ROUND_UP(priv->dma_conf.dma_buf_sz, PAGE_SIZE) * PAGE_SIZE; in stmmac_rx()
5430 limit = min(priv->dma_conf.dma_rx_size - 1, (unsigned int)limit); in stmmac_rx()
5444 stmmac_display_ring(priv, rx_head, priv->dma_conf.dma_rx_size, true, in stmmac_rx()
5487 priv->dma_conf.dma_rx_size); in stmmac_rx()
5618 priv->dma_conf.dma_buf_sz); in stmmac_rx()
5630 priv->dma_conf.dma_buf_sz); in stmmac_rx()
5855 struct stmmac_dma_conf *dma_conf; in stmmac_change_mtu() local
5878 dma_conf = stmmac_setup_dma_desc(priv, mtu); in stmmac_change_mtu()
5879 if (IS_ERR(dma_conf)) { in stmmac_change_mtu()
5882 return PTR_ERR(dma_conf); in stmmac_change_mtu()
5887 ret = __stmmac_open(dev, dma_conf); in stmmac_change_mtu()
5889 free_dma_desc_resources(priv, dma_conf); in stmmac_change_mtu()
5890 kfree(dma_conf); in stmmac_change_mtu()
5895 kfree(dma_conf); in stmmac_change_mtu()
6117 struct stmmac_dma_conf *dma_conf; in stmmac_msi_intr_tx() local
6122 dma_conf = container_of(tx_q, struct stmmac_dma_conf, tx_queue[chan]); in stmmac_msi_intr_tx()
6123 priv = container_of(dma_conf, struct stmmac_priv, dma_conf); in stmmac_msi_intr_tx()
6144 struct stmmac_dma_conf *dma_conf; in stmmac_msi_intr_rx() local
6148 dma_conf = container_of(rx_q, struct stmmac_dma_conf, rx_queue[chan]); in stmmac_msi_intr_rx()
6149 priv = container_of(dma_conf, struct stmmac_priv, dma_conf); in stmmac_msi_intr_rx()
6327 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_rings_status_show()
6334 priv->dma_conf.dma_rx_size, 1, seq, rx_q->dma_rx_phy); in stmmac_rings_status_show()
6338 priv->dma_conf.dma_rx_size, 0, seq, rx_q->dma_rx_phy); in stmmac_rings_status_show()
6343 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_rings_status_show()
6350 priv->dma_conf.dma_tx_size, 1, seq, tx_q->dma_tx_phy); in stmmac_rings_status_show()
6354 priv->dma_conf.dma_tx_size, 0, seq, tx_q->dma_tx_phy); in stmmac_rings_status_show()
6768 __free_dma_rx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_disable_rx_queue()
6773 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_enable_rx_queue()
6779 ret = __alloc_dma_rx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_enable_rx_queue()
6785 ret = __init_dma_rx_desc_rings(priv, &priv->dma_conf, queue, GFP_KERNEL); in stmmac_enable_rx_queue()
6787 __free_dma_rx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_enable_rx_queue()
6793 stmmac_clear_rx_descriptors(priv, &priv->dma_conf, queue); in stmmac_enable_rx_queue()
6810 priv->dma_conf.dma_buf_sz, in stmmac_enable_rx_queue()
6831 __free_dma_tx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_disable_tx_queue()
6836 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_enable_tx_queue()
6841 ret = __alloc_dma_tx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_enable_tx_queue()
6847 ret = __init_dma_tx_desc_rings(priv, &priv->dma_conf, queue); in stmmac_enable_tx_queue()
6849 __free_dma_tx_desc_resources(priv, &priv->dma_conf, queue); in stmmac_enable_tx_queue()
6855 stmmac_clear_tx_descriptors(priv, &priv->dma_conf, queue); in stmmac_enable_tx_queue()
6886 hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); in stmmac_xdp_release()
6895 free_dma_desc_resources(priv, &priv->dma_conf); in stmmac_xdp_release()
6920 ret = alloc_dma_desc_resources(priv, &priv->dma_conf); in stmmac_xdp_open()
6927 ret = init_dma_desc_rings(dev, &priv->dma_conf, GFP_KERNEL); in stmmac_xdp_open()
6947 rx_q = &priv->dma_conf.rx_queue[chan]; in stmmac_xdp_open()
6965 priv->dma_conf.dma_buf_sz, in stmmac_xdp_open()
6974 tx_q = &priv->dma_conf.tx_queue[chan]; in stmmac_xdp_open()
7007 hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); in stmmac_xdp_open()
7011 free_dma_desc_resources(priv, &priv->dma_conf); in stmmac_xdp_open()
7034 rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_xsk_wakeup()
7035 tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_xsk_wakeup()
7344 priv->dma_conf.dma_rx_size = rx_size; in stmmac_reinit_ringparam()
7345 priv->dma_conf.dma_tx_size = tx_size; in stmmac_reinit_ringparam()
7860 hrtimer_cancel(&priv->dma_conf.tx_queue[chan].txtimer); in stmmac_suspend()
7904 struct stmmac_rx_queue *rx_q = &priv->dma_conf.rx_queue[queue]; in stmmac_reset_rx_queue()
7912 struct stmmac_tx_queue *tx_q = &priv->dma_conf.tx_queue[queue]; in stmmac_reset_tx_queue()
7996 stmmac_clear_descriptors(priv, &priv->dma_conf); in stmmac_resume()