Home
last modified time | relevance | path

Searched refs:dma_tx (Results 1 – 25 of 31) sorted by relevance

12

/linux/drivers/spi/
H A Dspi-pxa2xx-dma.c98 chan = drv_data->controller->dma_tx; in pxa2xx_spi_dma_prepare_one()
127 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_transfer()
165 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_prepare()
173 dma_async_issue_pending(drv_data->controller->dma_tx); in pxa2xx_spi_dma_start()
182 dmaengine_terminate_sync(drv_data->controller->dma_tx); in pxa2xx_spi_dma_stop()
195 controller->dma_tx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup()
197 if (!controller->dma_tx) in pxa2xx_spi_dma_setup()
203 dma_release_channel(controller->dma_tx); in pxa2xx_spi_dma_setup()
204 controller->dma_tx = NULL; in pxa2xx_spi_dma_setup()
220 if (controller->dma_tx) { in pxa2xx_spi_dma_release()
[all …]
H A Dspi-bcm2835.c617 dmaengine_terminate_async(ctlr->dma_tx); in bcm2835_spi_dma_rx_done()
688 chan = ctlr->dma_tx; in bcm2835_spi_prepare_sg()
809 dma_async_issue_pending(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma()
823 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma()
867 if (ctlr->dma_tx) { in bcm2835_dma_release()
868 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_dma_release()
874 dma_unmap_page_attrs(ctlr->dma_tx->device->dev, in bcm2835_dma_release()
879 dma_release_channel(ctlr->dma_tx); in bcm2835_dma_release()
880 ctlr->dma_tx = NULL; in bcm2835_dma_release()
908 ctlr->dma_tx = dma_request_chan(dev, "tx"); in bcm2835_dma_init()
[all …]
H A Dspi-at91-usart.c134 ctlr->dma_tx = dma_request_chan(dev, "tx"); in at91_usart_spi_configure_dma()
135 if (IS_ERR(ctlr->dma_tx)) { in at91_usart_spi_configure_dma()
136 err = PTR_ERR(ctlr->dma_tx); in at91_usart_spi_configure_dma()
163 if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) { in at91_usart_spi_configure_dma()
174 if (!IS_ERR_OR_NULL(ctlr->dma_tx)) in at91_usart_spi_configure_dma()
175 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_configure_dma()
178 ctlr->dma_tx = NULL; in at91_usart_spi_configure_dma()
189 if (ctlr->dma_tx) in at91_usart_spi_release_dma()
190 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_release_dma()
197 if (ctlr->dma_tx) in at91_usart_spi_stop_dma()
[all …]
H A Dspi-sun6i.c248 dmaengine_slave_config(host->dma_tx, &txconf); in sun6i_spi_prepare_dma()
250 txdesc = dmaengine_prep_slave_sg(host->dma_tx, in sun6i_spi_prepare_dma()
269 dma_async_issue_pending(host->dma_tx); in sun6i_spi_prepare_dma()
523 dmaengine_terminate_sync(host->dma_tx); in sun6i_spi_transfer_one()
703 host->dma_tx = dma_request_chan(&pdev->dev, "tx"); in sun6i_spi_probe()
704 if (IS_ERR(host->dma_tx)) { in sun6i_spi_probe()
706 if (PTR_ERR(host->dma_tx) == -EPROBE_DEFER) { in sun6i_spi_probe()
711 host->dma_tx = NULL; in sun6i_spi_probe()
724 if (host->dma_tx && host->dma_rx) { in sun6i_spi_probe()
760 if (host->dma_tx) in sun6i_spi_probe()
[all …]
H A Dspi-fsl-lpspi.c406 ret = dmaengine_slave_config(controller->dma_tx, &tx); in fsl_lpspi_dma_configure()
651 desc_tx = dmaengine_prep_slave_sg(controller->dma_tx, in fsl_lpspi_dma_transfer()
655 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
663 dma_async_issue_pending(controller->dma_tx); in fsl_lpspi_dma_transfer()
676 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
686 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
696 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
706 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer()
725 if (controller->dma_tx) { in fsl_lpspi_dma_exit()
726 dma_release_channel(controller->dma_tx); in fsl_lpspi_dma_exit()
[all …]
H A Dspi-omap2-mcspi.c94 struct dma_chan *dma_tx; member
428 dmaengine_slave_config(mcspi_dma->dma_tx, &cfg); in omap2_mcspi_tx_dma()
430 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, in omap2_mcspi_tx_dma()
441 dma_async_issue_pending(mcspi_dma->dma_tx); in omap2_mcspi_tx_dma()
656 dmaengine_terminate_sync(mcspi_dma->dma_tx); in omap2_mcspi_txrx_dma()
1021 mcspi_dma->dma_tx = dma_request_chan(mcspi->dev, in omap2_mcspi_request_dma()
1023 if (IS_ERR(mcspi_dma->dma_tx)) { in omap2_mcspi_request_dma()
1024 ret = PTR_ERR(mcspi_dma->dma_tx); in omap2_mcspi_request_dma()
1025 mcspi_dma->dma_tx = NULL; in omap2_mcspi_request_dma()
1050 if (mcspi_dma->dma_tx) { in omap2_mcspi_release_dma()
[all …]
H A Dspi-davinci.c180 struct dma_chan *dma_tx; member
455 if (dspi->dma_rx && dspi->dma_tx) in davinci_spi_of_setup()
682 dmaengine_slave_config(dspi->dma_tx, &dma_tx_conf); in davinci_spi_bufs()
700 txdesc = dmaengine_prep_slave_sg(dspi->dma_tx, in davinci_spi_bufs()
718 dma_async_issue_pending(dspi->dma_tx); in davinci_spi_bufs()
817 dspi->dma_tx = dma_request_chan(sdev, "tx"); in davinci_spi_request_dma()
818 if (IS_ERR(dspi->dma_tx)) { in davinci_spi_request_dma()
820 return PTR_ERR(dspi->dma_tx); in davinci_spi_request_dma()
1016 dspi->dma_tx = NULL; in davinci_spi_probe()
1057 dma_release_channel(dspi->dma_tx); in davinci_spi_probe()
[all …]
H A Dspi-stm32.c373 struct dma_chan *dma_tx; member
862 if (spi->cur_usedma && spi->dma_tx) in stm32fx_spi_disable()
863 dmaengine_terminate_async(spi->dma_tx); in stm32fx_spi_disable()
905 if (spi->cur_usedma && spi->dma_tx) in stm32h7_spi_disable()
906 dmaengine_terminate_async(spi->dma_tx); in stm32h7_spi_disable()
1675 if (spi->tx_buf && spi->dma_tx) { in stm32_spi_transfer_one_dma()
1676 stm32_spi_dma_config(spi, spi->dma_tx, &tx_dma_conf, DMA_MEM_TO_DEV); in stm32_spi_transfer_one_dma()
1677 dmaengine_slave_config(spi->dma_tx, &tx_dma_conf); in stm32_spi_transfer_one_dma()
1678 tx_dma_desc = dmaengine_prep_slave_sg(spi->dma_tx, xfer->tx_sg.sgl, in stm32_spi_transfer_one_dma()
1684 if ((spi->tx_buf && spi->dma_tx && !tx_dma_desc) || in stm32_spi_transfer_one_dma()
[all …]
H A Dspi-atmel.c539 if (dmaengine_slave_config(host->dma_tx, &slave_config)) { in atmel_spi_dma_slave_config()
568 host->dma_tx = dma_request_chan(dev, "tx"); in atmel_spi_configure_dma()
569 if (IS_ERR(host->dma_tx)) { in atmel_spi_configure_dma()
570 err = PTR_ERR(host->dma_tx); in atmel_spi_configure_dma()
592 dma_chan_name(host->dma_tx), in atmel_spi_configure_dma()
599 if (!IS_ERR(host->dma_tx)) in atmel_spi_configure_dma()
600 dma_release_channel(host->dma_tx); in atmel_spi_configure_dma()
602 host->dma_tx = host->dma_rx = NULL; in atmel_spi_configure_dma()
610 if (host->dma_tx) in atmel_spi_stop_dma()
611 dmaengine_terminate_all(host->dma_tx); in atmel_spi_stop_dma()
[all …]
H A Dspi-imx.c1399 if (controller->dma_tx) { in spi_imx_sdma_exit()
1400 dma_release_channel(controller->dma_tx); in spi_imx_sdma_exit()
1401 controller->dma_tx = NULL; in spi_imx_sdma_exit()
1413 controller->dma_tx = dma_request_chan(dev, "tx"); in spi_imx_sdma_init()
1414 if (IS_ERR(controller->dma_tx)) { in spi_imx_sdma_init()
1415 ret = PTR_ERR(controller->dma_tx); in spi_imx_sdma_init()
1417 controller->dma_tx = NULL; in spi_imx_sdma_init()
1472 struct device *tx_dev = spi_imx->controller->dma_tx->device->dev; in spi_imx_dma_unmap()
1531 struct device *tx_dev = controller->dma_tx->device->dev; in spi_imx_dma_map()
1752 desc_tx = dmaengine_prep_slave_single(controller->dma_tx, dma_data->dma_tx_addr, in spi_imx_dma_submit()
[all …]
H A Dspi-atcspi200.c280 chan = spi->host->dma_tx; in atcspi_dma_config()
327 dma_ch = spi->host->dma_tx; in atcspi_dma_trans()
512 spi->host->dma_tx = devm_dma_request_chan(spi->dev, "tx"); in atcspi_configure_dma()
513 if (IS_ERR(spi->host->dma_tx)) in atcspi_configure_dma()
514 return PTR_ERR(spi->host->dma_tx); in atcspi_configure_dma()
H A Dspi-dw-dma.c113 struct dw_dma_slave dma_tx = { .dst_id = 1 }, *tx = &dma_tx; in dw_spi_dma_init_mfld() local
143 dws->ctlr->dma_tx = dws->txchan; in dw_spi_dma_init_mfld()
187 dws->ctlr->dma_tx = dws->txchan; in dw_spi_dma_init_generic()
H A Dspi-mem.c48 if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx) in spi_controller_dma_map_mem_op_data()
49 dmadev = ctlr->dma_tx->device->dev; in spi_controller_dma_map_mem_op_data()
94 if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx) in spi_controller_dma_unmap_mem_op_data()
95 dmadev = ctlr->dma_tx->device->dev; in spi_controller_dma_unmap_mem_op_data()
/linux/arch/powerpc/platforms/512x/
H A Dmpc512x_lpbfifo.c169 struct dma_async_tx_descriptor *dma_tx = NULL; in mpc512x_lpbfifo_kick() local
257 dma_tx = dmaengine_prep_slave_sg(lpbfifo.chan, &sg, in mpc512x_lpbfifo_kick()
259 if (!dma_tx) { in mpc512x_lpbfifo_kick()
263 dma_tx->callback = mpc512x_lpbfifo_callback; in mpc512x_lpbfifo_kick()
264 dma_tx->callback_param = NULL; in mpc512x_lpbfifo_kick()
313 cookie = dma_tx->tx_submit(dma_tx); in mpc512x_lpbfifo_kick()
/linux/drivers/mmc/host/
H A Djz4740_mmc.c179 struct dma_chan *dma_tx; member
223 dma_release_channel(host->dma_tx); in jz4740_mmc_release_dma_channels()
232 host->dma_tx = dma_request_chan(dev, "tx-rx"); in jz4740_mmc_acquire_dma_channels()
233 if (!IS_ERR(host->dma_tx)) in jz4740_mmc_acquire_dma_channels()
236 if (PTR_ERR(host->dma_tx) != -ENODEV) { in jz4740_mmc_acquire_dma_channels()
238 return PTR_ERR(host->dma_tx); in jz4740_mmc_acquire_dma_channels()
241 host->dma_tx = dma_request_chan(mmc_dev(host->mmc), "tx"); in jz4740_mmc_acquire_dma_channels()
242 if (IS_ERR(host->dma_tx)) { in jz4740_mmc_acquire_dma_channels()
244 return PTR_ERR(host->dma_tx); in jz4740_mmc_acquire_dma_channels()
250 dma_release_channel(host->dma_tx); in jz4740_mmc_acquire_dma_channels()
[all …]
/linux/drivers/mtd/nand/raw/
H A Dintel-nand-controller.c118 struct dma_chan *dma_tx; member
282 dmaengine_terminate_async(ebu_host->dma_tx); in ebu_dma_tx_callback()
305 chan = ebu_host->dma_tx; in ebu_dma_start()
572 if (ebu_host->dma_tx) in ebu_dma_cleanup()
573 dma_release_channel(ebu_host->dma_tx); in ebu_dma_cleanup()
641 ebu_host->dma_tx = dma_request_chan(dev, "tx"); in ebu_nand_probe()
642 if (IS_ERR(ebu_host->dma_tx)) { in ebu_nand_probe()
643 ret = dev_err_probe(dev, PTR_ERR(ebu_host->dma_tx), in ebu_nand_probe()
/linux/drivers/net/ethernet/samsung/sxgbe/
H A Dsxgbe_dma.c44 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument
66 writel(upper_32_bits(dma_tx), in sxgbe_dma_channel_init()
68 writel(lower_32_bits(dma_tx), in sxgbe_dma_channel_init()
80 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
H A Dsxgbe_main.c333 priv->hw->desc->init_tx_desc(&priv->txq[j]->dma_tx[i]); in sxgbe_clear_descriptors()
404 tx_ring->dma_tx = dma_alloc_coherent(dev, in init_tx_ring()
407 if (!tx_ring->dma_tx) in init_tx_ring()
433 tx_ring->dma_tx, tx_ring->dma_tx_phy); in init_tx_ring()
549 tx_ring->dma_tx, tx_ring->dma_tx_phy); in free_tx_ring()
618 struct sxgbe_tx_norm_desc *tdesc = txqueue->dma_tx + dma_desc; in tx_free_ring_skbufs()
750 p = tqueue->dma_tx + entry; in sxgbe_tx_queue_clean()
1314 tx_desc = tqueue->dma_tx + entry; in sxgbe_xmit()
1339 first_desc = tqueue->dma_tx + entry; in sxgbe_xmit()
1361 tx_desc = tqueue->dma_tx + entry; in sxgbe_xmit()
H A Dsxgbe_dma.h24 int pbl, dma_addr_t dma_tx, dma_addr_t dma_rx,
/linux/drivers/i2c/busses/
H A Di2c-sh_mobile.c136 struct dma_chan *dma_tx; member
449 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_cleanup_dma()
512 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma()
522 chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV, in sh_mobile_i2c_xfer_dma()
815 if (!IS_ERR(pd->dma_tx)) { in sh_mobile_i2c_release_dma()
816 dma_release_channel(pd->dma_tx); in sh_mobile_i2c_release_dma()
817 pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma()
920 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
H A Di2c-rcar.c162 struct dma_chan *dma_tx; member
450 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_cleanup_dma()
485 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma()
864 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma()
874 priv->dma_tx = chan; in rcar_i2c_request_dma()
879 if (!IS_ERR(priv->dma_tx)) { in rcar_i2c_release_dma()
880 dma_release_channel(priv->dma_tx); in rcar_i2c_release_dma()
881 priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma()
1162 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
/linux/drivers/net/ethernet/calxeda/
H A Dxgmac.c360 struct xgmac_dma_desc *dma_tx; member
748 priv->dma_tx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init()
753 if (!priv->dma_tx) in xgmac_dma_desc_rings_init()
758 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init()
769 desc_init_tx_desc(priv->dma_tx, DMA_TX_RING_SZ); in xgmac_dma_desc_rings_init()
820 p = priv->dma_tx + i; in xgmac_free_tx_skbufs()
841 if (priv->dma_tx) { in xgmac_free_dma_desc_rings()
844 priv->dma_tx, priv->dma_tx_phy); in xgmac_free_dma_desc_rings()
845 priv->dma_tx = NULL; in xgmac_free_dma_desc_rings()
869 struct xgmac_dma_desc *p = priv->dma_tx + entry; in xgmac_tx_complete()
[all …]
/linux/drivers/crypto/
H A Dsa2ul.c1127 ddev = dmaengine_get_dma_device(pdata->dma_tx); in sa_run()
1260 tx_out = dmaengine_prep_slave_sg(pdata->dma_tx, src, in sa_run()
1287 dma_async_issue_pending(pdata->dma_tx); in sa_run()
2254 dd->dma_tx = NULL; in sa_dma_init()
2273 dd->dma_tx = dma_request_chan(dd->dev, "tx"); in sa_dma_init()
2274 if (IS_ERR(dd->dma_tx)) { in sa_dma_init()
2275 ret = dev_err_probe(dd->dev, PTR_ERR(dd->dma_tx), in sa_dma_init()
2301 ret = dmaengine_slave_config(dd->dma_tx, &cfg); in sa_dma_init()
2311 dma_release_channel(dd->dma_tx); in sa_dma_init()
2428 dma_release_channel(dev_data->dma_tx); in sa_ul_probe()
[all …]
H A Dsa2ul.h210 struct dma_chan *dma_tx; member
/linux/drivers/net/ethernet/stmicro/stmmac/
H A Dstmmac_main.c1461 head_tx = (void *)tx_q->dma_tx; in stmmac_display_tx_rings()
1560 p = &tx_q->dma_tx[i]; in stmmac_clear_tx_descriptors()
1957 stmmac_mode_init(priv, tx_q->dma_tx, in __init_dma_tx_desc_rings()
1972 p = tx_q->dma_tx + i; in __init_dma_tx_desc_rings()
2144 addr = tx_q->dma_tx; in __free_dma_tx_desc_resources()
2328 tx_q->dma_tx = addr; in __alloc_dma_tx_desc_resources()
2694 tx_desc = tx_q->dma_tx + entry; in stmmac_xdp_xmit_zc()
2831 p = tx_q->dma_tx + entry; in stmmac_tx_clean()
4255 p = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_vlan_insert()
4296 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_allocator()
[all …]

12