| /linux/drivers/spi/ |
| H A D | spi-pxa2xx-dma.c | 98 chan = drv_data->controller->dma_tx; in pxa2xx_spi_dma_prepare_one() 127 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_transfer() 165 dmaengine_terminate_async(drv_data->controller->dma_tx); in pxa2xx_spi_dma_prepare() 173 dma_async_issue_pending(drv_data->controller->dma_tx); in pxa2xx_spi_dma_start() 182 dmaengine_terminate_sync(drv_data->controller->dma_tx); in pxa2xx_spi_dma_stop() 195 controller->dma_tx = dma_request_slave_channel_compat(mask, in pxa2xx_spi_dma_setup() 197 if (!controller->dma_tx) in pxa2xx_spi_dma_setup() 203 dma_release_channel(controller->dma_tx); in pxa2xx_spi_dma_setup() 204 controller->dma_tx = NULL; in pxa2xx_spi_dma_setup() 220 if (controller->dma_tx) { in pxa2xx_spi_dma_release() [all …]
|
| H A D | spi-bcm2835.c | 617 dmaengine_terminate_async(ctlr->dma_tx); in bcm2835_spi_dma_rx_done() 688 chan = ctlr->dma_tx; in bcm2835_spi_prepare_sg() 809 dma_async_issue_pending(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma() 823 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_spi_transfer_one_dma() 867 if (ctlr->dma_tx) { in bcm2835_dma_release() 868 dmaengine_terminate_sync(ctlr->dma_tx); in bcm2835_dma_release() 874 dma_unmap_page_attrs(ctlr->dma_tx->device->dev, in bcm2835_dma_release() 879 dma_release_channel(ctlr->dma_tx); in bcm2835_dma_release() 880 ctlr->dma_tx = NULL; in bcm2835_dma_release() 908 ctlr->dma_tx = dma_request_chan(dev, "tx"); in bcm2835_dma_init() [all …]
|
| H A D | spi-at91-usart.c | 134 ctlr->dma_tx = dma_request_chan(dev, "tx"); in at91_usart_spi_configure_dma() 135 if (IS_ERR(ctlr->dma_tx)) { in at91_usart_spi_configure_dma() 136 err = PTR_ERR(ctlr->dma_tx); in at91_usart_spi_configure_dma() 163 if (dmaengine_slave_config(ctlr->dma_tx, &slave_config)) { in at91_usart_spi_configure_dma() 174 if (!IS_ERR_OR_NULL(ctlr->dma_tx)) in at91_usart_spi_configure_dma() 175 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_configure_dma() 178 ctlr->dma_tx = NULL; in at91_usart_spi_configure_dma() 189 if (ctlr->dma_tx) in at91_usart_spi_release_dma() 190 dma_release_channel(ctlr->dma_tx); in at91_usart_spi_release_dma() 197 if (ctlr->dma_tx) in at91_usart_spi_stop_dma() [all …]
|
| H A D | spi-sun6i.c | 248 dmaengine_slave_config(host->dma_tx, &txconf); in sun6i_spi_prepare_dma() 250 txdesc = dmaengine_prep_slave_sg(host->dma_tx, in sun6i_spi_prepare_dma() 269 dma_async_issue_pending(host->dma_tx); in sun6i_spi_prepare_dma() 523 dmaengine_terminate_sync(host->dma_tx); in sun6i_spi_transfer_one() 703 host->dma_tx = dma_request_chan(&pdev->dev, "tx"); in sun6i_spi_probe() 704 if (IS_ERR(host->dma_tx)) { in sun6i_spi_probe() 706 if (PTR_ERR(host->dma_tx) == -EPROBE_DEFER) { in sun6i_spi_probe() 711 host->dma_tx = NULL; in sun6i_spi_probe() 724 if (host->dma_tx && host->dma_rx) { in sun6i_spi_probe() 760 if (host->dma_tx) in sun6i_spi_probe() [all …]
|
| H A D | spi-fsl-lpspi.c | 406 ret = dmaengine_slave_config(controller->dma_tx, &tx); in fsl_lpspi_dma_configure() 651 desc_tx = dmaengine_prep_slave_sg(controller->dma_tx, in fsl_lpspi_dma_transfer() 655 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer() 663 dma_async_issue_pending(controller->dma_tx); in fsl_lpspi_dma_transfer() 676 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer() 686 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer() 696 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer() 706 dmaengine_terminate_all(controller->dma_tx); in fsl_lpspi_dma_transfer() 725 if (controller->dma_tx) { in fsl_lpspi_dma_exit() 726 dma_release_channel(controller->dma_tx); in fsl_lpspi_dma_exit() [all …]
|
| H A D | spi-omap2-mcspi.c | 94 struct dma_chan *dma_tx; member 428 dmaengine_slave_config(mcspi_dma->dma_tx, &cfg); in omap2_mcspi_tx_dma() 430 tx = dmaengine_prep_slave_sg(mcspi_dma->dma_tx, xfer->tx_sg.sgl, in omap2_mcspi_tx_dma() 441 dma_async_issue_pending(mcspi_dma->dma_tx); in omap2_mcspi_tx_dma() 656 dmaengine_terminate_sync(mcspi_dma->dma_tx); in omap2_mcspi_txrx_dma() 1021 mcspi_dma->dma_tx = dma_request_chan(mcspi->dev, in omap2_mcspi_request_dma() 1023 if (IS_ERR(mcspi_dma->dma_tx)) { in omap2_mcspi_request_dma() 1024 ret = PTR_ERR(mcspi_dma->dma_tx); in omap2_mcspi_request_dma() 1025 mcspi_dma->dma_tx = NULL; in omap2_mcspi_request_dma() 1050 if (mcspi_dma->dma_tx) { in omap2_mcspi_release_dma() [all …]
|
| H A D | spi-davinci.c | 180 struct dma_chan *dma_tx; member 455 if (dspi->dma_rx && dspi->dma_tx) in davinci_spi_of_setup() 682 dmaengine_slave_config(dspi->dma_tx, &dma_tx_conf); in davinci_spi_bufs() 700 txdesc = dmaengine_prep_slave_sg(dspi->dma_tx, in davinci_spi_bufs() 718 dma_async_issue_pending(dspi->dma_tx); in davinci_spi_bufs() 817 dspi->dma_tx = dma_request_chan(sdev, "tx"); in davinci_spi_request_dma() 818 if (IS_ERR(dspi->dma_tx)) { in davinci_spi_request_dma() 820 return PTR_ERR(dspi->dma_tx); in davinci_spi_request_dma() 1016 dspi->dma_tx = NULL; in davinci_spi_probe() 1057 dma_release_channel(dspi->dma_tx); in davinci_spi_probe() [all …]
|
| H A D | spi-stm32.c | 373 struct dma_chan *dma_tx; member 862 if (spi->cur_usedma && spi->dma_tx) in stm32fx_spi_disable() 863 dmaengine_terminate_async(spi->dma_tx); in stm32fx_spi_disable() 905 if (spi->cur_usedma && spi->dma_tx) in stm32h7_spi_disable() 906 dmaengine_terminate_async(spi->dma_tx); in stm32h7_spi_disable() 1675 if (spi->tx_buf && spi->dma_tx) { in stm32_spi_transfer_one_dma() 1676 stm32_spi_dma_config(spi, spi->dma_tx, &tx_dma_conf, DMA_MEM_TO_DEV); in stm32_spi_transfer_one_dma() 1677 dmaengine_slave_config(spi->dma_tx, &tx_dma_conf); in stm32_spi_transfer_one_dma() 1678 tx_dma_desc = dmaengine_prep_slave_sg(spi->dma_tx, xfer->tx_sg.sgl, in stm32_spi_transfer_one_dma() 1684 if ((spi->tx_buf && spi->dma_tx && !tx_dma_desc) || in stm32_spi_transfer_one_dma() [all …]
|
| H A D | spi-atmel.c | 539 if (dmaengine_slave_config(host->dma_tx, &slave_config)) { in atmel_spi_dma_slave_config() 568 host->dma_tx = dma_request_chan(dev, "tx"); in atmel_spi_configure_dma() 569 if (IS_ERR(host->dma_tx)) { in atmel_spi_configure_dma() 570 err = PTR_ERR(host->dma_tx); in atmel_spi_configure_dma() 592 dma_chan_name(host->dma_tx), in atmel_spi_configure_dma() 599 if (!IS_ERR(host->dma_tx)) in atmel_spi_configure_dma() 600 dma_release_channel(host->dma_tx); in atmel_spi_configure_dma() 602 host->dma_tx = host->dma_rx = NULL; in atmel_spi_configure_dma() 610 if (host->dma_tx) in atmel_spi_stop_dma() 611 dmaengine_terminate_all(host->dma_tx); in atmel_spi_stop_dma() [all …]
|
| H A D | spi-imx.c | 1399 if (controller->dma_tx) { in spi_imx_sdma_exit() 1400 dma_release_channel(controller->dma_tx); in spi_imx_sdma_exit() 1401 controller->dma_tx = NULL; in spi_imx_sdma_exit() 1413 controller->dma_tx = dma_request_chan(dev, "tx"); in spi_imx_sdma_init() 1414 if (IS_ERR(controller->dma_tx)) { in spi_imx_sdma_init() 1415 ret = PTR_ERR(controller->dma_tx); in spi_imx_sdma_init() 1417 controller->dma_tx = NULL; in spi_imx_sdma_init() 1472 struct device *tx_dev = spi_imx->controller->dma_tx->device->dev; in spi_imx_dma_unmap() 1531 struct device *tx_dev = controller->dma_tx->device->dev; in spi_imx_dma_map() 1752 desc_tx = dmaengine_prep_slave_single(controller->dma_tx, dma_data->dma_tx_addr, in spi_imx_dma_submit() [all …]
|
| H A D | spi-atcspi200.c | 280 chan = spi->host->dma_tx; in atcspi_dma_config() 327 dma_ch = spi->host->dma_tx; in atcspi_dma_trans() 512 spi->host->dma_tx = devm_dma_request_chan(spi->dev, "tx"); in atcspi_configure_dma() 513 if (IS_ERR(spi->host->dma_tx)) in atcspi_configure_dma() 514 return PTR_ERR(spi->host->dma_tx); in atcspi_configure_dma()
|
| H A D | spi-dw-dma.c | 113 struct dw_dma_slave dma_tx = { .dst_id = 1 }, *tx = &dma_tx; in dw_spi_dma_init_mfld() local 143 dws->ctlr->dma_tx = dws->txchan; in dw_spi_dma_init_mfld() 187 dws->ctlr->dma_tx = dws->txchan; in dw_spi_dma_init_generic()
|
| H A D | spi-mem.c | 48 if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx) in spi_controller_dma_map_mem_op_data() 49 dmadev = ctlr->dma_tx->device->dev; in spi_controller_dma_map_mem_op_data() 94 if (op->data.dir == SPI_MEM_DATA_OUT && ctlr->dma_tx) in spi_controller_dma_unmap_mem_op_data() 95 dmadev = ctlr->dma_tx->device->dev; in spi_controller_dma_unmap_mem_op_data()
|
| /linux/arch/powerpc/platforms/512x/ |
| H A D | mpc512x_lpbfifo.c | 169 struct dma_async_tx_descriptor *dma_tx = NULL; in mpc512x_lpbfifo_kick() local 257 dma_tx = dmaengine_prep_slave_sg(lpbfifo.chan, &sg, in mpc512x_lpbfifo_kick() 259 if (!dma_tx) { in mpc512x_lpbfifo_kick() 263 dma_tx->callback = mpc512x_lpbfifo_callback; in mpc512x_lpbfifo_kick() 264 dma_tx->callback_param = NULL; in mpc512x_lpbfifo_kick() 313 cookie = dma_tx->tx_submit(dma_tx); in mpc512x_lpbfifo_kick()
|
| /linux/drivers/mmc/host/ |
| H A D | jz4740_mmc.c | 179 struct dma_chan *dma_tx; member 223 dma_release_channel(host->dma_tx); in jz4740_mmc_release_dma_channels() 232 host->dma_tx = dma_request_chan(dev, "tx-rx"); in jz4740_mmc_acquire_dma_channels() 233 if (!IS_ERR(host->dma_tx)) in jz4740_mmc_acquire_dma_channels() 236 if (PTR_ERR(host->dma_tx) != -ENODEV) { in jz4740_mmc_acquire_dma_channels() 238 return PTR_ERR(host->dma_tx); in jz4740_mmc_acquire_dma_channels() 241 host->dma_tx = dma_request_chan(mmc_dev(host->mmc), "tx"); in jz4740_mmc_acquire_dma_channels() 242 if (IS_ERR(host->dma_tx)) { in jz4740_mmc_acquire_dma_channels() 244 return PTR_ERR(host->dma_tx); in jz4740_mmc_acquire_dma_channels() 250 dma_release_channel(host->dma_tx); in jz4740_mmc_acquire_dma_channels() [all …]
|
| /linux/drivers/mtd/nand/raw/ |
| H A D | intel-nand-controller.c | 118 struct dma_chan *dma_tx; member 282 dmaengine_terminate_async(ebu_host->dma_tx); in ebu_dma_tx_callback() 305 chan = ebu_host->dma_tx; in ebu_dma_start() 572 if (ebu_host->dma_tx) in ebu_dma_cleanup() 573 dma_release_channel(ebu_host->dma_tx); in ebu_dma_cleanup() 641 ebu_host->dma_tx = dma_request_chan(dev, "tx"); in ebu_nand_probe() 642 if (IS_ERR(ebu_host->dma_tx)) { in ebu_nand_probe() 643 ret = dev_err_probe(dev, PTR_ERR(ebu_host->dma_tx), in ebu_nand_probe()
|
| /linux/drivers/net/ethernet/samsung/sxgbe/ |
| H A D | sxgbe_dma.c | 44 int fix_burst, int pbl, dma_addr_t dma_tx, in sxgbe_dma_channel_init() argument 66 writel(upper_32_bits(dma_tx), in sxgbe_dma_channel_init() 68 writel(lower_32_bits(dma_tx), in sxgbe_dma_channel_init() 80 dma_addr = dma_tx + ((t_rsize - 1) * SXGBE_DESC_SIZE_BYTES); in sxgbe_dma_channel_init()
|
| H A D | sxgbe_main.c | 333 priv->hw->desc->init_tx_desc(&priv->txq[j]->dma_tx[i]); in sxgbe_clear_descriptors() 404 tx_ring->dma_tx = dma_alloc_coherent(dev, in init_tx_ring() 407 if (!tx_ring->dma_tx) in init_tx_ring() 433 tx_ring->dma_tx, tx_ring->dma_tx_phy); in init_tx_ring() 549 tx_ring->dma_tx, tx_ring->dma_tx_phy); in free_tx_ring() 618 struct sxgbe_tx_norm_desc *tdesc = txqueue->dma_tx + dma_desc; in tx_free_ring_skbufs() 750 p = tqueue->dma_tx + entry; in sxgbe_tx_queue_clean() 1314 tx_desc = tqueue->dma_tx + entry; in sxgbe_xmit() 1339 first_desc = tqueue->dma_tx + entry; in sxgbe_xmit() 1361 tx_desc = tqueue->dma_tx + entry; in sxgbe_xmit()
|
| H A D | sxgbe_dma.h | 24 int pbl, dma_addr_t dma_tx, dma_addr_t dma_rx,
|
| /linux/drivers/i2c/busses/ |
| H A D | i2c-sh_mobile.c | 136 struct dma_chan *dma_tx; member 449 ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_cleanup_dma() 512 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx; in sh_mobile_i2c_xfer_dma() 522 chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV, in sh_mobile_i2c_xfer_dma() 815 if (!IS_ERR(pd->dma_tx)) { in sh_mobile_i2c_release_dma() 816 dma_release_channel(pd->dma_tx); in sh_mobile_i2c_release_dma() 817 pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_release_dma() 920 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER); in sh_mobile_i2c_probe()
|
| H A D | i2c-rcar.c | 162 struct dma_chan *dma_tx; member 450 ? priv->dma_rx : priv->dma_tx; in rcar_i2c_cleanup_dma() 485 struct dma_chan *chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_dma() 864 chan = read ? priv->dma_rx : priv->dma_tx; in rcar_i2c_request_dma() 874 priv->dma_tx = chan; in rcar_i2c_request_dma() 879 if (!IS_ERR(priv->dma_tx)) { in rcar_i2c_release_dma() 880 dma_release_channel(priv->dma_tx); in rcar_i2c_release_dma() 881 priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_release_dma() 1162 priv->dma_rx = priv->dma_tx = ERR_PTR(-EPROBE_DEFER); in rcar_i2c_probe()
|
| /linux/drivers/net/ethernet/calxeda/ |
| H A D | xgmac.c | 360 struct xgmac_dma_desc *dma_tx; member 748 priv->dma_tx = dma_alloc_coherent(priv->device, in xgmac_dma_desc_rings_init() 753 if (!priv->dma_tx) in xgmac_dma_desc_rings_init() 758 priv->dma_rx, priv->dma_tx, in xgmac_dma_desc_rings_init() 769 desc_init_tx_desc(priv->dma_tx, DMA_TX_RING_SZ); in xgmac_dma_desc_rings_init() 820 p = priv->dma_tx + i; in xgmac_free_tx_skbufs() 841 if (priv->dma_tx) { in xgmac_free_dma_desc_rings() 844 priv->dma_tx, priv->dma_tx_phy); in xgmac_free_dma_desc_rings() 845 priv->dma_tx = NULL; in xgmac_free_dma_desc_rings() 869 struct xgmac_dma_desc *p = priv->dma_tx + entry; in xgmac_tx_complete() [all …]
|
| /linux/drivers/crypto/ |
| H A D | sa2ul.c | 1127 ddev = dmaengine_get_dma_device(pdata->dma_tx); in sa_run() 1260 tx_out = dmaengine_prep_slave_sg(pdata->dma_tx, src, in sa_run() 1287 dma_async_issue_pending(pdata->dma_tx); in sa_run() 2254 dd->dma_tx = NULL; in sa_dma_init() 2273 dd->dma_tx = dma_request_chan(dd->dev, "tx"); in sa_dma_init() 2274 if (IS_ERR(dd->dma_tx)) { in sa_dma_init() 2275 ret = dev_err_probe(dd->dev, PTR_ERR(dd->dma_tx), in sa_dma_init() 2301 ret = dmaengine_slave_config(dd->dma_tx, &cfg); in sa_dma_init() 2311 dma_release_channel(dd->dma_tx); in sa_dma_init() 2428 dma_release_channel(dev_data->dma_tx); in sa_ul_probe() [all …]
|
| H A D | sa2ul.h | 210 struct dma_chan *dma_tx; member
|
| /linux/drivers/net/ethernet/stmicro/stmmac/ |
| H A D | stmmac_main.c | 1461 head_tx = (void *)tx_q->dma_tx; in stmmac_display_tx_rings() 1560 p = &tx_q->dma_tx[i]; in stmmac_clear_tx_descriptors() 1957 stmmac_mode_init(priv, tx_q->dma_tx, in __init_dma_tx_desc_rings() 1972 p = tx_q->dma_tx + i; in __init_dma_tx_desc_rings() 2144 addr = tx_q->dma_tx; in __free_dma_tx_desc_resources() 2328 tx_q->dma_tx = addr; in __alloc_dma_tx_desc_resources() 2694 tx_desc = tx_q->dma_tx + entry; in stmmac_xdp_xmit_zc() 2831 p = tx_q->dma_tx + entry; in stmmac_tx_clean() 4255 p = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_vlan_insert() 4296 desc = &tx_q->dma_tx[tx_q->cur_tx]; in stmmac_tso_allocator() [all …]
|