Home
last modified time | relevance | path

Searched refs:tx_dma (Results 1 – 24 of 24) sorted by relevance

/linux/drivers/tty/serial/
H A Dsprd_serial.c134 struct sprd_uart_dma tx_dma; member
234 dmaengine_pause(sp->tx_dma.chn); in sprd_stop_tx_dma()
236 dmaengine_tx_status(sp->tx_dma.chn, sp->tx_dma.cookie, &state); in sprd_stop_tx_dma()
238 trans_len = state.residue - sp->tx_dma.phys_addr; in sprd_stop_tx_dma()
240 dma_unmap_single(port->dev, sp->tx_dma.phys_addr, in sprd_stop_tx_dma()
241 sp->tx_dma.trans_len, DMA_TO_DEVICE); in sprd_stop_tx_dma()
244 dmaengine_terminate_all(sp->tx_dma.chn); in sprd_stop_tx_dma()
245 sp->tx_dma.trans_len = 0; in sprd_stop_tx_dma()
255 sp->tx_dma.trans_len = kfifo_out_linear_ptr(&tport->xmit_fifo, &tail, in sprd_tx_buf_remap()
258 sp->tx_dma.phys_addr = dma_map_single(port->dev, tail, in sprd_tx_buf_remap()
[all …]
/linux/drivers/spi/
H A Dspi-fsl-cpm.c89 if (mspi->tx_dma == mspi->dma_dummy_tx) in fsl_spi_cpm_bufs_start()
90 iowrite32be(mspi->tx_dma, &tx_bd->cbd_bufaddr); in fsl_spi_cpm_bufs_start()
92 iowrite32be(mspi->tx_dma + xfer_ofs, &tx_bd->cbd_bufaddr); in fsl_spi_cpm_bufs_start()
110 mspi->tx_dma = mspi->dma_dummy_tx; in fsl_spi_cpm_bufs()
137 mspi->tx_dma = dma_map_single(dev, nonconst_tx, t->len, in fsl_spi_cpm_bufs()
139 if (dma_mapping_error(dev, mspi->tx_dma)) { in fsl_spi_cpm_bufs()
144 mspi->tx_dma = 0; in fsl_spi_cpm_bufs()
171 dma_unmap_single(dev, mspi->tx_dma, t->len, DMA_TO_DEVICE); in fsl_spi_cpm_bufs()
182 dma_unmap_single(dev, mspi->tx_dma, t->len, DMA_TO_DEVICE); in fsl_spi_cpm_bufs_complete()
H A Dspi-s3c64xx.c221 struct s3c64xx_spi_dma_data tx_dma; member
288 struct s3c64xx_spi_driver_data, tx_dma); in s3c64xx_spi_dmacb()
323 struct s3c64xx_spi_driver_data, tx_dma); in s3c64xx_prepare_dma()
395 sdd->tx_dma.ch = dma_request_chan(&sdd->pdev->dev, "tx"); in s3c64xx_spi_prepare_transfer()
396 if (IS_ERR(sdd->tx_dma.ch)) { in s3c64xx_spi_prepare_transfer()
399 sdd->tx_dma.ch = NULL; in s3c64xx_spi_prepare_transfer()
405 spi->dma_tx = sdd->tx_dma.ch; in s3c64xx_spi_prepare_transfer()
418 if (sdd->rx_dma.ch && sdd->tx_dma.ch) { in s3c64xx_spi_unprepare_transfer()
420 dma_release_channel(sdd->tx_dma.ch); in s3c64xx_spi_unprepare_transfer()
422 sdd->tx_dma.ch = NULL; in s3c64xx_spi_unprepare_transfer()
[all …]
H A Dspi-atmel.c823 xfer, xfer->len, xfer->tx_buf, (unsigned long long)xfer->tx_dma, in atmel_spi_next_xfer_dma_submit()
854 dma_addr_t *tx_dma, in atmel_spi_next_xfer_data() argument
859 *tx_dma = xfer->tx_dma + xfer->len - *plen; in atmel_spi_next_xfer_data()
921 dma_addr_t tx_dma, rx_dma; in atmel_spi_pdc_next_xfer() local
926 atmel_spi_next_xfer_data(host, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer()
930 spi_writel(as, TPR, tx_dma); in atmel_spi_pdc_next_xfer()
940 (unsigned long long)xfer->tx_dma, xfer->rx_buf, in atmel_spi_pdc_next_xfer()
945 atmel_spi_next_xfer_data(host, xfer, &tx_dma, &rx_dma, &len); in atmel_spi_pdc_next_xfer()
949 spi_writel(as, TNPR, tx_dma); in atmel_spi_pdc_next_xfer()
959 (unsigned long long)xfer->tx_dma, xfer->rx_buf, in atmel_spi_pdc_next_xfer()
[all …]
H A Dspi-mt65xx.c176 dma_addr_t tx_dma; member
644 writel((u32)(xfer->tx_dma & MTK_SPI_32BITS_MASK), in mtk_spi_setup_dma_addr()
648 writel((u32)(xfer->tx_dma >> 32), in mtk_spi_setup_dma_addr()
723 xfer->tx_dma = sg_dma_address(mdata->tx_sgl); in mtk_spi_dma_transfer()
844 xfer->tx_dma += mdata->xfer_len; in mtk_spi_interrupt_thread()
851 xfer->tx_dma = sg_dma_address(mdata->tx_sgl); in mtk_spi_interrupt_thread()
948 writel((u32)(mdata->tx_dma & MTK_SPI_32BITS_MASK), in mtk_spi_mem_setup_dma_xfer()
952 writel((u32)(mdata->tx_dma >> 32), in mtk_spi_mem_setup_dma_xfer()
1093 mdata->tx_dma = dma_map_single(mdata->dev, tx_tmp_buf, in mtk_spi_mem_exec_op()
1095 if (dma_mapping_error(mdata->dev, mdata->tx_dma)) { in mtk_spi_mem_exec_op()
[all …]
H A Dspi-fsl-lib.h38 dma_addr_t tx_dma; member
H A Dspi-tegra210-quad.c488 tx_dma_phys = t->tx_dma; in tegra_qspi_start_tx_dma()
577 t->tx_dma = dma_map_single(tqspi->dev, (void *)tx_buf, len, DMA_TO_DEVICE); in tegra_qspi_dma_map_xfer()
578 if (dma_mapping_error(tqspi->dev, t->tx_dma)) in tegra_qspi_dma_map_xfer()
585 dma_unmap_single(tqspi->dev, t->tx_dma, len, DMA_TO_DEVICE); in tegra_qspi_dma_map_xfer()
600 dma_unmap_single(tqspi->dev, t->tx_dma, len, DMA_TO_DEVICE); in tegra_qspi_dma_unmap_xfer()
670 tx_dma_phys = t->tx_dma; in tegra_qspi_start_dma_based_transfer()
/linux/drivers/net/ethernet/amd/
H A Dau1000_eth.h39 struct tx_dma { struct
74 struct tx_dma *tx_dma_ring[NUM_TX_DMA];
H A Dau1000_eth.c640 aup->tx_dma_ring[i] = (struct tx_dma *) in au1000_setup_hw_rings()
641 (tx_base + sizeof(struct tx_dma) * i); in au1000_setup_hw_rings()
864 struct tx_dma *ptxd; in au1000_tx_ack()
960 struct tx_dma *ptxd; in au1000_tx()
/linux/drivers/net/ethernet/apple/
H A Dmace.c46 volatile struct dbdma_regs __iomem *tx_dma; member
180 mp->tx_dma = ioremap(macio_resource_start(mdev, 1), 0x1000); in mace_probe()
181 if (mp->tx_dma == NULL) { in mace_probe()
264 iounmap(mp->tx_dma); in mace_probe()
293 iounmap(mp->tx_dma); in mace_remove()
436 volatile struct dbdma_regs __iomem *td = mp->tx_dma; in mace_open()
506 volatile struct dbdma_regs __iomem *td = mp->tx_dma; in mace_close()
535 volatile struct dbdma_regs __iomem *td = mp->tx_dma; in mace_xmit_start()
665 volatile struct dbdma_regs __iomem *td = mp->tx_dma; in mace_interrupt()
811 volatile struct dbdma_regs __iomem *td = mp->tx_dma; in mace_tx_timeout()
H A Dbmac.c60 volatile struct dbdma_regs __iomem *tx_dma; member
225 volatile struct dbdma_regs __iomem *td = bp->tx_dma; in bmac_enable_and_reset_chip()
473 volatile struct dbdma_regs __iomem *td = bp->tx_dma; in bmac_suspend()
581 volatile struct dbdma_regs __iomem *td = bp->tx_dma; in bmac_init_tx_ring()
637 volatile struct dbdma_regs __iomem *td = bp->tx_dma; in bmac_transmit_packet()
770 if (cp == bus_to_virt(in_le32(&bp->tx_dma->cmdptr))) in bmac_txdma_intr()
1247 bp->tx_dma = ioremap(macio_resource_start(mdev, 1), macio_resource_len(mdev, 1)); in bmac_probe()
1248 if (!bp->tx_dma) in bmac_probe()
1306 iounmap(bp->tx_dma); in bmac_probe()
1333 volatile struct dbdma_regs __iomem *td = bp->tx_dma; in bmac_close()
[all …]
/linux/drivers/atm/
H A Deni.h82 void __iomem *tx_dma; /* TX DMA queue */ member
H A Deni.c1166 writel(eni_dev->dma[i*2],eni_dev->tx_dma+dma_wr*8); in do_tx()
1167 writel(eni_dev->dma[i*2+1],eni_dev->tx_dma+dma_wr*8+4); in do_tx()
1836 eni_dev->tx_dma = eni_dev->rx_dma+NR_DMA_RX*8; in eni_start()
1837 eni_dev->service = eni_dev->tx_dma+NR_DMA_TX*8; in eni_start()
1840 eni_dev->vci,eni_dev->rx_dma,eni_dev->tx_dma, in eni_start()
/linux/drivers/net/ethernet/sis/
H A Dsis190.c276 dma_addr_t tx_dma; member
890 SIS_W32(TxDescStartAddr, tp->tx_dma); in sis190_hw_start()
1076 &tp->tx_dma, GFP_KERNEL); in sis190_open()
1107 tp->tx_dma); in sis190_open()
1170 tp->tx_dma); in sis190_close()
/linux/drivers/i2c/busses/
H A Di2c-qcom-geni.c483 dma_addr_t tx_dma = 0; in geni_i2c_tx_one_msg() local
499 if (dma_buf && geni_se_tx_dma_prep(se, dma_buf, len, &tx_dma)) { in geni_i2c_tx_one_msg()
505 gi2c->dma_addr = tx_dma; in geni_i2c_tx_one_msg()
/linux/drivers/scsi/mvsas/
H A Dmv_64xx.c315 mw32(MVS_TX_LO, mvi->tx_dma); in mvs_64xx_init()
316 mw32(MVS_TX_HI, (mvi->tx_dma >> 16) >> 16); in mvs_64xx_init()
H A Dmv_init.c103 mvi->tx, mvi->tx_dma); in mvs_free()
229 &mvi->tx_dma, GFP_KERNEL); in mvs_alloc()
H A Dmv_sas.h350 dma_addr_t tx_dma; member
H A Dmv_94xx.c453 mw32(MVS_TX_LO, mvi->tx_dma); in mvs_94xx_init()
454 mw32(MVS_TX_HI, (mvi->tx_dma >> 16) >> 16); in mvs_94xx_init()
/linux/drivers/tty/serial/8250/
H A D8250.h18 int (*tx_dma)(struct uart_8250_port *p); member
H A D8250_port.c1418 if (up->dma && !up->dma->tx_dma(up)) in __start_tx()
3180 if (!up->dma->tx_dma) in serial8250_set_defaults()
3181 up->dma->tx_dma = serial8250_tx_dma; in serial8250_set_defaults()
/linux/drivers/net/ethernet/natsemi/
H A Dnatsemi.c548 dma_addr_t tx_dma[TX_RING_SIZE]; member
2013 dma_unmap_single(&np->pci_dev->dev, np->tx_dma[i], in drain_tx()
2100 np->tx_dma[entry] = dma_map_single(&np->pci_dev->dev, skb->data, in start_tx()
2102 if (dma_mapping_error(&np->pci_dev->dev, np->tx_dma[entry])) { in start_tx()
2109 np->tx_ring[entry].addr = cpu_to_le32(np->tx_dma[entry]); in start_tx()
2168 dma_unmap_single(&np->pci_dev->dev, np->tx_dma[entry], in netdev_tx_done()
/linux/include/linux/spi/
H A Dspi.h1123 dma_addr_t tx_dma; member
/linux/drivers/net/ethernet/cadence/
H A Dmacb_main.c2508 dma_addr_t tx_dma, rx_dma; in macb_alloc_consistent() local
2522 tx = dma_alloc_coherent(dev, size, &tx_dma, GFP_KERNEL); in macb_alloc_consistent()
2523 if (!tx || upper_32_bits(tx_dma) != upper_32_bits(tx_dma + size - 1)) in macb_alloc_consistent()
2526 size, bp->num_queues, (unsigned long)tx_dma, tx); in macb_alloc_consistent()
2537 queue->tx_ring_dma = tx_dma + macb_tx_ring_size_per_queue(bp) * q; in macb_alloc_consistent()