Home
last modified time | relevance | path

Searched refs:tx_sgl (Results 1 – 8 of 8) sorted by relevance

/linux/drivers/spi/
H A Dspi-mt65xx.c168 struct scatterlist *tx_sgl, *rx_sgl; member
643 if (mdata->tx_sgl) { in mtk_spi_setup_dma_addr()
701 mdata->tx_sgl = NULL; in mtk_spi_dma_transfer()
718 mdata->tx_sgl = xfer->tx_sg.sgl; in mtk_spi_dma_transfer()
722 if (mdata->tx_sgl) { in mtk_spi_dma_transfer()
723 xfer->tx_dma = sg_dma_address(mdata->tx_sgl); in mtk_spi_dma_transfer()
724 mdata->tx_sgl_len = sg_dma_len(mdata->tx_sgl); in mtk_spi_dma_transfer()
843 if (mdata->tx_sgl) in mtk_spi_interrupt_thread()
848 if (mdata->tx_sgl && (mdata->tx_sgl_len == 0)) { in mtk_spi_interrupt_thread()
849 mdata->tx_sgl = sg_next(mdata->tx_sgl); in mtk_spi_interrupt_thread()
[all …]
H A Dspi-pci1xxxx.c149 struct scatterlist *tx_sgl, *rx_sgl; member
549 p->tx_sgl = xfer->tx_sg.sgl; in pci1xxxx_spi_transfer_with_dma()
556 if (!xfer->tx_buf || !p->tx_sgl) { in pci1xxxx_spi_transfer_with_dma()
568 tx_dma_addr = sg_dma_address(p->tx_sgl); in pci1xxxx_spi_transfer_with_dma()
569 p->tx_sgl_len = sg_dma_len(p->tx_sgl); in pci1xxxx_spi_transfer_with_dma()
666 p->tx_sgl = sg_next(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer()
667 if (p->tx_sgl) { in pci1xxxx_spi_setup_next_dma_to_io_transfer()
668 tx_dma_addr = sg_dma_address(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer()
670 p->tx_sgl_len = sg_dma_len(p->tx_sgl); in pci1xxxx_spi_setup_next_dma_to_io_transfer()
/linux/drivers/tty/serial/
H A Dimx.c221 struct scatterlist rx_sgl, tx_sgl[2]; member
588 struct scatterlist *sgl = &sport->tx_sgl[0]; in imx_uart_dma_tx_callback()
625 struct scatterlist *sgl = sport->tx_sgl; in imx_uart_dma_tx()
639 sg_init_table(sgl, ARRAY_SIZE(sport->tx_sgl)); in imx_uart_dma_tx()
642 ARRAY_SIZE(sport->tx_sgl), sport->tx_bytes); in imx_uart_dma_tx()
1599 dma_unmap_sg(sport->port.dev, &sport->tx_sgl[0], in imx_uart_shutdown()
1717 struct scatterlist *sgl = &sport->tx_sgl[0]; in imx_uart_flush_buffer()
H A Dmxs-auart.c431 struct scatterlist tx_sgl; member
522 dma_unmap_sg(s->dev, &s->tx_sgl, 1, DMA_TO_DEVICE); in dma_tx_callback()
538 struct scatterlist *sgl = &s->tx_sgl; in mxs_auart_dma_tx()
H A Dfsl_lpuart.c285 struct scatterlist rx_sgl, tx_sgl[2]; member
481 struct scatterlist *sgl = sport->tx_sgl; in lpuart_dma_tx()
489 sg_init_table(sgl, ARRAY_SIZE(sport->tx_sgl)); in lpuart_dma_tx()
492 ARRAY_SIZE(sport->tx_sgl), sport->dma_tx_bytes); in lpuart_dma_tx()
527 struct scatterlist *sgl = &sport->tx_sgl[0]; in lpuart_dma_tx_complete()
609 dma_unmap_sg(chan->device->dev, &sport->tx_sgl[0], in lpuart_flush_buffer()
/linux/drivers/net/ethernet/hisilicon/hns3/
H A Dhns3_enet.h434 u64 tx_sgl; member
H A Dhns3_enet.c60 static unsigned int tx_sgl = 1; variable
61 module_param(tx_sgl, uint, 0600);
62 MODULE_PARM_DESC(tx_sgl, "Minimum number of frags when using dma_map_sg() to optimize the IOMMU map…
1028 if (skb->len <= ring->tx_copybreak || !tx_sgl || in hns3_can_use_tx_sgl()
1030 skb_shinfo(skb)->nr_frags < tx_sgl)) in hns3_can_use_tx_sgl()
2263 hns3_ring_stats_update(ring, tx_sgl); in hns3_handle_tx_sgl()
H A Dhns3_ethtool.c41 HNS3_TQP_STAT("sgl", tx_sgl),