| /linux/drivers/iommu/ |
| H A D | iommu-pages.h | 104 int iommu_pages_start_incoherent(void *virt, struct device *dma_dev); 106 struct device *dma_dev); 112 static inline void iommu_pages_flush_incoherent(struct device *dma_dev, in iommu_pages_flush_incoherent() argument 120 struct device *dma_dev) in iommu_pages_stop_incoherent_list() argument 128 struct device *dma_dev) in iommu_pages_free_incoherent() argument 136 static inline void iommu_pages_flush_incoherent(struct device *dma_dev, in iommu_pages_flush_incoherent() argument 140 dma_sync_single_for_device(dma_dev, (uintptr_t)virt + offset, len, in iommu_pages_flush_incoherent() 144 struct device *dma_dev); 145 void iommu_pages_free_incoherent(void *virt, struct device *dma_dev);
|
| H A D | iommu-pages.c | 146 int iommu_pages_start_incoherent(void *virt, struct device *dma_dev) in iommu_pages_start_incoherent() argument 155 iommu_pages_flush_incoherent(dma_dev, virt, 0, in iommu_pages_start_incoherent() 158 dma = dma_map_single(dma_dev, virt, ioptdesc_mem_size(iopt), in iommu_pages_start_incoherent() 160 if (dma_mapping_error(dma_dev, dma)) in iommu_pages_start_incoherent() 169 dma_unmap_single(dma_dev, dma, ioptdesc_mem_size(iopt), in iommu_pages_start_incoherent() 190 struct device *dma_dev) in iommu_pages_start_incoherent_list() argument 200 folio_address(ioptdesc_folio(cur)), dma_dev); in iommu_pages_start_incoherent_list() 218 struct device *dma_dev) in iommu_pages_stop_incoherent_list() argument 227 dma_unmap_single(dma_dev, virt_to_phys(folio_address(folio)), in iommu_pages_stop_incoherent_list() 241 void iommu_pages_free_incoherent(void *virt, struct device *dma_dev) in iommu_pages_free_incoherent() argument [all …]
|
| /linux/drivers/spi/ |
| H A D | spi-pxa2xx-pci.c | 84 if (dws->dma_dev != chan->device->dev) in lpss_dma_filter() 91 static void lpss_dma_put_device(void *dma_dev) in lpss_dma_put_device() argument 93 pci_dev_put(dma_dev); in lpss_dma_put_device() 100 struct pci_dev *dma_dev; in lpss_spi_setup() local 152 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(PCI_SLOT(dev->devfn), 0)); in lpss_spi_setup() 153 ret = devm_add_action_or_reset(&dev->dev, lpss_dma_put_device, dma_dev); in lpss_spi_setup() 158 tx->dma_dev = &dma_dev->dev; in lpss_spi_setup() 163 rx->dma_dev = &dma_dev->dev; in lpss_spi_setup() 196 struct pci_dev *dma_dev; in mrfld_spi_setup() local 228 dma_dev = pci_get_slot(dev->bus, PCI_DEVFN(21, 0)); in mrfld_spi_setup() [all …]
|
| /linux/drivers/dma/amd/ptdma/ |
| H A D | ptdma-dmaengine.c | 564 struct dma_device *dma_dev = &pt->dma_dev; in pt_dmaengine_register() local 596 dma_dev->dev = pt->dev; in pt_dmaengine_register() 597 dma_dev->src_addr_widths = DMA_SLAVE_BUSWIDTH_64_BYTES; in pt_dmaengine_register() 598 dma_dev->dst_addr_widths = DMA_SLAVE_BUSWIDTH_64_BYTES; in pt_dmaengine_register() 599 dma_dev->directions = DMA_MEM_TO_MEM; in pt_dmaengine_register() 600 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in pt_dmaengine_register() 601 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in pt_dmaengine_register() 602 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in pt_dmaengine_register() 608 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in pt_dmaengine_register() 610 INIT_LIST_HEAD(&dma_dev->channels); in pt_dmaengine_register() [all …]
|
| /linux/drivers/crypto/ccp/ |
| H A D | ccp-dmaengine.c | 668 struct dma_device *dma_dev = &ccp->dma_dev; in ccp_dmaengine_register() local 714 dma_dev->dev = ccp->dev; in ccp_dmaengine_register() 715 dma_dev->src_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register() 716 dma_dev->dst_addr_widths = CCP_DMA_WIDTH(dma_get_mask(ccp->dev)); in ccp_dmaengine_register() 717 dma_dev->directions = DMA_MEM_TO_MEM; in ccp_dmaengine_register() 718 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in ccp_dmaengine_register() 719 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in ccp_dmaengine_register() 720 dma_cap_set(DMA_INTERRUPT, dma_dev->cap_mask); in ccp_dmaengine_register() 729 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in ccp_dmaengine_register() 731 INIT_LIST_HEAD(&dma_dev->channels); in ccp_dmaengine_register() [all …]
|
| /linux/drivers/dma/sf-pdma/ |
| H A D | sf-pdma.c | 93 dev_err(chan->pdma->dma_dev.dev, in sf_pdma_prep_dma_memcpy() 260 dev_err(chan->pdma->dma_dev.dev, "NULL desc.\n"); in sf_pdma_xfer_desc() 456 INIT_LIST_HEAD(&pdma->dma_dev.channels); in sf_pdma_setup_chans() 487 vchan_init(&chan->vchan, &pdma->dma_dev); in sf_pdma_setup_chans() 542 pdma->dma_dev.dev = &pdev->dev; in sf_pdma_probe() 545 dma_cap_set(DMA_MEMCPY, pdma->dma_dev.cap_mask); in sf_pdma_probe() 546 pdma->dma_dev.copy_align = 2; in sf_pdma_probe() 547 pdma->dma_dev.src_addr_widths = widths; in sf_pdma_probe() 548 pdma->dma_dev.dst_addr_widths = widths; in sf_pdma_probe() 549 pdma->dma_dev.directions = BIT(DMA_MEM_TO_MEM); in sf_pdma_probe() [all …]
|
| /linux/drivers/dma/ |
| H A D | dma-axi-dmac.c | 173 struct dma_device dma_dev; member 180 dma_dev); in chan_to_axi_dmac() 360 dev_dbg(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers() 364 dev_warn(dmac->dma_dev.dev, in axi_dmac_dequeue_partial_xfers() 537 struct device *dev = dmac->dma_dev.dev; in axi_dmac_alloc_desc() 576 struct device *dev = dmac->dma_dev.dev; in axi_dmac_free_desc() 992 struct device *dev = dmac->dma_dev.dev; in axi_dmac_detect_caps() 1016 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps() 1024 dev_err(dmac->dma_dev.dev, in axi_dmac_detect_caps() 1071 struct dma_device *dma_dev; in axi_dmac_probe() local [all …]
|
| H A D | altera-msgdma.c | 812 struct dma_device *dma_dev; in msgdma_probe() local 860 dma_dev = &mdev->dmadev; in msgdma_probe() 863 dma_cap_zero(dma_dev->cap_mask); in msgdma_probe() 864 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in msgdma_probe() 865 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in msgdma_probe() 867 dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe() 868 dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in msgdma_probe() 869 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | in msgdma_probe() 871 dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in msgdma_probe() 874 INIT_LIST_HEAD(&dma_dev->channels); in msgdma_probe() [all …]
|
| H A D | img-mdc-dma.c | 133 struct dma_device dma_dev; member 182 return mdma->dma_dev.dev; in mdma2dev() 816 list_for_each_entry(chan, &mdma->dma_dev.channels, device_node) { in mdc_of_xlate() 912 dma_cap_zero(mdma->dma_dev.cap_mask); in mdc_dma_probe() 913 dma_cap_set(DMA_SLAVE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 914 dma_cap_set(DMA_PRIVATE, mdma->dma_dev.cap_mask); in mdc_dma_probe() 915 dma_cap_set(DMA_CYCLIC, mdma->dma_dev.cap_mask); in mdc_dma_probe() 916 dma_cap_set(DMA_MEMCPY, mdma->dma_dev.cap_mask); in mdc_dma_probe() 946 mdma->dma_dev.dev = &pdev->dev; in mdc_dma_probe() 947 mdma->dma_dev.device_prep_slave_sg = mdc_prep_slave_sg; in mdc_dma_probe() [all …]
|
| H A D | nbpfaxi.c | 229 struct dma_device dma_dev; member 325 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_read() 334 dev_dbg(nbpf->dma_dev.dev, "%s(0x%p + 0x%x) = 0x%x\n", in nbpf_write() 400 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): next 0x%x, cur 0x%x\n", __func__, in nbpf_start() 869 dev_dbg(chan->nbpf->dma_dev.dev, "%s(): force-free desc %p cookie %d\n", in nbpf_chan_idle() 1110 dchan = dma_get_any_slave_channel(&nbpf->dma_dev); in nbpf_of_xlate() 1241 dev_warn(nbpf->dma_dev.dev, "DMA error IRQ %u\n", irq); in nbpf_err_irq() 1259 struct dma_device *dma_dev = &nbpf->dma_dev; in nbpf_chan_probe() local 1267 chan->dma_chan.device = dma_dev; in nbpf_chan_probe() 1271 dev_dbg(dma_dev->dev, "%s(): channel %d: -> %p\n", __func__, n, chan->base); in nbpf_chan_probe() [all …]
|
| H A D | ep93xx_dma.c | 218 * @dma_dev: holds the dmaengine device 234 struct dma_device dma_dev; member 1373 struct dma_device *dma_dev; in ep93xx_dma_of_probe() local 1388 dma_dev = &edma->dma_dev; in ep93xx_dma_of_probe() 1390 INIT_LIST_HEAD(&dma_dev->channels); in ep93xx_dma_of_probe() 1395 edmac->chan.device = dma_dev; in ep93xx_dma_of_probe() 1427 &dma_dev->channels); in ep93xx_dma_of_probe() 1449 dma_cap_mask_t mask = edma->dma_dev.cap_mask; in ep93xx_m2p_dma_of_xlate() 1480 dma_cap_mask_t mask = edma->dma_dev in ep93xx_m2m_dma_of_xlate() 1505 struct dma_device *dma_dev; ep93xx_dma_probe() local [all...] |
| H A D | fsl_raid.c | 629 struct dma_device *dma_dev; in fsl_re_chan_probe() local 637 dma_dev = &re_priv->dma_dev; in fsl_re_chan_probe() 683 chan->chan.device = dma_dev; in fsl_re_chan_probe() 752 struct dma_device *dma_dev; in fsl_re_probe() local 781 dma_dev = &re_priv->dma_dev; in fsl_re_probe() 782 dma_dev->dev = dev; in fsl_re_probe() 783 INIT_LIST_HEAD(&dma_dev->channels); in fsl_re_probe() 786 dma_dev in fsl_re_probe() [all...] |
| H A D | sprd-dma.c | 209 struct dma_device dma_dev; member 358 dev_warn(sdev->dma_dev.dev, in sprd_dma_pause_resume() 422 dev_warn(sdev->dma_dev.dev, "incorrect dma interrupt type\n"); in sprd_dma_get_int_type() 481 dev_err(sdev->dma_dev.dev, "invalid channel mode setting %d\n", in sprd_dma_set_2stage_config() 762 dev_err(sdev->dma_dev.dev, "invalid source step\n"); in sprd_dma_fill_desc() 778 dev_err(sdev->dma_dev.dev, "invalid destination step\n"); in sprd_dma_fill_desc() 786 dev_err(sdev->dma_dev.dev, "invalid source datawidth\n"); in sprd_dma_fill_desc() 792 dev_err(sdev->dma_dev.dev, "invalid destination datawidth\n"); in sprd_dma_fill_desc() 1177 dma_cap_set(DMA_MEMCPY, sdev->dma_dev.cap_mask); in sprd_dma_probe() 1179 INIT_LIST_HEAD(&sdev->dma_dev.channels); in sprd_dma_probe() [all …]
|
| H A D | hisi_dma.c | 161 struct dma_device dma_dev; member 712 vchan_init(&hdma_dev->chan[i].vc, &hdma_dev->dma_dev); in hisi_dma_enable_qps() 824 struct dma_device *dma_dev; in hisi_dma_init_dma_dev() local 826 dma_dev = &hdma_dev->dma_dev; in hisi_dma_init_dma_dev() 827 dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); in hisi_dma_init_dma_dev() 828 dma_dev->device_free_chan_resources = hisi_dma_free_chan_resources; in hisi_dma_init_dma_dev() 829 dma_dev->device_prep_dma_memcpy = hisi_dma_prep_dma_memcpy; in hisi_dma_init_dma_dev() 830 dma_dev->device_tx_status = hisi_dma_tx_status; in hisi_dma_init_dma_dev() 831 dma_dev->device_issue_pending = hisi_dma_issue_pending; in hisi_dma_init_dma_dev() 832 dma_dev->device_terminate_all = hisi_dma_terminate_all; in hisi_dma_init_dma_dev() [all …]
|
| H A D | mv_xor.c | 1040 struct dma_device *dma_dev; in mv_xor_channel_add() local 1053 dma_dev = &mv_chan->dmadev; in mv_xor_channel_add() 1054 dma_dev->dev = &pdev->dev; in mv_xor_channel_add() 1062 mv_chan->dummy_src_addr = dma_map_single(dma_dev->dev, in mv_xor_channel_add() 1064 if (dma_mapping_error(dma_dev->dev, mv_chan->dummy_src_addr)) in mv_xor_channel_add() 1067 mv_chan->dummy_dst_addr = dma_map_single(dma_dev->dev, in mv_xor_channel_add() 1069 if (dma_mapping_error(dma_dev->dev, mv_chan->dummy_dst_addr)) { in mv_xor_channel_add() 1088 dma_dev->cap_mask = cap_mask; in mv_xor_channel_add() 1090 INIT_LIST_HEAD(&dma_dev->channels); in mv_xor_channel_add() 1093 dma_dev in mv_xor_channel_add() [all...] |
| H A D | tegra210-adma.c | 187 struct dma_device dma_dev; member 829 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate() 1111 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_adma_probe() 1136 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_adma_probe() 1151 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_adma_probe() 1152 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_adma_probe() 1153 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_adma_probe() 1155 tdma->dma_dev.dev = &pdev->dev; in tegra_adma_probe() 1156 tdma->dma_dev.device_alloc_chan_resources = in tegra_adma_probe() 1158 tdma->dma_dev.device_free_chan_resources = in tegra_adma_probe() [all …]
|
| H A D | tegra186-gpc-dma.c | 256 struct dma_device dma_dev; member 1297 chan = dma_get_any_slave_channel(&tdma->dma_dev); in tegra_dma_of_xlate() 1391 tdma->dma_dev.dev = &pdev->dev; in tegra_dma_probe() 1407 INIT_LIST_HEAD(&tdma->dma_dev.channels); in tegra_dma_probe() 1426 vchan_init(&tdc->vc, &tdma->dma_dev); in tegra_dma_probe() 1434 dma_cap_set(DMA_SLAVE, tdma->dma_dev.cap_mask); in tegra_dma_probe() 1435 dma_cap_set(DMA_PRIVATE, tdma->dma_dev.cap_mask); in tegra_dma_probe() 1436 dma_cap_set(DMA_MEMCPY, tdma->dma_dev.cap_mask); in tegra_dma_probe() 1437 dma_cap_set(DMA_MEMSET, tdma->dma_dev.cap_mask); in tegra_dma_probe() 1438 dma_cap_set(DMA_CYCLIC, tdma->dma_dev.cap_mask); in tegra_dma_probe() [all …]
|
| H A D | dmaengine.c | 71 static void dmaengine_debug_register(struct dma_device *dma_dev) in dmaengine_debug_register() argument 73 dma_dev->dbg_dev_root = debugfs_create_dir(dev_name(dma_dev->dev), in dmaengine_debug_register() 75 if (IS_ERR(dma_dev->dbg_dev_root)) in dmaengine_debug_register() 76 dma_dev->dbg_dev_root = NULL; in dmaengine_debug_register() 79 static void dmaengine_debug_unregister(struct dma_device *dma_dev) in dmaengine_debug_unregister() argument 81 debugfs_remove_recursive(dma_dev->dbg_dev_root); in dmaengine_debug_unregister() 82 dma_dev->dbg_dev_root = NULL; in dmaengine_debug_unregister() 86 struct dma_device *dma_dev) in dmaengine_dbg_summary_show() argument 90 list_for_each_entry(chan, &dma_dev in dmaengine_dbg_summary_show() 106 struct dma_device *dma_dev = NULL; dmaengine_summary_show() local 138 dmaengine_debug_register(struct dma_device * dma_dev) dmaengine_debug_register() argument 143 dmaengine_debug_unregister(struct dma_device * dma_dev) dmaengine_debug_unregister() argument [all...] |
| /linux/drivers/tty/serial/8250/ |
| H A D | 8250_mid.c | 44 struct pci_dev *dma_dev; member 70 mid->dma_dev = pci_get_slot(pdev->bus, in pnw_setup() 77 pci_dev_put(mid->dma_dev); in pnw_exit() 89 chip = pci_get_drvdata(mid->dma_dev); in tng_handle_irq() 125 mid->dma_dev = pci_get_slot(pdev->bus, PCI_DEVFN(5, 0)); in tng_setup() 133 pci_dev_put(mid->dma_dev); in tng_exit() 192 mid->dma_dev = pdev; in dnv_setup() 200 if (!mid->dma_dev) in dnv_exit() 246 if (s->dma_dev != chan->device->dev || s->chan_id != chan->chan_id) in mid8250_dma_filter() 260 if (!mid->dma_dev) in mid8250_dma_setup() [all …]
|
| /linux/drivers/gpu/drm/exynos/ |
| H A D | exynos_drm_dma.c | 48 if (get_dma_ops(priv->dma_dev) != get_dma_ops(subdrv_dev)) { in drm_iommu_attach_device() 100 if (!priv->dma_dev) { in exynos_drm_register_dma() 101 priv->dma_dev = dev; in exynos_drm_register_dma() 116 mapping = iommu_get_domain_for_dev(priv->dma_dev); in exynos_drm_register_dma() 142 priv->dma_dev = NULL; in exynos_drm_cleanup_dma()
|
| /linux/drivers/dma/xilinx/ |
| H A D | xdma.c | 128 struct dma_device dma_dev; member 477 vchan_init(&xchan->vchan, &xdev->dma_dev); in xdma_alloc_channels() 812 struct device *dev = xdev->dma_dev.dev; in xdma_alloc_chan_resources() 1187 dma_async_device_unregister(&xdev->dma_dev); in xdma_remove() 1241 INIT_LIST_HEAD(&xdev->dma_dev.channels); in xdma_probe() 1255 dma_cap_set(DMA_SLAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1256 dma_cap_set(DMA_PRIVATE, xdev->dma_dev.cap_mask); in xdma_probe() 1257 dma_cap_set(DMA_CYCLIC, xdev->dma_dev.cap_mask); in xdma_probe() 1258 dma_cap_set(DMA_INTERLEAVE, xdev->dma_dev.cap_mask); in xdma_probe() 1259 dma_cap_set(DMA_REPEAT, xdev->dma_dev.cap_mask); in xdma_probe() [all …]
|
| /linux/drivers/dma/sh/ |
| H A D | shdmac.c | 164 dev_warn(shdev->shdma_dev.dma_dev.dev, "Can't initialize DMAOR.\n"); in sh_dmae_rst() 168 dev_warn(shdev->shdma_dev.dma_dev.dev, in sh_dmae_rst() 530 struct platform_device *pdev = to_platform_device(sdev->dma_dev.dev); in sh_dmae_chan_probe() 535 sh_chan = devm_kzalloc(sdev->dma_dev.dev, sizeof(struct sh_dmae_chan), in sh_dmae_chan_probe() 557 dev_err(sdev->dma_dev.dev, in sh_dmae_chan_probe() 688 struct dma_device *dma_dev; in sh_dmae_probe() local 727 dma_dev = &shdev->shdma_dev.dma_dev; in sh_dmae_probe() 738 dma_dev->src_addr_widths = widths; in sh_dmae_probe() 739 dma_dev->dst_addr_widths = widths; in sh_dmae_probe() 740 dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM); in sh_dmae_probe() [all …]
|
| H A D | shdma-base.c | 37 #define to_shdma_dev(d) container_of(d, struct shdma_dev, dma_dev) 319 dev_warn(sdev->dma_dev.dev, "invalid slave ID passed to dma_request_slave\n"); in shdma_chan_filter() 965 schan->dma_chan.device = &sdev->dma_dev; in shdma_chan_probe() 968 schan->dev = sdev->dma_dev.dev; in shdma_chan_probe() 982 &sdev->dma_dev.channels); in shdma_chan_probe() 996 struct dma_device *dma_dev = &sdev->dma_dev; in shdma_init() local 1019 INIT_LIST_HEAD(&dma_dev->channels); in shdma_init() 1022 dma_dev->device_alloc_chan_resources in shdma_init() 1024 dma_dev in shdma_init() [all...] |
| /linux/drivers/dma/stm32/ |
| H A D | stm32-dma3.c | 306 struct dma_device dma_dev; member 318 return container_of(chan->vchan.chan.device, struct stm32_dma3_ddata, dma_dev); in to_stm32_dma3_ddata() 338 return ddata->dma_dev.dev; in ddata2dev() 573 struct dma_device dma_device = ddata->dma_dev; in stm32_dma3_chan_prep_hw() 1638 dma_cap_mask_t mask = ddata->dma_dev.cap_mask; in stm32_dma3_of_xlate() 1736 struct dma_device *dma_dev; in stm32_dma3_probe() local 1746 dma_dev = &ddata->dma_dev; in stm32_dma3_probe() 1766 INIT_LIST_HEAD(&dma_dev->channels); in stm32_dma3_probe() 1768 dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); in stm32_dma3_probe() 1769 dma_cap_set(DMA_PRIVATE, dma_dev->cap_mask); in stm32_dma3_probe() [all …]
|
| /linux/drivers/pci/endpoint/functions/ |
| H A D | pci-epf-mhi.c | 313 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; in pci_epf_mhi_edma_read() local 337 dst_addr = dma_map_single(dma_dev, buf_info->dev_addr, buf_info->size, in pci_epf_mhi_edma_read() 339 ret = dma_mapping_error(dma_dev, dst_addr); in pci_epf_mhi_edma_read() 373 dma_unmap_single(dma_dev, dst_addr, buf_info->size, DMA_FROM_DEVICE); in pci_epf_mhi_edma_read() 384 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; in pci_epf_mhi_edma_write() local 408 src_addr = dma_map_single(dma_dev, buf_info->dev_addr, buf_info->size, in pci_epf_mhi_edma_write() 410 ret = dma_mapping_error(dma_dev, src_addr); in pci_epf_mhi_edma_write() 444 dma_unmap_single(dma_dev, src_addr, buf_info->size, DMA_TO_DEVICE); in pci_epf_mhi_edma_write() 454 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; in pci_epf_mhi_dma_worker() local 466 dma_unmap_single(dma_dev, it in pci_epf_mhi_dma_worker() 489 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; pci_epf_mhi_edma_read_async() local 568 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; pci_epf_mhi_edma_write_async() local 662 struct device *dma_dev = epf_mhi->epf->epc->dev.parent; pci_epf_mhi_dma_init() local [all...] |