Home
last modified time | relevance | path

Searched refs:dma_device (Results 1 – 25 of 130) sorted by relevance

123456

/linux/drivers/dma/
H A Dmxs-dma.c142 struct dma_device dma_device; member
367 dev_dbg(mxs_dma->dma_device.dev, in mxs_dma_int_handler()
401 mxs_chan->ccw = dma_alloc_coherent(mxs_dma->dma_device.dev, in mxs_dma_alloc_chan_resources()
431 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_alloc_chan_resources()
446 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_free_chan_resources()
491 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_slave_sg()
536 dev_err(mxs_dma->dma_device.dev, "maximum bytes for sg entry exceeded: %d > %d\n", in mxs_dma_prep_slave_sg()
591 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
598 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
728 dma_cap_mask_t mask = mxs_dma->dma_device.cap_mask; in mxs_dma_xlate()
[all …]
H A Ddmaengine.c71 static void dmaengine_debug_register(struct dma_device *dma_dev) in dmaengine_debug_register()
79 static void dmaengine_debug_unregister(struct dma_device *dma_dev) in dmaengine_debug_unregister()
86 struct dma_device *dma_dev) in dmaengine_dbg_summary_show()
106 struct dma_device *dma_dev = NULL; in dmaengine_summary_show()
138 static inline int dmaengine_debug_register(struct dma_device *dma_dev) in dmaengine_debug_register()
143 static inline void dmaengine_debug_unregister(struct dma_device *dma_dev) { } in dmaengine_debug_unregister()
324 struct dma_device *device; in min_chan()
366 struct dma_device *device; in dma_channel_rebalance()
394 static int dma_device_satisfies_mask(struct dma_device *device, in dma_device_satisfies_mask()
427 struct dma_device *device = container_of(ref, struct dma_device, ref); in dma_device_release()
[all …]
H A Dst_fdma.c178 if (fdev->dma_device.dev->of_node != dma_spec->np) in st_fdma_of_xlate()
187 chan = dma_get_any_slave_channel(&fdev->dma_device); in st_fdma_of_xlate()
787 INIT_LIST_HEAD(&fdev->dma_device.channels); in st_fdma_probe()
793 vchan_init(&fchan->vchan, &fdev->dma_device); in st_fdma_probe()
799 dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask); in st_fdma_probe()
800 dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask); in st_fdma_probe()
801 dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask); in st_fdma_probe()
803 fdev->dma_device.dev = &pdev->dev; in st_fdma_probe()
804 fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res; in st_fdma_probe()
805 fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res; in st_fdma_probe()
[all …]
H A Dat_hdmac.c348 struct dma_device dma_device; member
376 static inline struct at_dma *to_at_dma(struct dma_device *ddev) in to_at_dma()
378 return container_of(ddev, struct at_dma, dma_device); in to_at_dma()
842 dev_vdbg(atdma->dma_device.dev, in at_dma_interrupt()
846 for (i = 0; i < atdma->dma_device.chancnt; i++) { in at_dma_interrupt()
1748 BUG_ON(!atslave->dma_dev || atslave->dma_dev != atdma->dma_device.dev); in atc_alloc_chan_resources()
1980 atdma->dma_device.cap_mask = plat_dat->cap_mask; in at_dma_probe()
2024 INIT_LIST_HEAD(&atdma->dma_device.channels); in at_dma_probe()
2036 vchan_init(&atchan->vc, &atdma->dma_device); in at_dma_probe()
2041 atdma->dma_device.device_alloc_chan_resources = atc_alloc_chan_resources; in at_dma_probe()
[all …]
H A Dimx-dma.c176 struct dma_device dma_device; member
1032 return dma_request_channel(imxdma->dma_device.cap_mask, in imxdma_xlate()
1108 INIT_LIST_HEAD(&imxdma->dma_device.channels); in imxdma_probe()
1110 dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); in imxdma_probe()
1111 dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); in imxdma_probe()
1112 dma_cap_set(DMA_MEMCPY, imxdma->dma_device.cap_mask); in imxdma_probe()
1113 dma_cap_set(DMA_INTERLEAVE, imxdma->dma_device.cap_mask); in imxdma_probe()
1146 imxdmac->chan.device = &imxdma->dma_device; in imxdma_probe()
1152 &imxdma->dma_device.channels); in imxdma_probe()
1155 imxdma->dma_device.dev = &pdev->dev; in imxdma_probe()
[all …]
H A Dimx-sdma.c530 struct dma_device dma_device; member
2213 dma_cap_mask_t mask = sdma->dma_device.cap_mask; in sdma_xlate()
2302 dma_cap_set(DMA_SLAVE, sdma->dma_device.cap_mask); in sdma_probe()
2303 dma_cap_set(DMA_CYCLIC, sdma->dma_device.cap_mask); in sdma_probe()
2304 dma_cap_set(DMA_MEMCPY, sdma->dma_device.cap_mask); in sdma_probe()
2305 dma_cap_set(DMA_PRIVATE, sdma->dma_device.cap_mask); in sdma_probe()
2307 INIT_LIST_HEAD(&sdma->dma_device.channels); in sdma_probe()
2325 vchan_init(&sdmac->vc, &sdma->dma_device); in sdma_probe()
2345 sdma->dma_device.dev = &pdev->dev; in sdma_probe()
2347 sdma->dma_device.device_alloc_chan_resources = sdma_alloc_chan_resources; in sdma_probe()
[all …]
H A Ddmaengine.h183 struct dma_chan *dma_get_any_slave_channel(struct dma_device *device);
189 dmaengine_get_debugfs_root(struct dma_device *dma_dev) { in dmaengine_get_debugfs_root()
195 dmaengine_get_debugfs_root(struct dma_device *dma_dev) in dmaengine_get_debugfs_root()
H A Ddma-jz4780.c149 struct dma_device dma_device; member
180 dma_device); in jz4780_dma_chan_parent()
809 dma_cap_mask_t mask = jzdma->dma_device.cap_mask; in jz4780_of_dma_xlate()
826 dev_err(jzdma->dma_device.dev, in jz4780_of_dma_xlate()
834 dev_err(jzdma->dma_device.dev, in jz4780_of_dma_xlate()
857 struct dma_device *dd; in jz4780_dma_probe()
912 dd = &jzdma->dma_device; in jz4780_dma_probe()
H A Didma64.h184 struct dma_device dma;
193 static inline struct idma64 *to_idma64(struct dma_device *ddev) in to_idma64()
/linux/include/linux/
H A Ddmaengine.h339 struct dma_device *device;
868 struct dma_device { struct
956 void (*device_release)(struct dma_device *dev); argument
958 void (*dbg_summary_show)(struct seq_file *s, struct dma_device *dev); argument
1258 static inline bool is_dma_copy_aligned(struct dma_device *dev, size_t off1, in is_dma_copy_aligned()
1264 static inline bool is_dma_xor_aligned(struct dma_device *dev, size_t off1, in is_dma_xor_aligned()
1270 static inline bool is_dma_pq_aligned(struct dma_device *dev, size_t off1, in is_dma_pq_aligned()
1276 static inline bool is_dma_fill_aligned(struct dma_device *dev, size_t off1, in is_dma_fill_aligned()
1283 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) in dma_set_maxpq()
1302 static inline bool dma_dev_has_pq_continue(struct dma_device *dma) in dma_dev_has_pq_continue()
[all …]
/linux/drivers/infiniband/core/
H A Dumem_dmabuf.c118 struct device *dma_device, in ib_umem_dmabuf_get_with_dma_device() argument
157 dma_device, in ib_umem_dmabuf_get_with_dma_device()
179 return ib_umem_dmabuf_get_with_dma_device(device, device->dma_device, in ib_umem_dmabuf_get()
200 struct device *dma_device, in ib_umem_dmabuf_get_pinned_with_dma_device() argument
207 umem_dmabuf = ib_umem_dmabuf_get_with_dma_device(device, dma_device, offset, in ib_umem_dmabuf_get_pinned_with_dma_device()
238 return ib_umem_dmabuf_get_pinned_with_dma_device(device, device->dma_device, in ib_umem_dmabuf_get_pinned()
H A Dumem_odp.c102 ret = hmm_dma_map_alloc(dev->dma_device, map, in ib_init_umem_odp()
120 hmm_dma_map_free(dev->dma_device, map); in ib_init_umem_odp()
292 hmm_dma_map_free(dev->dma_device, &umem_odp->map); in ib_umem_odp_free()
447 if (!hmm_dma_unmap_pfn(dev->dma_device, &umem_odp->map, idx)) in ib_umem_odp_unmap_dma_pages()
/linux/drivers/dma/ioat/
H A Dsysfs.c19 struct dma_device *dma = c->device; in cap_show()
33 struct dma_device *dma = c->device; in version_show()
77 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_kobject_add()
98 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_kobject_del()
H A Dinit.c305 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_dma_self_test()
496 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_probe()
540 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_dma_remove()
557 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_enumerate_channels()
758 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_init_channel()
790 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_xor_val_self_test()
1052 struct dma_device *dma; in ioat_intr_quirk()
1080 struct dma_device *dma; in ioat3_dma_probe()
1320 static void release_ioatdma(struct dma_device *device) in release_ioatdma()
/linux/drivers/dma/hsu/
H A Dhsu.h115 struct dma_device dma;
122 static inline struct hsu_dma *to_hsu_dma(struct dma_device *ddev) in to_hsu_dma()
/linux/drivers/dma/qcom/
H A Dgpi.c470 struct dma_device dma_device; member
2209 INIT_LIST_HEAD(&gpi_dev->dma_device.channels); in gpi_probe()
2242 vchan_init(&gchan->vc, &gpi_dev->dma_device); in gpi_probe()
2261 dma_cap_zero(gpi_dev->dma_device.cap_mask); in gpi_probe()
2262 dma_cap_set(DMA_SLAVE, gpi_dev->dma_device.cap_mask); in gpi_probe()
2265 gpi_dev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); in gpi_probe()
2266 gpi_dev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in gpi_probe()
2267 gpi_dev->dma_device.src_addr_widths = DMA_SLAVE_BUSWIDTH_8_BYTES; in gpi_probe()
2268 gpi_dev->dma_device.dst_addr_widths = DMA_SLAVE_BUSWIDTH_8_BYTES; in gpi_probe()
2269 gpi_dev->dma_device.device_alloc_chan_resources = gpi_alloc_chan_resources; in gpi_probe()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c51 struct device *dma_device; in mlx5_fpga_conn_map_buf() local
57 dma_device = mlx5_core_dma_dev(conn->fdev->mdev); in mlx5_fpga_conn_map_buf()
58 buf->sg[0].dma_addr = dma_map_single(dma_device, buf->sg[0].data, in mlx5_fpga_conn_map_buf()
60 err = dma_mapping_error(dma_device, buf->sg[0].dma_addr); in mlx5_fpga_conn_map_buf()
70 buf->sg[1].dma_addr = dma_map_single(dma_device, buf->sg[1].data, in mlx5_fpga_conn_map_buf()
72 err = dma_mapping_error(dma_device, buf->sg[1].dma_addr); in mlx5_fpga_conn_map_buf()
75 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_map_buf()
87 struct device *dma_device; in mlx5_fpga_conn_unmap_buf() local
89 dma_device = mlx5_core_dma_dev(conn->fdev->mdev); in mlx5_fpga_conn_unmap_buf()
91 dma_unmap_single(dma_device, buf->sg[1].dma_addr, in mlx5_fpga_conn_unmap_buf()
[all …]
/linux/crypto/async_tx/
H A Dasync_pq.c42 struct dma_device *dma = chan->device; in do_async_gen_syndrome()
143 is_dma_pq_aligned_offs(struct dma_device *dev, unsigned int *offs, in is_dma_pq_aligned_offs()
184 struct dma_device *device = chan ? chan->device : NULL; in async_gen_syndrome()
303 struct dma_device *device = chan ? chan->device : NULL; in async_syndrome_val()
H A Dasync_tx.c69 struct dma_device *device = chan->device; in async_tx_channel_switch()
224 struct dma_device *device; in async_trigger_callback()
H A Dasync_memcpy.c38 struct dma_device *device = chan ? chan->device : NULL; in async_memcpy()
/linux/drivers/dma/idxd/
H A Ddma.c190 static void idxd_dma_release(struct dma_device *device) in idxd_dma_release()
200 struct dma_device *dma; in idxd_register_dma_device()
251 struct dma_device *dma = &idxd->idxd_dma->dma; in idxd_register_dma_channel()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_db.c71 struct device *dma_device) in hns_roce_alloc_db_pgdir() argument
83 pgdir->page = dma_alloc_coherent(dma_device, PAGE_SIZE, in hns_roce_alloc_db_pgdir()
/linux/include/rdma/
H A Dib_verbs.h2813 /* Do not access @dma_device directly from ULP nor from HW drivers. */
2814 struct device *dma_device;
2997 struct device *dma_device);
4157 * Drivers that don't need a DMA mapping at the RDMA layer, set dma_device to in ib_virt_dma_to_ptr()
4163 return IS_ENABLED(CONFIG_INFINIBAND_VIRT_DMA) && !dev->dma_device;
4174 return dma_pci_p2pdma_supported(dev->dma_device);
4211 return dma_mapping_error(dev->dma_device, dma_addr); in ib_dma_unmap_single()
4227 return dma_map_single(dev->dma_device, cpu_addr, size, direction); in ib_dma_map_page()
4242 dma_unmap_single(dev->dma_device, addr, size, direction);
4261 return dma_map_page(dev->dma_device, pag in ib_dma_map_sg_attrs()
2787 struct device *dma_device; global() member
[all...]
/linux/include/linux/platform_data/
H A Ddma-iop32x.h37 struct dma_device common;
/linux/drivers/dma/sf-pdma/
H A Dsf-pdma.h113 struct dma_device dma_dev;

123456