Home
last modified time | relevance | path

Searched refs:dma_device (Results 1 – 25 of 141) sorted by relevance

123456

/linux/drivers/dma/
H A Dmxs-dma.c142 struct dma_device dma_device; member
367 dev_dbg(mxs_dma->dma_device.dev, in mxs_dma_int_handler()
401 mxs_chan->ccw = dma_alloc_coherent(mxs_dma->dma_device.dev, in mxs_dma_alloc_chan_resources()
431 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_alloc_chan_resources()
446 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_free_chan_resources()
491 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_slave_sg()
536 dev_err(mxs_dma->dma_device.dev, "maximum bytes for sg entry exceeded: %d > %d\n", in mxs_dma_prep_slave_sg()
591 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
598 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
728 dma_cap_mask_t mask = mxs_dma->dma_device.cap_mask; in mxs_dma_xlate()
[all …]
H A Ddmaengine.c69 static void dmaengine_debug_register(struct dma_device *dma_dev) in dmaengine_debug_register()
77 static void dmaengine_debug_unregister(struct dma_device *dma_dev) in dmaengine_debug_unregister()
84 struct dma_device *dma_dev) in dmaengine_dbg_summary_show()
104 struct dma_device *dma_dev = NULL; in dmaengine_summary_show()
136 static inline int dmaengine_debug_register(struct dma_device *dma_dev) in dmaengine_debug_register()
141 static inline void dmaengine_debug_unregister(struct dma_device *dma_dev) { } in dmaengine_debug_unregister()
322 struct dma_device *device; in min_chan()
364 struct dma_device *device; in dma_channel_rebalance()
392 static int dma_device_satisfies_mask(struct dma_device *device, in dma_device_satisfies_mask()
425 struct dma_device *device = container_of(ref, struct dma_device, ref); in dma_device_release()
[all …]
H A Dst_fdma.c178 if (fdev->dma_device.dev->of_node != dma_spec->np) in st_fdma_of_xlate()
187 chan = dma_get_any_slave_channel(&fdev->dma_device); in st_fdma_of_xlate()
787 INIT_LIST_HEAD(&fdev->dma_device.channels); in st_fdma_probe()
793 vchan_init(&fchan->vchan, &fdev->dma_device); in st_fdma_probe()
799 dma_cap_set(DMA_SLAVE, fdev->dma_device.cap_mask); in st_fdma_probe()
800 dma_cap_set(DMA_CYCLIC, fdev->dma_device.cap_mask); in st_fdma_probe()
801 dma_cap_set(DMA_MEMCPY, fdev->dma_device.cap_mask); in st_fdma_probe()
803 fdev->dma_device.dev = &pdev->dev; in st_fdma_probe()
804 fdev->dma_device.device_alloc_chan_resources = st_fdma_alloc_chan_res; in st_fdma_probe()
805 fdev->dma_device.device_free_chan_resources = st_fdma_free_chan_res; in st_fdma_probe()
[all …]
H A Dat_hdmac.c348 struct dma_device dma_device; member
376 static inline struct at_dma *to_at_dma(struct dma_device *ddev) in to_at_dma()
378 return container_of(ddev, struct at_dma, dma_device); in to_at_dma()
842 dev_vdbg(atdma->dma_device.dev, in at_dma_interrupt()
846 for (i = 0; i < atdma->dma_device.chancnt; i++) { in at_dma_interrupt()
1748 BUG_ON(!atslave->dma_dev || atslave->dma_dev != atdma->dma_device.dev); in atc_alloc_chan_resources()
1975 atdma->dma_device.cap_mask = plat_dat->cap_mask; in at_dma_probe()
2019 INIT_LIST_HEAD(&atdma->dma_device.channels); in at_dma_probe()
2031 vchan_init(&atchan->vc, &atdma->dma_device); in at_dma_probe()
2036 atdma->dma_device.device_alloc_chan_resources = atc_alloc_chan_resources; in at_dma_probe()
[all …]
H A Dimx-dma.c175 struct dma_device dma_device; member
1030 return dma_request_channel(imxdma->dma_device.cap_mask, in imxdma_xlate()
1106 INIT_LIST_HEAD(&imxdma->dma_device.channels); in imxdma_probe()
1108 dma_cap_set(DMA_SLAVE, imxdma->dma_device.cap_mask); in imxdma_probe()
1109 dma_cap_set(DMA_CYCLIC, imxdma->dma_device.cap_mask); in imxdma_probe()
1110 dma_cap_set(DMA_MEMCPY, imxdma->dma_device.cap_mask); in imxdma_probe()
1111 dma_cap_set(DMA_INTERLEAVE, imxdma->dma_device.cap_mask); in imxdma_probe()
1144 imxdmac->chan.device = &imxdma->dma_device; in imxdma_probe()
1150 &imxdma->dma_device.channels); in imxdma_probe()
1153 imxdma->dma_device.dev = &pdev->dev; in imxdma_probe()
[all …]
H A Dimx-sdma.c530 struct dma_device dma_device; member
2214 dma_cap_mask_t mask = sdma->dma_device.cap_mask; in sdma_xlate()
2303 dma_cap_set(DMA_SLAVE, sdma->dma_device.cap_mask); in sdma_probe()
2304 dma_cap_set(DMA_CYCLIC, sdma->dma_device.cap_mask); in sdma_probe()
2305 dma_cap_set(DMA_MEMCPY, sdma->dma_device.cap_mask); in sdma_probe()
2306 dma_cap_set(DMA_PRIVATE, sdma->dma_device.cap_mask); in sdma_probe()
2308 INIT_LIST_HEAD(&sdma->dma_device.channels); in sdma_probe()
2326 vchan_init(&sdmac->vc, &sdma->dma_device); in sdma_probe()
2346 sdma->dma_device.dev = &pdev->dev; in sdma_probe()
2348 sdma->dma_device.device_alloc_chan_resources = sdma_alloc_chan_resources; in sdma_probe()
[all …]
H A Ddmaengine.h183 struct dma_chan *dma_get_any_slave_channel(struct dma_device *device);
189 dmaengine_get_debugfs_root(struct dma_device *dma_dev) { in dmaengine_get_debugfs_root()
195 dmaengine_get_debugfs_root(struct dma_device *dma_dev) in dmaengine_get_debugfs_root()
H A Ddma-jz4780.c149 struct dma_device dma_device; member
180 dma_device); in jz4780_dma_chan_parent()
809 dma_cap_mask_t mask = jzdma->dma_device.cap_mask; in jz4780_of_dma_xlate()
826 dev_err(jzdma->dma_device.dev, in jz4780_of_dma_xlate()
834 dev_err(jzdma->dma_device.dev, in jz4780_of_dma_xlate()
857 struct dma_device *dd; in jz4780_dma_probe()
912 dd = &jzdma->dma_device; in jz4780_dma_probe()
H A Didma64.h184 struct dma_device dma;
193 static inline struct idma64 *to_idma64(struct dma_device *ddev) in to_idma64()
H A Dst_fdma.h133 struct dma_device dma_device; member
/linux/include/linux/
H A Ddmaengine.h339 struct dma_device *device;
862 struct dma_device { struct
953 void (*device_release)(struct dma_device *dev); argument
955 void (*dbg_summary_show)(struct seq_file *s, struct dma_device *dev); argument
1255 static inline bool is_dma_copy_aligned(struct dma_device *dev, size_t off1, in is_dma_copy_aligned()
1261 static inline bool is_dma_xor_aligned(struct dma_device *dev, size_t off1, in is_dma_xor_aligned()
1267 static inline bool is_dma_pq_aligned(struct dma_device *dev, size_t off1, in is_dma_pq_aligned()
1273 static inline bool is_dma_fill_aligned(struct dma_device *dev, size_t off1, in is_dma_fill_aligned()
1280 dma_set_maxpq(struct dma_device *dma, int maxpq, int has_pq_continue) in dma_set_maxpq()
1299 static inline bool dma_dev_has_pq_continue(struct dma_device *dma) in dma_dev_has_pq_continue()
[all …]
/linux/drivers/infiniband/core/
H A Dumem_dmabuf.c118 struct device *dma_device, in ib_umem_dmabuf_get_with_dma_device() argument
160 dma_device, in ib_umem_dmabuf_get_with_dma_device()
182 return ib_umem_dmabuf_get_with_dma_device(device, device->dma_device, in ib_umem_dmabuf_get()
203 struct device *dma_device, in ib_umem_dmabuf_get_pinned_with_dma_device() argument
210 umem_dmabuf = ib_umem_dmabuf_get_with_dma_device(device, dma_device, offset, in ib_umem_dmabuf_get_pinned_with_dma_device()
243 return ib_umem_dmabuf_get_pinned_with_dma_device(device, device->dma_device, in ib_umem_dmabuf_get_pinned()
/linux/drivers/dma/ioat/
H A Dsysfs.c19 struct dma_device *dma = c->device; in cap_show()
33 struct dma_device *dma = c->device; in version_show()
77 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_kobject_add()
98 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_kobject_del()
H A Dinit.c305 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_dma_self_test()
496 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_probe()
540 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_dma_remove()
557 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_enumerate_channels()
758 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_init_channel()
790 struct dma_device *dma = &ioat_dma->dma_dev; in ioat_xor_val_self_test()
1052 struct dma_device *dma; in ioat_intr_quirk()
1080 struct dma_device *dma; in ioat3_dma_probe()
1321 static void release_ioatdma(struct dma_device *device) in release_ioatdma()
/linux/drivers/dma/hsu/
H A Dhsu.h115 struct dma_device dma;
122 static inline struct hsu_dma *to_hsu_dma(struct dma_device *ddev) in to_hsu_dma()
/linux/drivers/dma/qcom/
H A Dgpi.c468 struct dma_device dma_device; member
2192 INIT_LIST_HEAD(&gpi_dev->dma_device.channels); in gpi_probe()
2225 vchan_init(&gchan->vc, &gpi_dev->dma_device); in gpi_probe()
2244 dma_cap_zero(gpi_dev->dma_device.cap_mask); in gpi_probe()
2245 dma_cap_set(DMA_SLAVE, gpi_dev->dma_device.cap_mask); in gpi_probe()
2248 gpi_dev->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); in gpi_probe()
2249 gpi_dev->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; in gpi_probe()
2250 gpi_dev->dma_device.src_addr_widths = DMA_SLAVE_BUSWIDTH_8_BYTES; in gpi_probe()
2251 gpi_dev->dma_device.dst_addr_widths = DMA_SLAVE_BUSWIDTH_8_BYTES; in gpi_probe()
2252 gpi_dev->dma_device.device_alloc_chan_resources = gpi_alloc_chan_resources; in gpi_probe()
[all …]
/linux/crypto/async_tx/
H A Dasync_xor.c26 struct dma_device *dma = chan->device; in do_async_xor()
141 dma_xor_aligned_offsets(struct dma_device *device, unsigned int offset, in dma_xor_aligned_offsets()
189 struct dma_device *device = chan ? chan->device : NULL; in async_xor_offs()
321 struct dma_device *device = chan ? chan->device : NULL; in async_xor_val_offs()
H A Dasync_tx.c69 struct dma_device *device = chan->device; in async_tx_channel_switch()
224 struct dma_device *device; in async_trigger_callback()
H A Dasync_pq.c42 struct dma_device *dma = chan->device; in do_async_gen_syndrome()
143 is_dma_pq_aligned_offs(struct dma_device *dev, unsigned int *offs, in is_dma_pq_aligned_offs()
184 struct dma_device *device = chan ? chan->device : NULL; in async_gen_syndrome()
303 struct dma_device *device = chan ? chan->device : NULL; in async_syndrome_val()
/linux/drivers/net/ethernet/mellanox/mlx5/core/fpga/
H A Dconn.c51 struct device *dma_device; in mlx5_fpga_conn_map_buf() local
57 dma_device = mlx5_core_dma_dev(conn->fdev->mdev); in mlx5_fpga_conn_map_buf()
58 buf->sg[0].dma_addr = dma_map_single(dma_device, buf->sg[0].data, in mlx5_fpga_conn_map_buf()
60 err = dma_mapping_error(dma_device, buf->sg[0].dma_addr); in mlx5_fpga_conn_map_buf()
70 buf->sg[1].dma_addr = dma_map_single(dma_device, buf->sg[1].data, in mlx5_fpga_conn_map_buf()
72 err = dma_mapping_error(dma_device, buf->sg[1].dma_addr); in mlx5_fpga_conn_map_buf()
75 dma_unmap_single(dma_device, buf->sg[0].dma_addr, in mlx5_fpga_conn_map_buf()
87 struct device *dma_device; in mlx5_fpga_conn_unmap_buf() local
89 dma_device = mlx5_core_dma_dev(conn->fdev->mdev); in mlx5_fpga_conn_unmap_buf()
91 dma_unmap_single(dma_device, buf->sg[1].dma_addr, in mlx5_fpga_conn_unmap_buf()
[all …]
/linux/include/rdma/
H A Dib_umem.h156 struct device *dma_device,
209 struct device *dma_device, in ib_umem_dmabuf_get_pinned_with_dma_device() argument
H A Dib_verbs.h2704 struct device *dma_device; member
2879 struct device *dma_device);
4037 return IS_ENABLED(CONFIG_INFINIBAND_VIRT_DMA) && !dev->dma_device; in ib_uses_virt_dma()
4048 return dma_pci_p2pdma_supported(dev->dma_device); in ib_dma_pci_p2p_dma_supported()
4085 return dma_mapping_error(dev->dma_device, dma_addr); in ib_dma_mapping_error()
4101 return dma_map_single(dev->dma_device, cpu_addr, size, direction); in ib_dma_map_single()
4116 dma_unmap_single(dev->dma_device, addr, size, direction); in ib_dma_unmap_single()
4135 return dma_map_page(dev->dma_device, page, offset, size, direction); in ib_dma_map_page()
4150 dma_unmap_page(dev->dma_device, addr, size, direction); in ib_dma_unmap_page()
4161 return dma_map_sg_attrs(dev->dma_device, sg, nents, direction, in ib_dma_map_sg_attrs()
[all …]
/linux/drivers/dma/idxd/
H A Ddma.c190 static void idxd_dma_release(struct dma_device *device) in idxd_dma_release()
200 struct dma_device *dma; in idxd_register_dma_device()
251 struct dma_device *dma = &idxd->idxd_dma->dma; in idxd_register_dma_channel()
/linux/drivers/infiniband/hw/hns/
H A Dhns_roce_db.c71 struct device *dma_device) in hns_roce_alloc_db_pgdir() argument
83 pgdir->page = dma_alloc_coherent(dma_device, PAGE_SIZE, in hns_roce_alloc_db_pgdir()
/linux/include/linux/platform_data/
H A Ddma-iop32x.h37 struct dma_device common;

123456