/linux/include/linux/ |
H A D | dmaengine.h | 575 struct dma_async_tx_descriptor; 578 int (*attach)(struct dma_async_tx_descriptor *desc, void *data, 581 void *(*get_ptr)(struct dma_async_tx_descriptor *desc, 583 int (*set_len)(struct dma_async_tx_descriptor *desc, 614 struct dma_async_tx_descriptor { struct 619 dma_cookie_t (*tx_submit)(struct dma_async_tx_descriptor *tx); argument 620 int (*desc_free)(struct dma_async_tx_descriptor *tx); argument 628 struct dma_async_tx_descriptor *next; argument 629 struct dma_async_tx_descriptor *parent; argument 635 static inline void dma_set_unmap(struct dma_async_tx_descriptor *tx, in dma_set_unmap() argument [all …]
|
/linux/crypto/async_tx/ |
H A D | async_tx.c | 46 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in __async_tx_find_channel() 65 async_tx_channel_switch(struct dma_async_tx_descriptor *depend_tx, in async_tx_channel_switch() 66 struct dma_async_tx_descriptor *tx) in async_tx_channel_switch() 70 struct dma_async_tx_descriptor *intr_tx = (void *) ~0; in async_tx_channel_switch() 143 async_tx_submit(struct dma_chan *chan, struct dma_async_tx_descriptor *tx, in async_tx_submit() 146 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_tx_submit() 220 struct dma_async_tx_descriptor * 225 struct dma_async_tx_descriptor *tx; in async_trigger_callback() 226 struct dma_async_tx_descriptor *depend_tx = submit->depend_tx; in async_trigger_callback() 263 void async_tx_quiesce(struct dma_async_tx_descriptor **tx) in async_tx_quiesce()
|
H A D | async_raid6_recov.c | 17 static struct dma_async_tx_descriptor * 36 struct dma_async_tx_descriptor *tx; in async_sum_product() 86 static struct dma_async_tx_descriptor * 104 struct dma_async_tx_descriptor *tx; in async_mult() 152 static struct dma_async_tx_descriptor * 157 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_4() 202 static struct dma_async_tx_descriptor * 207 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_5() 293 static struct dma_async_tx_descriptor * 298 struct dma_async_tx_descriptor *tx = NULL; in __2data_recov_n() [all …]
|
H A D | async_pq.c | 34 static __async_inline struct dma_async_tx_descriptor * 41 struct dma_async_tx_descriptor *tx = NULL; in do_async_gen_syndrome() 176 struct dma_async_tx_descriptor * 197 struct dma_async_tx_descriptor *tx; in async_gen_syndrome() 297 struct dma_async_tx_descriptor * 304 struct dma_async_tx_descriptor *tx; in async_syndrome_val()
|
H A D | async_memcpy.c | 31 struct dma_async_tx_descriptor * 39 struct dma_async_tx_descriptor *tx = NULL; in async_memcpy()
|
H A D | raid6test.c | 62 struct dma_async_tx_descriptor *tx = NULL; in raid6_dual_recov() 154 struct dma_async_tx_descriptor *tx; in test()
|
/linux/drivers/dma/ioat/ |
H A D | dma.h | 188 struct dma_async_tx_descriptor txd; 220 struct dma_async_tx_descriptor *tx, int id) in __dump_desc_dbg() 356 struct dma_async_tx_descriptor * 359 struct dma_async_tx_descriptor * 361 struct dma_async_tx_descriptor * 364 struct dma_async_tx_descriptor * 368 struct dma_async_tx_descriptor * 372 struct dma_async_tx_descriptor * 376 struct dma_async_tx_descriptor * 379 struct dma_async_tx_descriptor *
|
H A D | prep.c | 100 struct dma_async_tx_descriptor * 151 static struct dma_async_tx_descriptor * 244 struct dma_async_tx_descriptor * 256 struct dma_async_tx_descriptor * 338 static struct dma_async_tx_descriptor * 462 static struct dma_async_tx_descriptor * 573 struct dma_async_tx_descriptor * 618 struct dma_async_tx_descriptor * 646 struct dma_async_tx_descriptor * 672 struct dma_async_tx_descriptor * [all …]
|
/linux/drivers/dma/ |
H A D | dmaengine.h | 29 static inline dma_cookie_t dma_cookie_assign(struct dma_async_tx_descriptor *tx) in dma_cookie_assign() 52 static inline void dma_cookie_complete(struct dma_async_tx_descriptor *tx) in dma_cookie_complete() 114 dmaengine_desc_get_callback(struct dma_async_tx_descriptor *tx, in dmaengine_desc_get_callback() 160 dmaengine_desc_get_callback_invoke(struct dma_async_tx_descriptor *tx, in dmaengine_desc_get_callback_invoke()
|
H A D | virt-dma.c | 14 static struct virt_dma_desc *to_virt_desc(struct dma_async_tx_descriptor *tx) in to_virt_desc() 19 dma_cookie_t vchan_tx_submit(struct dma_async_tx_descriptor *tx) in vchan_tx_submit() 49 int vchan_tx_desc_free(struct dma_async_tx_descriptor *tx) in vchan_tx_desc_free()
|
H A D | dmaengine.c | 1473 void dma_async_tx_descriptor_init(struct dma_async_tx_descriptor *tx, in dma_async_tx_descriptor_init() 1484 struct dma_async_tx_descriptor *desc, enum dma_desc_metadata_mode mode) in desc_check_and_set_metadata_mode() 1499 int dmaengine_desc_attach_metadata(struct dma_async_tx_descriptor *desc, in dmaengine_desc_attach_metadata() 1518 void *dmaengine_desc_get_metadata_ptr(struct dma_async_tx_descriptor *desc, in dmaengine_desc_get_metadata_ptr() 1537 int dmaengine_desc_set_metadata_len(struct dma_async_tx_descriptor *desc, in dmaengine_desc_set_metadata_len() 1561 dma_wait_for_async_tx(struct dma_async_tx_descriptor *tx) in dma_wait_for_async_tx() 1588 void dma_run_dependencies(struct dma_async_tx_descriptor *tx) in dma_run_dependencies() 1590 struct dma_async_tx_descriptor *dep = txd_next(tx); in dma_run_dependencies() 1591 struct dma_async_tx_descriptor *dep_next; in dma_run_dependencies()
|
H A D | plx_dma.c | 98 struct dma_async_tx_descriptor txd; 124 static struct plx_dma_desc *to_plx_desc(struct dma_async_tx_descriptor *txd) in to_plx_desc() 251 static struct dma_async_tx_descriptor *plx_dma_prep_memcpy(struct dma_chan *c, in plx_dma_prep_memcpy() 300 static dma_cookie_t plx_dma_tx_submit(struct dma_async_tx_descriptor *desc) in plx_dma_tx_submit()
|
H A D | imx-dma.c | 123 struct dma_async_tx_descriptor desc; 157 struct dma_async_tx_descriptor desc; 723 static dma_cookie_t imxdma_tx_submit(struct dma_async_tx_descriptor *tx) in imxdma_tx_submit() 793 static struct dma_async_tx_descriptor *imxdma_prep_slave_sg( in imxdma_prep_slave_sg() 846 static struct dma_async_tx_descriptor *imxdma_prep_dma_cyclic( in imxdma_prep_dma_cyclic() 904 static struct dma_async_tx_descriptor *imxdma_prep_dma_memcpy( in imxdma_prep_dma_memcpy() 935 static struct dma_async_tx_descriptor *imxdma_prep_dma_interleaved( in imxdma_prep_dma_interleaved()
|
H A D | at_xdmac.c | 275 struct dma_async_tx_descriptor tx_dma_desc; 337 static inline struct at_xdmac_desc *txd_to_at_desc(struct dma_async_tx_descriptor *txd) in txd_to_at_desc() 545 static dma_cookie_t at_xdmac_tx_submit(struct dma_async_tx_descriptor *tx) in at_xdmac_tx_submit() 761 static struct dma_async_tx_descriptor * 772 struct dma_async_tx_descriptor *ret = NULL; in at_xdmac_prep_slave_sg() 863 static struct dma_async_tx_descriptor * 1071 static struct dma_async_tx_descriptor * 1168 static struct dma_async_tx_descriptor * 1352 static struct dma_async_tx_descriptor * 1377 static struct dma_async_tx_descriptor * [all …]
|
/linux/drivers/dma/idxd/ |
H A D | dma.c | 28 struct dma_async_tx_descriptor *tx; in idxd_dma_complete_txd() 85 static struct dma_async_tx_descriptor * 106 static struct dma_async_tx_descriptor * 171 static dma_cookie_t idxd_dma_tx_submit(struct dma_async_tx_descriptor *tx) in idxd_dma_tx_submit()
|
/linux/Documentation/driver-api/dmaengine/ |
H A D | client.rst | 104 struct dma_async_tx_descriptor *dmaengine_prep_slave_sg( 109 struct dma_async_tx_descriptor *dmaengine_prep_peripheral_dma_vec( 114 struct dma_async_tx_descriptor *dmaengine_prep_dma_cyclic( 118 struct dma_async_tx_descriptor *dmaengine_prep_interleaved_dma( 178 int dmaengine_desc_attach_metadata(struct dma_async_tx_descriptor *desc, 197 void *dmaengine_desc_get_metadata_ptr(struct dma_async_tx_descriptor *desc, 200 int dmaengine_desc_set_metadata_len(struct dma_async_tx_descriptor *desc, 269 dma_cookie_t dmaengine_submit(struct dma_async_tx_descriptor *desc) 280 (``struct dma_async_tx_descriptor``) belongs to the DMA engine.
|
/linux/include/linux/mtd/ |
H A D | nand-qpic-common.h | 260 struct dma_async_tx_descriptor *last_data_desc; 261 struct dma_async_tx_descriptor *last_cmd_desc; 293 struct dma_async_tx_descriptor *dma_desc;
|
/linux/drivers/md/ |
H A D | raid5-log.h | 34 struct dma_async_tx_descriptor * 36 struct dma_async_tx_descriptor *tx);
|
/linux/include/linux/dma/ |
H A D | mxs-dma.h | 16 static inline struct dma_async_tx_descriptor *mxs_dmaengine_prep_pio( in mxs_dmaengine_prep_pio()
|
/linux/drivers/mmc/host/ |
H A D | mxs-mmc.c | 213 static struct dma_async_tx_descriptor *mxs_mmc_prep_dma( in mxs_mmc_prep_dma() 217 struct dma_async_tx_descriptor *desc; in mxs_mmc_prep_dma() 252 struct dma_async_tx_descriptor *desc; in mxs_mmc_bc() 286 struct dma_async_tx_descriptor *desc; in mxs_mmc_ac() 347 struct dma_async_tx_descriptor *desc; in mxs_mmc_adtc()
|
/linux/drivers/spi/ |
H A D | spi-pxa2xx-dma.c | 66 static struct dma_async_tx_descriptor * 139 struct dma_async_tx_descriptor *tx_desc, *rx_desc; in pxa2xx_spi_dma_prepare()
|
H A D | spi-ingenic.c | 121 static struct dma_async_tx_descriptor * 132 struct dma_async_tx_descriptor *desc; in spi_ingenic_prepare_dma() 178 struct dma_async_tx_descriptor *rx_desc, *tx_desc; in spi_ingenic_dma_tx()
|
/linux/include/linux/platform_data/ |
H A D | dma-iop32x.h | 92 struct dma_async_tx_descriptor async_tx;
|
/linux/drivers/crypto/ccp/ |
H A D | ccp-dmaengine.c | 193 struct dma_async_tx_descriptor *tx_desc; in ccp_handle_active_desc() 297 static dma_cookie_t ccp_tx_submit(struct dma_async_tx_descriptor *tx_desc) in ccp_tx_submit() 478 static struct dma_async_tx_descriptor *ccp_prep_dma_memcpy( in ccp_prep_dma_memcpy() 506 static struct dma_async_tx_descriptor *ccp_prep_dma_interrupt( in ccp_prep_dma_interrupt()
|
/linux/drivers/dma/amd/ptdma/ |
H A D | ptdma-dmaengine.c | 199 struct dma_async_tx_descriptor *tx_desc; in pt_handle_active_desc() 338 struct dma_async_tx_descriptor *tx_desc; in pt_cmd_callback_work() 422 static struct dma_async_tx_descriptor * 435 static struct dma_async_tx_descriptor *
|