| /linux/drivers/dma/amd/qdma/ |
| H A D | qdma.c | 34 static inline struct qdma_mm_vdesc *to_qdma_vdesc(struct virt_dma_desc *vdesc) in to_qdma_vdesc() argument 36 return container_of(vdesc, struct qdma_mm_vdesc, vdesc); in to_qdma_vdesc() 429 static void qdma_free_vdesc(struct virt_dma_desc *vdesc) in qdma_free_vdesc() argument 431 struct qdma_mm_vdesc *vd = to_qdma_vdesc(vdesc); in qdma_free_vdesc() 658 static int qdma_hw_enqueue(struct qdma_queue *q, struct qdma_mm_vdesc *vdesc) in qdma_hw_enqueue() argument 667 if (!vdesc->sg_len) in qdma_hw_enqueue() 671 dst = &vdesc->dev_addr; in qdma_hw_enqueue() 675 src = &vdesc->dev_addr; in qdma_hw_enqueue() 678 for_each_sg(vdesc->sgl, sg, vdesc->sg_len, i) { in qdma_hw_enqueue() 679 addr = sg_dma_address(sg) + vdesc->sg_off; in qdma_hw_enqueue() [all …]
|
| H A D | qdma.h | 191 struct virt_dma_desc vdesc; member 202 #define QDMA_VDESC_QUEUED(vdesc) (!(vdesc)->sg_len) argument
|
| /linux/drivers/dma/ |
| H A D | dma-jz4780.c | 119 struct virt_dma_desc vdesc; member 171 struct virt_dma_desc *vdesc) in to_jz4780_dma_desc() argument 173 return container_of(vdesc, struct jz4780_dma_desc, vdesc); in to_jz4780_dma_desc() 262 static void jz4780_dma_desc_free(struct virt_dma_desc *vdesc) in jz4780_dma_desc_free() argument 264 struct jz4780_dma_desc *desc = to_jz4780_dma_desc(vdesc); in jz4780_dma_desc_free() 265 struct jz4780_dma_chan *jzchan = to_jz4780_dma_chan(vdesc->tx.chan); in jz4780_dma_desc_free() 382 jz4780_dma_desc_free(&jzchan->desc->vdesc); in jz4780_dma_prep_slave_sg() 403 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_slave_sg() 429 jz4780_dma_desc_free(&jzchan->desc->vdesc); in jz4780_dma_prep_dma_cyclic() 455 return vchan_tx_prep(&jzchan->vchan, &desc->vdesc, flags); in jz4780_dma_prep_dma_cyclic() [all …]
|
| H A D | idma64.c | 115 struct virt_dma_desc *vdesc; in idma64_start_transfer() local 118 vdesc = vchan_next_desc(&idma64c->vchan); in idma64_start_transfer() 119 if (!vdesc) { in idma64_start_transfer() 124 list_del(&vdesc->node); in idma64_start_transfer() 125 idma64c->desc = to_idma64_desc(vdesc); in idma64_start_transfer() 154 vchan_cookie_complete(&desc->vdesc); in idma64_chan_irq() 230 static void idma64_vdesc_free(struct virt_dma_desc *vdesc) in idma64_vdesc_free() argument 232 struct idma64_chan *idma64c = to_idma64_chan(vdesc->tx.chan); in idma64_vdesc_free() 234 idma64_desc_free(idma64c, to_idma64_desc(vdesc)); in idma64_vdesc_free() 334 return vchan_tx_prep(&idma64c->vchan, &desc->vdesc, flags); in idma64_prep_slave_sg() [all …]
|
| H A D | st_fdma.c | 30 return container_of(vd, struct st_fdma_desc, vdesc); in to_st_fdma_desc() 77 struct virt_dma_desc *vdesc; in st_fdma_xfer_desc() local 80 vdesc = vchan_next_desc(&fchan->vchan); in st_fdma_xfer_desc() 81 if (!vdesc) in st_fdma_xfer_desc() 84 fchan->fdesc = to_st_fdma_desc(vdesc); in st_fdma_xfer_desc() 145 list_del(&fchan->fdesc->vdesc.node); in st_fdma_irq_handler() 146 vchan_cookie_complete(&fchan->fdesc->vdesc); in st_fdma_irq_handler() 150 vchan_cyclic_callback(&fchan->fdesc->vdesc); in st_fdma_irq_handler() 226 static void st_fdma_free_desc(struct virt_dma_desc *vdesc) in st_fdma_free_desc() argument 231 fdesc = to_st_fdma_desc(vdesc); in st_fdma_free_desc() [all …]
|
| H A D | idma64.h | 114 struct virt_dma_desc vdesc; member 122 static inline struct idma64_desc *to_idma64_desc(struct virt_dma_desc *vdesc) in to_idma64_desc() argument 124 return container_of(vdesc, struct idma64_desc, vdesc); in to_idma64_desc()
|
| H A D | fsl-edma-common.h | 199 struct virt_dma_desc vdesc; member 475 return container_of(vd, struct fsl_edma_desc, vdesc); in to_fsl_edma_desc() 487 void fsl_edma_free_desc(struct virt_dma_desc *vdesc);
|
| H A D | tegra186-gpc-dma.c | 514 struct virt_dma_desc *vdesc; in tegra_dma_start() local 517 vdesc = vchan_next_desc(&tdc->vc); in tegra_dma_start() 518 if (!vdesc) in tegra_dma_start() 521 dma_desc = vd_to_tegra_dma_desc(vdesc); in tegra_dma_start() 522 list_del(&vdesc->node); in tegra_dma_start()
|
| H A D | img-mdc-dma.c | 175 struct virt_dma_desc *vdesc = container_of(t, struct virt_dma_desc, tx); in to_mdc_desc() local 177 return container_of(vdesc, struct mdc_tx_desc, vd); in to_mdc_desc()
|
| /linux/drivers/dma/hsu/ |
| H A D | hsu.c | 118 struct virt_dma_desc *vdesc; in hsu_dma_start_transfer() local 121 vdesc = vchan_next_desc(&hsuc->vchan); in hsu_dma_start_transfer() 122 if (!vdesc) { in hsu_dma_start_transfer() 127 list_del(&vdesc->node); in hsu_dma_start_transfer() 128 hsuc->desc = to_hsu_dma_desc(vdesc); in hsu_dma_start_transfer() 232 vchan_cookie_complete(&desc->vdesc); in hsu_dma_do_irq() 261 static void hsu_dma_desc_free(struct virt_dma_desc *vdesc) in hsu_dma_desc_free() argument 263 struct hsu_dma_desc *desc = to_hsu_dma_desc(vdesc); in hsu_dma_desc_free() 295 return vchan_tx_prep(&hsuc->vchan, &desc->vdesc, flags); in hsu_dma_prep_slave_sg() 330 struct virt_dma_desc *vdesc; in hsu_dma_tx_status() local [all …]
|
| H A D | hsu.h | 72 struct virt_dma_desc vdesc; member 81 static inline struct hsu_dma_desc *to_hsu_dma_desc(struct virt_dma_desc *vdesc) in to_hsu_dma_desc() argument 83 return container_of(vdesc, struct hsu_dma_desc, vdesc); in to_hsu_dma_desc()
|
| /linux/drivers/dma/xilinx/ |
| H A D | xilinx_dpdma.c | 199 struct virt_dma_desc vdesc; member 206 container_of(_desc, struct xilinx_dpdma_tx_desc, vdesc) 655 static void xilinx_dpdma_chan_free_tx_desc(struct virt_dma_desc *vdesc) in xilinx_dpdma_chan_free_tx_desc() argument 660 if (!vdesc) in xilinx_dpdma_chan_free_tx_desc() 663 desc = to_dpdma_tx_desc(vdesc); in xilinx_dpdma_chan_free_tx_desc() 743 return vchan_tx_prep(&chan->vchan, &tx_desc->vdesc, flags); in xilinx_dpdma_chan_prep_cyclic() 746 xilinx_dpdma_chan_free_tx_desc(&tx_desc->vdesc); in xilinx_dpdma_chan_prep_cyclic() 785 xilinx_dpdma_chan_free_tx_desc(&tx_desc->vdesc); in xilinx_dpdma_chan_prep_interleaved_dma() 909 struct virt_dma_desc *vdesc; in xilinx_dpdma_chan_queue_transfer() local 925 vdesc = vchan_next_desc(&chan->vchan); in xilinx_dpdma_chan_queue_transfer() [all …]
|
| H A D | xdma.c | 97 struct virt_dma_desc vdesc; member 215 static inline struct xdma_desc *to_xdma_desc(struct virt_dma_desc *vdesc) in to_xdma_desc() argument 217 return container_of(vdesc, struct xdma_desc, vdesc); in to_xdma_desc() 246 static void xdma_free_desc(struct virt_dma_desc *vdesc) in xdma_free_desc() argument 251 sw_desc = to_xdma_desc(vdesc); in xdma_free_desc() 319 xdma_free_desc(&sw_desc->vdesc); in xdma_alloc_desc() 639 tx_desc = vchan_tx_prep(&xdma_chan->vchan, &sw_desc->vdesc, flags); in xdma_prep_device_sg() 646 xdma_free_desc(&sw_desc->vdesc); in xdma_prep_device_sg() 715 tx_desc = vchan_tx_prep(&xdma_chan->vchan, &sw_desc->vdesc, flags); in xdma_prep_dma_cyclic() 722 xdma_free_desc(&sw_desc->vdesc); in xdma_prep_dma_cyclic() [all …]
|
| /linux/drivers/dma/sf-pdma/ |
| H A D | sf-pdma.c | 52 return container_of(vd, struct sf_pdma_desc, vdesc); in to_sf_pdma_desc() 103 desc->async_tx = vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in sf_pdma_prep_dma_memcpy() 244 struct virt_dma_desc *vdesc; in sf_pdma_get_first_pending_desc() local 249 vdesc = list_first_entry(&vchan->desc_issued, struct virt_dma_desc, node); in sf_pdma_get_first_pending_desc() 251 return container_of(vdesc, struct sf_pdma_desc, vdesc); in sf_pdma_get_first_pending_desc() 290 static void sf_pdma_free_desc(struct virt_dma_desc *vdesc) in sf_pdma_free_desc() argument 294 desc = to_sf_pdma_desc(vdesc); in sf_pdma_free_desc() 312 list_del(&chan->desc->vdesc.node); in sf_pdma_donebh_tasklet() 313 vchan_cookie_complete(&chan->desc->vdesc); in sf_pdma_donebh_tasklet()
|
| H A D | sf-pdma.h | 80 struct virt_dma_desc vdesc; member
|
| /linux/drivers/dma/stm32/ |
| H A D | stm32-dma.c | 190 struct virt_dma_desc vdesc; member 246 static struct stm32_dma_desc *to_stm32_dma_desc(struct virt_dma_desc *vdesc) in to_stm32_dma_desc() argument 248 return container_of(vdesc, struct stm32_dma_desc, vdesc); in to_stm32_dma_desc() 508 dma_cookie_complete(&chan->desc->vdesc.tx); in stm32_dma_terminate_all() 509 vchan_terminate_vdesc(&chan->desc->vdesc); in stm32_dma_terminate_all() 559 struct virt_dma_desc *vdesc; in stm32_dma_start_transfer() local 570 vdesc = vchan_next_desc(&chan->vchan); in stm32_dma_start_transfer() 571 if (!vdesc) in stm32_dma_start_transfer() 574 list_del(&vdesc->node); in stm32_dma_start_transfer() 576 chan->desc = to_stm32_dma_desc(vdesc); in stm32_dma_start_transfer() [all …]
|
| H A D | stm32-mdma.c | 223 struct virt_dma_desc vdesc; member 274 static struct stm32_mdma_desc *to_stm32_mdma_desc(struct virt_dma_desc *vdesc) in to_stm32_mdma_desc() argument 276 return container_of(vdesc, struct stm32_mdma_desc, vdesc); in to_stm32_mdma_desc() 345 static void stm32_mdma_desc_free(struct virt_dma_desc *vdesc) in stm32_mdma_desc_free() argument 347 struct stm32_mdma_desc *desc = to_stm32_mdma_desc(vdesc); in stm32_mdma_desc_free() 348 struct stm32_mdma_chan *chan = to_stm32_mdma_chan(vdesc->tx.chan); in stm32_mdma_desc_free() 830 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_slave_sg() 926 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_dma_cyclic() 1112 return vchan_tx_prep(&chan->vchan, &desc->vdesc, flags); in stm32_mdma_prep_dma_memcpy() 1144 struct virt_dma_desc *vdesc; in stm32_mdma_start_transfer() local [all …]
|
| H A D | stm32-dma3.c | 270 struct virt_dma_desc vdesc; member 326 static inline struct stm32_dma3_swdesc *to_stm32_dma3_swdesc(struct virt_dma_desc *vdesc) in to_stm32_dma3_swdesc() argument 328 return container_of(vdesc, struct stm32_dma3_swdesc, vdesc); in to_stm32_dma3_swdesc() 460 static void stm32_dma3_chan_vdesc_free(struct virt_dma_desc *vdesc) in stm32_dma3_chan_vdesc_free() argument 462 struct stm32_dma3_swdesc *swdesc = to_stm32_dma3_swdesc(vdesc); in stm32_dma3_chan_vdesc_free() 463 struct stm32_dma3_chan *chan = to_stm32_dma3_chan(vdesc->tx.chan); in stm32_dma3_chan_vdesc_free() 773 struct virt_dma_desc *vdesc; in stm32_dma3_chan_start() local 778 vdesc = vchan_next_desc(&chan->vchan); in stm32_dma3_chan_start() 779 if (!vdesc) { in stm32_dma3_chan_start() 783 list_del(&vdesc->node); in stm32_dma3_chan_start() [all …]
|
| /linux/drivers/sh/intc/ |
| H A D | virq.c | 122 struct irq_desc *vdesc = irq_to_desc(entry->irq); in intc_virq_handler() local 124 if (vdesc) { in intc_virq_handler() 125 handle = (unsigned long)irq_desc_get_handler_data(vdesc); in intc_virq_handler() 128 generic_handle_irq_desc(vdesc); in intc_virq_handler()
|
| /linux/drivers/dma/fsl-dpaa2-qdma/ |
| H A D | dpaa2-qdma.c | 26 return container_of(vd, struct dpaa2_qdma_comp, vdesc); in to_fsl_qdma_comp() 269 return vchan_tx_prep(&dpaa2_chan->vchan, &dpaa2_comp->vdesc, flags); in dpaa2_qdma_prep_memcpy() 276 struct virt_dma_desc *vdesc; in dpaa2_qdma_issue_pending() local 284 vdesc = vchan_next_desc(&dpaa2_chan->vchan); in dpaa2_qdma_issue_pending() 285 if (!vdesc) in dpaa2_qdma_issue_pending() 287 dpaa2_comp = to_fsl_qdma_comp(vdesc); in dpaa2_qdma_issue_pending() 291 list_del(&vdesc->node); in dpaa2_qdma_issue_pending() 443 dpaa2_comp->vdesc); in dpaa2_qdma_fqdan_cb() 621 static void dpaa2_qdma_free_desc(struct virt_dma_desc *vdesc) in dpaa2_qdma_free_desc() argument 627 dpaa2_comp = to_fsl_qdma_comp(vdesc); in dpaa2_qdma_free_desc()
|
| H A D | dpaa2-qdma.h | 73 struct virt_dma_desc vdesc; member 96 struct virt_dma_desc vdesc; member
|
| /linux/drivers/dma/ti/ |
| H A D | edma.c | 171 struct virt_dma_desc vdesc; member 709 return container_of(tx, struct edma_desc, vdesc.tx); in to_edma_desc() 712 static void edma_desc_free(struct virt_dma_desc *vdesc) in edma_desc_free() argument 714 kfree(container_of(vdesc, struct edma_desc, vdesc)); in edma_desc_free() 721 struct virt_dma_desc *vdesc; in edma_execute() local 728 vdesc = vchan_next_desc(&echan->vchan); in edma_execute() 729 if (!vdesc) in edma_execute() 731 list_del(&vdesc->node); in edma_execute() 732 echan->edesc = to_edma_desc(&vdesc->tx); in edma_execute() 830 vchan_terminate_vdesc(&echan->edesc->vdesc); in edma_terminate_all() [all …]
|
| /linux/drivers/dma/lgm/ |
| H A D | lgm-dma.c | 270 struct virt_dma_desc vdesc; member 300 static inline struct dw2_desc_sw *to_lgm_dma_desc(struct virt_dma_desc *vdesc) in to_lgm_dma_desc() argument 302 return container_of(vdesc, struct dw2_desc_sw, vdesc); in to_lgm_dma_desc() 688 tx = &ds->vdesc.tx; in ldma_chan_desc_cfg() 965 static void dma_free_desc_resource(struct virt_dma_desc *vdesc) in dma_free_desc_resource() argument 967 struct dw2_desc_sw *ds = to_lgm_dma_desc(vdesc); in dma_free_desc_resource() 1023 struct virt_dma_desc *vdesc; in ldma_issue_pending() local 1026 vdesc = vchan_next_desc(&c->vchan); in ldma_issue_pending() 1027 if (!vdesc) { in ldma_issue_pending() 1032 list_del(&vdesc->node); in ldma_issue_pending() [all …]
|
| /linux/drivers/dma/dw-edma/ |
| H A D | dw-edma-core.c | 169 static void vchan_free_desc(struct virt_dma_desc *vdesc) in vchan_free_desc() argument 171 dw_edma_free_desc(vd2dw_edma_desc(vdesc)); in vchan_free_desc()
|