Lines Matching +full:imx28 +full:- +full:dma +full:- +full:apbh
1 // SPDX-License-Identifier: GPL-2.0
5 // Refer to drivers/dma/imx-sdma.c
16 #include <linux/dma-mapping.h>
26 #include <linux/dma/mxs-dma.h>
33 * NOTE: The term "PIO" throughout the mxs-dma implementation means
34 * PIO mode of mxs apbh-dma and apbx-dma. With this working mode,
35 * dma can program the controller registers of peripheral devices.
38 #define dma_is_apbh(mxs_dma) ((mxs_dma)->type == MXS_DMA_APBH)
39 #define apbh_is_old(mxs_dma) ((mxs_dma)->dev_id == IMX23_DMA)
50 * The offset of NXTCMDAR register is different per both dma type and version,
67 * NAND_LOCK: 4 (1) - not implemented
68 * NAND_WAIT4READY: 5 (1) - not implemented
170 { .compatible = "fsl,imx23-dma-apbh", .data = &mxs_dma_types[0], },
171 { .compatible = "fsl,imx23-dma-apbx", .data = &mxs_dma_types[1], },
172 { .compatible = "fsl,imx28-dma-apbh", .data = &mxs_dma_types[2], },
173 { .compatible = "fsl,imx28-dma-apbx", .data = &mxs_dma_types[3], },
186 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_reset_chan()
187 int chan_id = mxs_chan->chan.chan_id; in mxs_dma_reset_chan()
190 * mxs dma channel resets can cause a channel stall. To recover from a in mxs_dma_reset_chan()
191 * channel stall, we have to reset the whole DMA engine. To avoid this, in mxs_dma_reset_chan()
192 * we use cyclic DMA with semaphores, that are enhanced in in mxs_dma_reset_chan()
196 if (mxs_chan->flags & MXS_DMA_USE_SEMAPHORE && in mxs_dma_reset_chan()
197 mxs_chan->flags & MXS_DMA_SG_LOOP) { in mxs_dma_reset_chan()
198 mxs_chan->reset = true; in mxs_dma_reset_chan()
201 mxs_dma->base + HW_APBHX_CTRL0 + STMP_OFFSET_REG_SET); in mxs_dma_reset_chan()
205 void __iomem *reg_dbg1 = mxs_dma->base + in mxs_dma_reset_chan()
209 * On i.MX28 APBX, the DMA channel can stop working if we reset in mxs_dma_reset_chan()
221 dev_err(&mxs_chan->mxs_dma->pdev->dev, in mxs_dma_reset_chan()
222 …"Failed waiting for the DMA channel %d to leave state READ_FLUSH, trying to reset channel in READ_… in mxs_dma_reset_chan()
226 mxs_dma->base + HW_APBHX_CHANNEL_CTRL + STMP_OFFSET_REG_SET); in mxs_dma_reset_chan()
229 mxs_chan->status = DMA_COMPLETE; in mxs_dma_reset_chan()
235 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_enable_chan()
236 int chan_id = mxs_chan->chan.chan_id; in mxs_dma_enable_chan()
239 writel(mxs_chan->ccw_phys, in mxs_dma_enable_chan()
240 mxs_dma->base + HW_APBHX_CHn_NXTCMDAR(mxs_dma, chan_id)); in mxs_dma_enable_chan()
243 if (mxs_chan->flags & MXS_DMA_USE_SEMAPHORE && in mxs_dma_enable_chan()
244 mxs_chan->flags & MXS_DMA_SG_LOOP) { in mxs_dma_enable_chan()
245 /* A cyclic DMA consists of at least 2 segments, so initialize in mxs_dma_enable_chan()
248 writel(2, mxs_dma->base + HW_APBHX_CHn_SEMA(mxs_dma, chan_id)); in mxs_dma_enable_chan()
250 writel(1, mxs_dma->base + HW_APBHX_CHn_SEMA(mxs_dma, chan_id)); in mxs_dma_enable_chan()
252 mxs_chan->reset = false; in mxs_dma_enable_chan()
259 mxs_chan->status = DMA_COMPLETE; in mxs_dma_disable_chan()
265 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_pause_chan()
266 int chan_id = mxs_chan->chan.chan_id; in mxs_dma_pause_chan()
271 mxs_dma->base + HW_APBHX_CTRL0 + STMP_OFFSET_REG_SET); in mxs_dma_pause_chan()
274 mxs_dma->base + HW_APBHX_CHANNEL_CTRL + STMP_OFFSET_REG_SET); in mxs_dma_pause_chan()
276 mxs_chan->status = DMA_PAUSED; in mxs_dma_pause_chan()
283 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_resume_chan()
284 int chan_id = mxs_chan->chan.chan_id; in mxs_dma_resume_chan()
289 mxs_dma->base + HW_APBHX_CTRL0 + STMP_OFFSET_REG_CLR); in mxs_dma_resume_chan()
292 mxs_dma->base + HW_APBHX_CHANNEL_CTRL + STMP_OFFSET_REG_CLR); in mxs_dma_resume_chan()
294 mxs_chan->status = DMA_IN_PROGRESS; in mxs_dma_resume_chan()
307 dmaengine_desc_get_callback_invoke(&mxs_chan->desc, NULL); in mxs_dma_tasklet()
314 for (i = 0; i != mxs_dma->nr_channels; ++i) in mxs_dma_irq_to_chan()
315 if (mxs_dma->mxs_chans[i].chan_irq == irq) in mxs_dma_irq_to_chan()
318 return -EINVAL; in mxs_dma_irq_to_chan()
333 completed = readl(mxs_dma->base + HW_APBHX_CTRL1); in mxs_dma_int_handler()
338 mxs_dma->base + HW_APBHX_CTRL1 + STMP_OFFSET_REG_CLR); in mxs_dma_int_handler()
341 err = readl(mxs_dma->base + HW_APBHX_CTRL2); in mxs_dma_int_handler()
353 mxs_dma->base + HW_APBHX_CTRL2 + STMP_OFFSET_REG_CLR); in mxs_dma_int_handler()
362 err -= err & completed; in mxs_dma_int_handler()
364 mxs_chan = &mxs_dma->mxs_chans[chan]; in mxs_dma_int_handler()
367 dev_dbg(mxs_dma->dma_device.dev, in mxs_dma_int_handler()
370 mxs_chan->status = DMA_ERROR; in mxs_dma_int_handler()
371 mxs_dma_reset_chan(&mxs_chan->chan); in mxs_dma_int_handler()
372 } else if (mxs_chan->status != DMA_COMPLETE) { in mxs_dma_int_handler()
373 if (mxs_chan->flags & MXS_DMA_SG_LOOP) { in mxs_dma_int_handler()
374 mxs_chan->status = DMA_IN_PROGRESS; in mxs_dma_int_handler()
375 if (mxs_chan->flags & MXS_DMA_USE_SEMAPHORE) in mxs_dma_int_handler()
376 writel(1, mxs_dma->base + in mxs_dma_int_handler()
379 mxs_chan->status = DMA_COMPLETE; in mxs_dma_int_handler()
383 if (mxs_chan->status == DMA_COMPLETE) { in mxs_dma_int_handler()
384 if (mxs_chan->reset) in mxs_dma_int_handler()
386 dma_cookie_complete(&mxs_chan->desc); in mxs_dma_int_handler()
390 tasklet_schedule(&mxs_chan->tasklet); in mxs_dma_int_handler()
398 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_alloc_chan_resources()
401 mxs_chan->ccw = dma_alloc_coherent(mxs_dma->dma_device.dev, in mxs_dma_alloc_chan_resources()
403 &mxs_chan->ccw_phys, GFP_KERNEL); in mxs_dma_alloc_chan_resources()
404 if (!mxs_chan->ccw) { in mxs_dma_alloc_chan_resources()
405 ret = -ENOMEM; in mxs_dma_alloc_chan_resources()
409 ret = request_irq(mxs_chan->chan_irq, mxs_dma_int_handler, in mxs_dma_alloc_chan_resources()
410 0, "mxs-dma", mxs_dma); in mxs_dma_alloc_chan_resources()
414 ret = clk_prepare_enable(mxs_dma->clk); in mxs_dma_alloc_chan_resources()
420 dma_async_tx_descriptor_init(&mxs_chan->desc, chan); in mxs_dma_alloc_chan_resources()
421 mxs_chan->desc.tx_submit = mxs_dma_tx_submit; in mxs_dma_alloc_chan_resources()
424 async_tx_ack(&mxs_chan->desc); in mxs_dma_alloc_chan_resources()
429 free_irq(mxs_chan->chan_irq, mxs_dma); in mxs_dma_alloc_chan_resources()
431 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_alloc_chan_resources()
432 mxs_chan->ccw, mxs_chan->ccw_phys); in mxs_dma_alloc_chan_resources()
440 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_free_chan_resources()
444 free_irq(mxs_chan->chan_irq, mxs_dma); in mxs_dma_free_chan_resources()
446 dma_free_coherent(mxs_dma->dma_device.dev, CCW_BLOCK_SIZE, in mxs_dma_free_chan_resources()
447 mxs_chan->ccw, mxs_chan->ccw_phys); in mxs_dma_free_chan_resources()
449 clk_disable_unprepare(mxs_dma->clk); in mxs_dma_free_chan_resources()
453 * How to use the flags for ->device_prep_slave_sg() :
454 * [1] If there is only one DMA command in the DMA chain, the code should be:
456 * ->device_prep_slave_sg(DMA_CTRL_ACK);
458 * [2] If there are two DMA commands in the DMA chain, the code should be
460 * ->device_prep_slave_sg(0);
462 * ->device_prep_slave_sg(DMA_CTRL_ACK);
464 * [3] If there are more than two DMA commands in the DMA chain, the code
467 * ->device_prep_slave_sg(0); // First
469 * ->device_prep_slave_sg(DMA_CTRL_ACK]);
471 * ->device_prep_slave_sg(DMA_CTRL_ACK); // Last
480 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_prep_slave_sg()
487 if (mxs_chan->status == DMA_IN_PROGRESS) in mxs_dma_prep_slave_sg()
488 idx = mxs_chan->desc_count; in mxs_dma_prep_slave_sg()
491 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_slave_sg()
497 mxs_chan->status = DMA_IN_PROGRESS; in mxs_dma_prep_slave_sg()
498 mxs_chan->flags = 0; in mxs_dma_prep_slave_sg()
506 ccw = &mxs_chan->ccw[idx - 1]; in mxs_dma_prep_slave_sg()
507 ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; in mxs_dma_prep_slave_sg()
508 ccw->bits |= CCW_CHAIN; in mxs_dma_prep_slave_sg()
509 ccw->bits &= ~CCW_IRQ; in mxs_dma_prep_slave_sg()
510 ccw->bits &= ~CCW_DEC_SEM; in mxs_dma_prep_slave_sg()
516 ccw = &mxs_chan->ccw[idx++]; in mxs_dma_prep_slave_sg()
520 ccw->pio_words[j++] = *pio++; in mxs_dma_prep_slave_sg()
522 ccw->bits = 0; in mxs_dma_prep_slave_sg()
523 ccw->bits |= CCW_IRQ; in mxs_dma_prep_slave_sg()
524 ccw->bits |= CCW_DEC_SEM; in mxs_dma_prep_slave_sg()
526 ccw->bits |= CCW_WAIT4END; in mxs_dma_prep_slave_sg()
527 ccw->bits |= CCW_HALT_ON_TERM; in mxs_dma_prep_slave_sg()
528 ccw->bits |= CCW_TERM_FLUSH; in mxs_dma_prep_slave_sg()
529 ccw->bits |= BF_CCW(sg_len, PIO_NUM); in mxs_dma_prep_slave_sg()
530 ccw->bits |= BF_CCW(MXS_DMA_CMD_NO_XFER, COMMAND); in mxs_dma_prep_slave_sg()
532 ccw->bits |= CCW_WAIT4RDY; in mxs_dma_prep_slave_sg()
536 dev_err(mxs_dma->dma_device.dev, "maximum bytes for sg entry exceeded: %d > %d\n", in mxs_dma_prep_slave_sg()
541 ccw = &mxs_chan->ccw[idx++]; in mxs_dma_prep_slave_sg()
543 ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * idx; in mxs_dma_prep_slave_sg()
544 ccw->bufaddr = sg->dma_address; in mxs_dma_prep_slave_sg()
545 ccw->xfer_bytes = sg_dma_len(sg); in mxs_dma_prep_slave_sg()
547 ccw->bits = 0; in mxs_dma_prep_slave_sg()
548 ccw->bits |= CCW_CHAIN; in mxs_dma_prep_slave_sg()
549 ccw->bits |= CCW_HALT_ON_TERM; in mxs_dma_prep_slave_sg()
550 ccw->bits |= CCW_TERM_FLUSH; in mxs_dma_prep_slave_sg()
551 ccw->bits |= BF_CCW(direction == DMA_DEV_TO_MEM ? in mxs_dma_prep_slave_sg()
556 ccw->bits &= ~CCW_CHAIN; in mxs_dma_prep_slave_sg()
557 ccw->bits |= CCW_IRQ; in mxs_dma_prep_slave_sg()
558 ccw->bits |= CCW_DEC_SEM; in mxs_dma_prep_slave_sg()
560 ccw->bits |= CCW_WAIT4END; in mxs_dma_prep_slave_sg()
564 mxs_chan->desc_count = idx; in mxs_dma_prep_slave_sg()
566 return &mxs_chan->desc; in mxs_dma_prep_slave_sg()
569 mxs_chan->status = DMA_ERROR; in mxs_dma_prep_slave_sg()
579 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_prep_dma_cyclic()
583 if (mxs_chan->status == DMA_IN_PROGRESS) in mxs_dma_prep_dma_cyclic()
586 mxs_chan->status = DMA_IN_PROGRESS; in mxs_dma_prep_dma_cyclic()
587 mxs_chan->flags |= MXS_DMA_SG_LOOP; in mxs_dma_prep_dma_cyclic()
588 mxs_chan->flags |= MXS_DMA_USE_SEMAPHORE; in mxs_dma_prep_dma_cyclic()
591 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
598 dev_err(mxs_dma->dma_device.dev, in mxs_dma_prep_dma_cyclic()
605 struct mxs_dma_ccw *ccw = &mxs_chan->ccw[i]; in mxs_dma_prep_dma_cyclic()
608 ccw->next = mxs_chan->ccw_phys; in mxs_dma_prep_dma_cyclic()
610 ccw->next = mxs_chan->ccw_phys + sizeof(*ccw) * (i + 1); in mxs_dma_prep_dma_cyclic()
612 ccw->bufaddr = dma_addr; in mxs_dma_prep_dma_cyclic()
613 ccw->xfer_bytes = period_len; in mxs_dma_prep_dma_cyclic()
615 ccw->bits = 0; in mxs_dma_prep_dma_cyclic()
616 ccw->bits |= CCW_CHAIN; in mxs_dma_prep_dma_cyclic()
617 ccw->bits |= CCW_IRQ; in mxs_dma_prep_dma_cyclic()
618 ccw->bits |= CCW_HALT_ON_TERM; in mxs_dma_prep_dma_cyclic()
619 ccw->bits |= CCW_TERM_FLUSH; in mxs_dma_prep_dma_cyclic()
620 ccw->bits |= CCW_DEC_SEM; in mxs_dma_prep_dma_cyclic()
621 ccw->bits |= BF_CCW(direction == DMA_DEV_TO_MEM ? in mxs_dma_prep_dma_cyclic()
629 mxs_chan->desc_count = i; in mxs_dma_prep_dma_cyclic()
631 return &mxs_chan->desc; in mxs_dma_prep_dma_cyclic()
634 mxs_chan->status = DMA_ERROR; in mxs_dma_prep_dma_cyclic()
650 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_tx_status()
653 if (mxs_chan->status == DMA_IN_PROGRESS && in mxs_dma_tx_status()
654 mxs_chan->flags & MXS_DMA_SG_LOOP) { in mxs_dma_tx_status()
658 last_ccw = &mxs_chan->ccw[mxs_chan->desc_count - 1]; in mxs_dma_tx_status()
659 residue = last_ccw->xfer_bytes + last_ccw->bufaddr; in mxs_dma_tx_status()
661 bar = readl(mxs_dma->base + in mxs_dma_tx_status()
662 HW_APBHX_CHn_BAR(mxs_dma, chan->chan_id)); in mxs_dma_tx_status()
663 residue -= bar; in mxs_dma_tx_status()
666 dma_set_tx_state(txstate, chan->completed_cookie, chan->cookie, in mxs_dma_tx_status()
669 return mxs_chan->status; in mxs_dma_tx_status()
676 ret = clk_prepare_enable(mxs_dma->clk); in mxs_dma_init()
680 ret = stmp_reset_block(mxs_dma->base); in mxs_dma_init()
684 /* enable apbh burst */ in mxs_dma_init()
687 mxs_dma->base + HW_APBHX_CTRL0 + STMP_OFFSET_REG_SET); in mxs_dma_init()
689 mxs_dma->base + HW_APBHX_CTRL0 + STMP_OFFSET_REG_SET); in mxs_dma_init()
694 mxs_dma->base + HW_APBHX_CTRL1 + STMP_OFFSET_REG_SET); in mxs_dma_init()
697 clk_disable_unprepare(mxs_dma->clk); in mxs_dma_init()
709 struct mxs_dma_engine *mxs_dma = mxs_chan->mxs_dma; in mxs_dma_filter_fn()
712 if (chan->chan_id != param->chan_id) in mxs_dma_filter_fn()
715 chan_irq = platform_get_irq(mxs_dma->pdev, param->chan_id); in mxs_dma_filter_fn()
719 mxs_chan->chan_irq = chan_irq; in mxs_dma_filter_fn()
727 struct mxs_dma_engine *mxs_dma = ofdma->of_dma_data; in mxs_dma_xlate()
728 dma_cap_mask_t mask = mxs_dma->dma_device.cap_mask; in mxs_dma_xlate()
731 if (dma_spec->args_count != 1) in mxs_dma_xlate()
734 param.chan_id = dma_spec->args[0]; in mxs_dma_xlate()
736 if (param.chan_id >= mxs_dma->nr_channels) in mxs_dma_xlate()
740 ofdma->of_node); in mxs_dma_xlate()
745 struct device_node *np = pdev->dev.of_node; in mxs_dma_probe()
750 mxs_dma = devm_kzalloc(&pdev->dev, sizeof(*mxs_dma), GFP_KERNEL); in mxs_dma_probe()
752 return -ENOMEM; in mxs_dma_probe()
754 ret = of_property_read_u32(np, "dma-channels", &mxs_dma->nr_channels); in mxs_dma_probe()
756 dev_err(&pdev->dev, "failed to read dma-channels\n"); in mxs_dma_probe()
760 dma_type = (struct mxs_dma_type *)of_device_get_match_data(&pdev->dev); in mxs_dma_probe()
761 mxs_dma->type = dma_type->type; in mxs_dma_probe()
762 mxs_dma->dev_id = dma_type->id; in mxs_dma_probe()
764 mxs_dma->base = devm_platform_ioremap_resource(pdev, 0); in mxs_dma_probe()
765 if (IS_ERR(mxs_dma->base)) in mxs_dma_probe()
766 return PTR_ERR(mxs_dma->base); in mxs_dma_probe()
768 mxs_dma->clk = devm_clk_get(&pdev->dev, NULL); in mxs_dma_probe()
769 if (IS_ERR(mxs_dma->clk)) in mxs_dma_probe()
770 return PTR_ERR(mxs_dma->clk); in mxs_dma_probe()
772 dma_cap_set(DMA_SLAVE, mxs_dma->dma_device.cap_mask); in mxs_dma_probe()
773 dma_cap_set(DMA_CYCLIC, mxs_dma->dma_device.cap_mask); in mxs_dma_probe()
775 INIT_LIST_HEAD(&mxs_dma->dma_device.channels); in mxs_dma_probe()
779 struct mxs_dma_chan *mxs_chan = &mxs_dma->mxs_chans[i]; in mxs_dma_probe()
781 mxs_chan->mxs_dma = mxs_dma; in mxs_dma_probe()
782 mxs_chan->chan.device = &mxs_dma->dma_device; in mxs_dma_probe()
783 dma_cookie_init(&mxs_chan->chan); in mxs_dma_probe()
785 tasklet_setup(&mxs_chan->tasklet, mxs_dma_tasklet); in mxs_dma_probe()
789 list_add_tail(&mxs_chan->chan.device_node, in mxs_dma_probe()
790 &mxs_dma->dma_device.channels); in mxs_dma_probe()
797 mxs_dma->pdev = pdev; in mxs_dma_probe()
798 mxs_dma->dma_device.dev = &pdev->dev; in mxs_dma_probe()
801 dma_set_max_seg_size(mxs_dma->dma_device.dev, MAX_XFER_BYTES); in mxs_dma_probe()
803 mxs_dma->dma_device.device_alloc_chan_resources = mxs_dma_alloc_chan_resources; in mxs_dma_probe()
804 mxs_dma->dma_device.device_free_chan_resources = mxs_dma_free_chan_resources; in mxs_dma_probe()
805 mxs_dma->dma_device.device_tx_status = mxs_dma_tx_status; in mxs_dma_probe()
806 mxs_dma->dma_device.device_prep_slave_sg = mxs_dma_prep_slave_sg; in mxs_dma_probe()
807 mxs_dma->dma_device.device_prep_dma_cyclic = mxs_dma_prep_dma_cyclic; in mxs_dma_probe()
808 mxs_dma->dma_device.device_pause = mxs_dma_pause_chan; in mxs_dma_probe()
809 mxs_dma->dma_device.device_resume = mxs_dma_resume_chan; in mxs_dma_probe()
810 mxs_dma->dma_device.device_terminate_all = mxs_dma_terminate_all; in mxs_dma_probe()
811 mxs_dma->dma_device.src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in mxs_dma_probe()
812 mxs_dma->dma_device.dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); in mxs_dma_probe()
813 mxs_dma->dma_device.directions = BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_DEV); in mxs_dma_probe()
814 mxs_dma->dma_device.residue_granularity = DMA_RESIDUE_GRANULARITY_BURST; in mxs_dma_probe()
815 mxs_dma->dma_device.device_issue_pending = mxs_dma_enable_chan; in mxs_dma_probe()
817 ret = dmaenginem_async_device_register(&mxs_dma->dma_device); in mxs_dma_probe()
819 dev_err(mxs_dma->dma_device.dev, "unable to register\n"); in mxs_dma_probe()
825 dev_err(mxs_dma->dma_device.dev, in mxs_dma_probe()
829 dev_info(mxs_dma->dma_device.dev, "initialized\n"); in mxs_dma_probe()
836 .name = "mxs-dma",