/*- * Copyright (c) 2016-2018 Ruslan Bukin * All rights reserved. * * This software was developed by SRI International and the University of * Cambridge Computer Laboratory under DARPA/AFRL contract FA8750-10-C-0237 * ("CTSRD"), as part of the DARPA CRASH research programme. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ /* Altera mSGDMA driver. */ #include #include "opt_platform.h" #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #ifdef FDT #include #include #include #endif #include #include "xdma_if.h" #include "opt_altera_msgdma.h" #include #define MSGDMA_DEBUG #undef MSGDMA_DEBUG #ifdef MSGDMA_DEBUG #define dprintf(fmt, ...) printf(fmt, ##__VA_ARGS__) #else #define dprintf(fmt, ...) #endif #define MSGDMA_NCHANNELS 1 struct msgdma_channel { struct msgdma_softc *sc; struct mtx mtx; xdma_channel_t *xchan; struct proc *p; int used; int index; int idx_head; int idx_tail; struct msgdma_desc **descs; bus_dma_segment_t *descs_phys; uint32_t descs_num; bus_dma_tag_t dma_tag; bus_dmamap_t *dma_map; uint32_t map_descr; uint8_t map_err; uint32_t descs_used_count; }; struct msgdma_softc { device_t dev; struct resource *res[3]; bus_space_tag_t bst; bus_space_handle_t bsh; bus_space_tag_t bst_d; bus_space_handle_t bsh_d; void *ih; struct msgdma_desc desc; struct msgdma_channel channels[MSGDMA_NCHANNELS]; }; static struct resource_spec msgdma_spec[] = { { SYS_RES_MEMORY, 0, RF_ACTIVE }, { SYS_RES_MEMORY, 1, RF_ACTIVE }, { SYS_RES_IRQ, 0, RF_ACTIVE }, { -1, 0 } }; #define HWTYPE_NONE 0 #define HWTYPE_STD 1 static struct ofw_compat_data compat_data[] = { { "altr,msgdma-16.0", HWTYPE_STD }, { "altr,msgdma-1.0", HWTYPE_STD }, { NULL, HWTYPE_NONE }, }; static int msgdma_probe(device_t dev); static int msgdma_attach(device_t dev); static int msgdma_detach(device_t dev); static inline uint32_t msgdma_next_desc(struct msgdma_channel *chan, uint32_t curidx) { return ((curidx + 1) % chan->descs_num); } static void msgdma_intr(void *arg) { xdma_transfer_status_t status; struct xdma_transfer_status st; struct msgdma_desc *desc; struct msgdma_channel *chan; struct xdma_channel *xchan; struct msgdma_softc *sc; uint32_t tot_copied; sc = arg; chan = &sc->channels[0]; xchan = chan->xchan; dprintf("%s(%d): status 0x%08x next_descr 0x%08x, control 0x%08x\n", __func__, device_get_unit(sc->dev), READ4_DESC(sc, PF_STATUS), READ4_DESC(sc, PF_NEXT_LO), READ4_DESC(sc, PF_CONTROL)); tot_copied = 0; while (chan->idx_tail != chan->idx_head) { dprintf("%s: idx_tail %d idx_head %d\n", __func__, chan->idx_tail, chan->idx_head); bus_dmamap_sync(chan->dma_tag, chan->dma_map[chan->idx_tail], BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE); desc = chan->descs[chan->idx_tail]; if ((le32toh(desc->control) & CONTROL_OWN) != 0) { break; } tot_copied += le32toh(desc->transferred); st.error = 0; st.transferred = le32toh(desc->transferred); xchan_seg_done(xchan, &st); chan->idx_tail = msgdma_next_desc(chan, chan->idx_tail); atomic_subtract_int(&chan->descs_used_count, 1); } WRITE4_DESC(sc, PF_STATUS, PF_STATUS_IRQ); /* Finish operation */ status.error = 0; status.transferred = tot_copied; xdma_callback(chan->xchan, &status); } static int msgdma_reset(struct msgdma_softc *sc) { int timeout; dprintf("%s: read status: %x\n", __func__, READ4(sc, 0x00)); dprintf("%s: read control: %x\n", __func__, READ4(sc, 0x04)); dprintf("%s: read 1: %x\n", __func__, READ4(sc, 0x08)); dprintf("%s: read 2: %x\n", __func__, READ4(sc, 0x0C)); WRITE4(sc, DMA_CONTROL, CONTROL_RESET); timeout = 100; do { if ((READ4(sc, DMA_STATUS) & STATUS_RESETTING) == 0) break; } while (timeout--); dprintf("timeout %d\n", timeout); if (timeout == 0) return (-1); dprintf("%s: read control after reset: %x\n", __func__, READ4(sc, DMA_CONTROL)); return (0); } static int msgdma_probe(device_t dev) { int hwtype; if (!ofw_bus_status_okay(dev)) return (ENXIO); hwtype = ofw_bus_search_compatible(dev, compat_data)->ocd_data; if (hwtype == HWTYPE_NONE) return (ENXIO); device_set_desc(dev, "Altera mSGDMA"); return (BUS_PROBE_DEFAULT); } static int msgdma_attach(device_t dev) { struct msgdma_softc *sc; phandle_t xref, node; int err; sc = device_get_softc(dev); sc->dev = dev; if (bus_alloc_resources(dev, msgdma_spec, sc->res)) { device_printf(dev, "could not allocate resources for device\n"); return (ENXIO); } /* CSR memory interface */ sc->bst = rman_get_bustag(sc->res[0]); sc->bsh = rman_get_bushandle(sc->res[0]); /* Descriptor memory interface */ sc->bst_d = rman_get_bustag(sc->res[1]); sc->bsh_d = rman_get_bushandle(sc->res[1]); /* Setup interrupt handler */ err = bus_setup_intr(dev, sc->res[2], INTR_TYPE_MISC | INTR_MPSAFE, NULL, msgdma_intr, sc, &sc->ih); if (err) { device_printf(dev, "Unable to alloc interrupt resource.\n"); return (ENXIO); } node = ofw_bus_get_node(dev); xref = OF_xref_from_node(node); OF_device_register_xref(xref, dev); if (msgdma_reset(sc) != 0) return (-1); WRITE4(sc, DMA_CONTROL, CONTROL_GIEM); return (0); } static int msgdma_detach(device_t dev) { struct msgdma_softc *sc; sc = device_get_softc(dev); return (0); } static void msgdma_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int err) { struct msgdma_channel *chan; chan = (struct msgdma_channel *)arg; KASSERT(chan != NULL, ("xchan is NULL")); if (err) { chan->map_err = 1; return; } chan->descs_phys[chan->map_descr].ds_addr = segs[0].ds_addr; chan->descs_phys[chan->map_descr].ds_len = segs[0].ds_len; dprintf("map desc %d: descs phys %lx len %ld\n", chan->map_descr, segs[0].ds_addr, segs[0].ds_len); } static int msgdma_desc_free(struct msgdma_softc *sc, struct msgdma_channel *chan) { struct msgdma_desc *desc; int nsegments; int i; nsegments = chan->descs_num; for (i = 0; i < nsegments; i++) { desc = chan->descs[i]; bus_dmamap_unload(chan->dma_tag, chan->dma_map[i]); bus_dmamem_free(chan->dma_tag, desc, chan->dma_map[i]); } bus_dma_tag_destroy(chan->dma_tag); free(chan->descs, M_DEVBUF); free(chan->dma_map, M_DEVBUF); free(chan->descs_phys, M_DEVBUF); return (0); } static int msgdma_desc_alloc(struct msgdma_softc *sc, struct msgdma_channel *chan, uint32_t desc_size, uint32_t align) { int nsegments; int err; int i; nsegments = chan->descs_num; dprintf("%s: nseg %d\n", __func__, nsegments); err = bus_dma_tag_create( bus_get_dma_tag(sc->dev), align, 0, /* alignment, boundary */ BUS_SPACE_MAXADDR_32BIT, /* lowaddr */ BUS_SPACE_MAXADDR, /* highaddr */ NULL, NULL, /* filter, filterarg */ desc_size, 1, /* maxsize, nsegments*/ desc_size, 0, /* maxsegsize, flags */ NULL, NULL, /* lockfunc, lockarg */ &chan->dma_tag); if (err) { device_printf(sc->dev, "%s: Can't create bus_dma tag.\n", __func__); return (-1); } /* Descriptors. */ chan->descs = malloc(nsegments * sizeof(struct msgdma_desc *), M_DEVBUF, (M_WAITOK | M_ZERO)); if (chan->descs == NULL) { device_printf(sc->dev, "%s: Can't allocate memory.\n", __func__); return (-1); } chan->dma_map = malloc(nsegments * sizeof(bus_dmamap_t), M_DEVBUF, (M_WAITOK | M_ZERO)); chan->descs_phys = malloc(nsegments * sizeof(bus_dma_segment_t), M_DEVBUF, (M_WAITOK | M_ZERO)); /* Allocate bus_dma memory for each descriptor. */ for (i = 0; i < nsegments; i++) { err = bus_dmamem_alloc(chan->dma_tag, (void **)&chan->descs[i], BUS_DMA_WAITOK | BUS_DMA_ZERO, &chan->dma_map[i]); if (err) { device_printf(sc->dev, "%s: Can't allocate memory for descriptors.\n", __func__); return (-1); } chan->map_err = 0; chan->map_descr = i; err = bus_dmamap_load(chan->dma_tag, chan->dma_map[i], chan->descs[i], desc_size, msgdma_dmamap_cb, chan, BUS_DMA_WAITOK); if (err) { device_printf(sc->dev, "%s: Can't load DMA map.\n", __func__); return (-1); } if (chan->map_err != 0) { device_printf(sc->dev, "%s: Can't load DMA map.\n", __func__); return (-1); } } return (0); } static int msgdma_channel_alloc(device_t dev, struct xdma_channel *xchan) { struct msgdma_channel *chan; struct msgdma_softc *sc; int i; sc = device_get_softc(dev); for (i = 0; i < MSGDMA_NCHANNELS; i++) { chan = &sc->channels[i]; if (chan->used == 0) { chan->xchan = xchan; xchan->chan = (void *)chan; if ((xchan->caps & XCHAN_CAP_IOMMU) == 0) xchan->caps |= XCHAN_CAP_BUSDMA; chan->index = i; chan->sc = sc; chan->used = 1; chan->idx_head = 0; chan->idx_tail = 0; chan->descs_used_count = 0; chan->descs_num = 1024; return (0); } } return (-1); } static int msgdma_channel_free(device_t dev, struct xdma_channel *xchan) { struct msgdma_channel *chan; struct msgdma_softc *sc; sc = device_get_softc(dev); chan = (struct msgdma_channel *)xchan->chan; msgdma_desc_free(sc, chan); chan->used = 0; return (0); } static int msgdma_channel_capacity(device_t dev, xdma_channel_t *xchan, uint32_t *capacity) { struct msgdma_channel *chan; uint32_t c; chan = (struct msgdma_channel *)xchan->chan; /* At least one descriptor must be left empty. */ c = (chan->descs_num - chan->descs_used_count - 1); *capacity = c; return (0); } static int msgdma_channel_submit_sg(device_t dev, struct xdma_channel *xchan, struct xdma_sglist *sg, uint32_t sg_n) { struct msgdma_channel *chan; struct msgdma_desc *desc; struct msgdma_softc *sc; bus_addr_t src_addr_lo; bus_addr_t dst_addr_lo; uint32_t len; uint32_t tmp; int i; sc = device_get_softc(dev); chan = (struct msgdma_channel *)xchan->chan; for (i = 0; i < sg_n; i++) { src_addr_lo = sg[i].src_addr; dst_addr_lo = sg[i].dst_addr; len = (uint32_t)sg[i].len; dprintf("%s: src %x dst %x len %d\n", __func__, src_addr_lo, dst_addr_lo, len); desc = chan->descs[chan->idx_head]; #if defined(ALTERA_MSGDMA_DESC_EXT) || defined(ALTERA_MSGDMA_DESC_PF_EXT) desc->read_hi = htole32(src_addr_lo >> 32); desc->write_hi = htole32(dst_addr_lo >> 32); #endif desc->read_lo = htole32(src_addr_lo); desc->write_lo = htole32(dst_addr_lo); desc->length = htole32(len); desc->transferred = 0; desc->status = 0; desc->reserved = 0; desc->control = 0; if (sg[i].direction == XDMA_MEM_TO_DEV) { if (sg[i].first == 1) { desc->control |= htole32(CONTROL_GEN_SOP); } if (sg[i].last == 1) { desc->control |= htole32(CONTROL_GEN_EOP); desc->control |= htole32(CONTROL_TC_IRQ_EN | CONTROL_ET_IRQ_EN | CONTROL_ERR_M); } } else { desc->control |= htole32(CONTROL_END_ON_EOP | (1 << 13)); desc->control |= htole32(CONTROL_TC_IRQ_EN | CONTROL_ET_IRQ_EN | CONTROL_ERR_M); } tmp = chan->idx_head; atomic_add_int(&chan->descs_used_count, 1); chan->idx_head = msgdma_next_desc(chan, chan->idx_head); desc->control |= htole32(CONTROL_OWN | CONTROL_GO); bus_dmamap_sync(chan->dma_tag, chan->dma_map[tmp], BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE); } return (0); } static int msgdma_channel_prep_sg(device_t dev, struct xdma_channel *xchan) { struct msgdma_channel *chan; struct msgdma_desc *desc; struct msgdma_softc *sc; uint32_t addr; uint32_t reg; int ret; int i; sc = device_get_softc(dev); dprintf("%s(%d)\n", __func__, device_get_unit(dev)); chan = (struct msgdma_channel *)xchan->chan; ret = msgdma_desc_alloc(sc, chan, sizeof(struct msgdma_desc), 16); if (ret != 0) { device_printf(sc->dev, "%s: Can't allocate descriptors.\n", __func__); return (-1); } for (i = 0; i < chan->descs_num; i++) { desc = chan->descs[i]; if (i == (chan->descs_num - 1)) { desc->next = htole32(chan->descs_phys[0].ds_addr); } else { desc->next = htole32(chan->descs_phys[i+1].ds_addr); } dprintf("%s(%d): desc %d vaddr %lx next paddr %x\n", __func__, device_get_unit(dev), i, (uint64_t)desc, le32toh(desc->next)); } addr = chan->descs_phys[0].ds_addr; WRITE4_DESC(sc, PF_NEXT_LO, addr); WRITE4_DESC(sc, PF_NEXT_HI, 0); WRITE4_DESC(sc, PF_POLL_FREQ, 1000); reg = (PF_CONTROL_GIEM | PF_CONTROL_DESC_POLL_EN); reg |= PF_CONTROL_RUN; WRITE4_DESC(sc, PF_CONTROL, reg); return (0); } static int msgdma_channel_control(device_t dev, xdma_channel_t *xchan, int cmd) { struct msgdma_channel *chan; struct msgdma_softc *sc; sc = device_get_softc(dev); chan = (struct msgdma_channel *)xchan->chan; switch (cmd) { case XDMA_CMD_BEGIN: case XDMA_CMD_TERMINATE: case XDMA_CMD_PAUSE: /* TODO: implement me */ return (-1); } return (0); } #ifdef FDT static int msgdma_ofw_md_data(device_t dev, pcell_t *cells, int ncells, void **ptr) { return (0); } #endif static device_method_t msgdma_methods[] = { /* Device interface */ DEVMETHOD(device_probe, msgdma_probe), DEVMETHOD(device_attach, msgdma_attach), DEVMETHOD(device_detach, msgdma_detach), /* xDMA Interface */ DEVMETHOD(xdma_channel_alloc, msgdma_channel_alloc), DEVMETHOD(xdma_channel_free, msgdma_channel_free), DEVMETHOD(xdma_channel_control, msgdma_channel_control), /* xDMA SG Interface */ DEVMETHOD(xdma_channel_capacity, msgdma_channel_capacity), DEVMETHOD(xdma_channel_prep_sg, msgdma_channel_prep_sg), DEVMETHOD(xdma_channel_submit_sg, msgdma_channel_submit_sg), #ifdef FDT DEVMETHOD(xdma_ofw_md_data, msgdma_ofw_md_data), #endif DEVMETHOD_END }; static driver_t msgdma_driver = { "msgdma", msgdma_methods, sizeof(struct msgdma_softc), }; EARLY_DRIVER_MODULE(msgdma, simplebus, msgdma_driver, 0, 0, BUS_PASS_INTERRUPT + BUS_PASS_ORDER_LATE);