1 /* 2 * Copyright (c) 2013 Daisuke Aoyama <aoyama@peach.ne.jp> 3 * Copyright (c) 2013 Oleksandr Tymoshenko <gonzo@bluezbox.com> 4 * 5 * Redistribution and use in source and binary forms, with or without 6 * modification, are permitted provided that the following conditions 7 * are met: 8 * 1. Redistributions of source code must retain the above copyright 9 * notice, this list of conditions and the following disclaimer. 10 * 2. Redistributions in binary form must reproduce the above copyright 11 * notice, this list of conditions and the following disclaimer in the 12 * documentation and/or other materials provided with the distribution. 13 * 14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 24 * SUCH DAMAGE. 25 * 26 */ 27 28 #include <sys/cdefs.h> 29 __FBSDID("$FreeBSD$"); 30 31 #include <sys/param.h> 32 #include <sys/systm.h> 33 #include <sys/bus.h> 34 #include <sys/kernel.h> 35 #include <sys/lock.h> 36 #include <sys/malloc.h> 37 #include <sys/module.h> 38 #include <sys/mutex.h> 39 #include <sys/queue.h> 40 #include <sys/resource.h> 41 #include <sys/rman.h> 42 43 #include <dev/fdt/fdt_common.h> 44 #include <dev/ofw/openfirm.h> 45 #include <dev/ofw/ofw_bus.h> 46 #include <dev/ofw/ofw_bus_subr.h> 47 48 #include <vm/vm.h> 49 #include <vm/pmap.h> 50 #include <machine/bus.h> 51 #include <machine/cpu.h> 52 #include <machine/cpufunc.h> 53 54 #include "bcm2835_dma.h" 55 #include "bcm2835_vcbus.h" 56 57 #define MAX_REG 9 58 59 /* private flags */ 60 #define BCM_DMA_CH_USED 0x00000001 61 #define BCM_DMA_CH_FREE 0x40000000 62 #define BCM_DMA_CH_UNMAP 0x80000000 63 64 /* Register Map (4.2.1.2) */ 65 #define BCM_DMA_CS(n) (0x100*(n) + 0x00) 66 #define CS_ACTIVE (1 << 0) 67 #define CS_END (1 << 1) 68 #define CS_INT (1 << 2) 69 #define CS_DREQ (1 << 3) 70 #define CS_ISPAUSED (1 << 4) 71 #define CS_ISHELD (1 << 5) 72 #define CS_ISWAIT (1 << 6) 73 #define CS_ERR (1 << 8) 74 #define CS_WAITWRT (1 << 28) 75 #define CS_DISDBG (1 << 29) 76 #define CS_ABORT (1 << 30) 77 #define CS_RESET (1U << 31) 78 #define BCM_DMA_CBADDR(n) (0x100*(n) + 0x04) 79 #define BCM_DMA_INFO(n) (0x100*(n) + 0x08) 80 #define INFO_INT_EN (1 << 0) 81 #define INFO_TDMODE (1 << 1) 82 #define INFO_WAIT_RESP (1 << 3) 83 #define INFO_D_INC (1 << 4) 84 #define INFO_D_WIDTH (1 << 5) 85 #define INFO_D_DREQ (1 << 6) 86 #define INFO_S_INC (1 << 8) 87 #define INFO_S_WIDTH (1 << 9) 88 #define INFO_S_DREQ (1 << 10) 89 #define INFO_WAITS_SHIFT (21) 90 #define INFO_PERMAP_SHIFT (16) 91 #define INFO_PERMAP_MASK (0x1f << INFO_PERMAP_SHIFT) 92 93 #define BCM_DMA_SRC(n) (0x100*(n) + 0x0C) 94 #define BCM_DMA_DST(n) (0x100*(n) + 0x10) 95 #define BCM_DMA_LEN(n) (0x100*(n) + 0x14) 96 #define BCM_DMA_STRIDE(n) (0x100*(n) + 0x18) 97 #define BCM_DMA_CBNEXT(n) (0x100*(n) + 0x1C) 98 #define BCM_DMA_DEBUG(n) (0x100*(n) + 0x20) 99 #define DEBUG_ERROR_MASK (7) 100 101 #define BCM_DMA_INT_STATUS 0xfe0 102 #define BCM_DMA_ENABLE 0xff0 103 104 /* relative offset from BCM_VC_DMA0_BASE (p.39) */ 105 #define BCM_DMA_CH(n) (0x100*(n)) 106 107 /* DMA Control Block - 256bit aligned (p.40) */ 108 struct bcm_dma_cb { 109 uint32_t info; /* Transfer Information */ 110 uint32_t src; /* Source Address */ 111 uint32_t dst; /* Destination Address */ 112 uint32_t len; /* Transfer Length */ 113 uint32_t stride; /* 2D Mode Stride */ 114 uint32_t next; /* Next Control Block Address */ 115 uint32_t rsvd1; /* Reserved */ 116 uint32_t rsvd2; /* Reserved */ 117 }; 118 119 #ifdef DEBUG 120 static void bcm_dma_cb_dump(struct bcm_dma_cb *cb); 121 static void bcm_dma_reg_dump(int ch); 122 #endif 123 124 /* DMA channel private info */ 125 struct bcm_dma_ch { 126 int ch; 127 uint32_t flags; 128 struct bcm_dma_cb * cb; 129 uint32_t vc_cb; 130 bus_dmamap_t dma_map; 131 void (*intr_func)(int, void *); 132 void * intr_arg; 133 }; 134 135 struct bcm_dma_softc { 136 device_t sc_dev; 137 struct mtx sc_mtx; 138 struct resource * sc_mem; 139 struct resource * sc_irq[BCM_DMA_CH_MAX]; 140 void * sc_intrhand[BCM_DMA_CH_MAX]; 141 struct bcm_dma_ch sc_dma_ch[BCM_DMA_CH_MAX]; 142 bus_dma_tag_t sc_dma_tag; 143 }; 144 145 static struct bcm_dma_softc *bcm_dma_sc = NULL; 146 147 static void 148 bcm_dmamap_cb(void *arg, bus_dma_segment_t *segs, 149 int nseg, int err) 150 { 151 bus_addr_t *addr; 152 153 if (err) 154 return; 155 156 addr = (bus_addr_t*)arg; 157 *addr = PHYS_TO_VCBUS(segs[0].ds_addr); 158 } 159 160 static void 161 bcm_dma_reset(device_t dev, int ch) 162 { 163 struct bcm_dma_softc *sc = device_get_softc(dev); 164 struct bcm_dma_cb *cb; 165 uint32_t cs; 166 int count; 167 168 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 169 return; 170 171 cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch)); 172 173 if (cs & CS_ACTIVE) { 174 /* pause current task */ 175 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch), 0); 176 177 count = 1000; 178 do { 179 cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch)); 180 } while (!(cs & CS_ISPAUSED) && (count-- > 0)); 181 182 if (!(cs & CS_ISPAUSED)) { 183 device_printf(dev, 184 "Can't abort DMA transfer at channel %d\n", ch); 185 } 186 187 bus_write_4(sc->sc_mem, BCM_DMA_CBNEXT(ch), 0); 188 189 /* Complete everything, clear interrupt */ 190 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch), 191 CS_ABORT | CS_INT | CS_END| CS_ACTIVE); 192 } 193 194 /* clear control blocks */ 195 bus_write_4(sc->sc_mem, BCM_DMA_CBADDR(ch), 0); 196 bus_write_4(sc->sc_mem, BCM_DMA_CBNEXT(ch), 0); 197 198 /* Reset control block */ 199 cb = sc->sc_dma_ch[ch].cb; 200 bzero(cb, sizeof(*cb)); 201 cb->info = INFO_WAIT_RESP; 202 } 203 204 static int 205 bcm_dma_init(device_t dev) 206 { 207 struct bcm_dma_softc *sc = device_get_softc(dev); 208 uint32_t mask; 209 struct bcm_dma_ch *ch; 210 void *cb_virt; 211 vm_paddr_t cb_phys; 212 int err; 213 int i; 214 215 /* disable and clear interrupt status */ 216 bus_write_4(sc->sc_mem, BCM_DMA_ENABLE, 0); 217 bus_write_4(sc->sc_mem, BCM_DMA_INT_STATUS, 0); 218 219 /* Allocate DMA chunks control blocks */ 220 /* p.40 of spec - control block should be 32-bit aligned */ 221 err = bus_dma_tag_create(bus_get_dma_tag(dev), 222 1, 0, BUS_SPACE_MAXADDR_32BIT, 223 BUS_SPACE_MAXADDR, NULL, NULL, 224 sizeof(struct bcm_dma_cb), 1, 225 sizeof(struct bcm_dma_cb), 226 BUS_DMA_ALLOCNOW, NULL, NULL, 227 &sc->sc_dma_tag); 228 229 if (err) { 230 device_printf(dev, "failed allocate DMA tag"); 231 return (err); 232 } 233 234 /* setup initial settings */ 235 for (i = 0; i < BCM_DMA_CH_MAX; i++) { 236 ch = &sc->sc_dma_ch[i]; 237 238 err = bus_dmamem_alloc(sc->sc_dma_tag, &cb_virt, 239 BUS_DMA_WAITOK | BUS_DMA_COHERENT | BUS_DMA_ZERO, 240 &ch->dma_map); 241 if (err) { 242 device_printf(dev, "cannot allocate DMA memory\n"); 243 break; 244 } 245 246 /* 247 * Least alignment for busdma-allocated stuff is cache 248 * line size, so just make sure nothing stupid happend 249 * and we got properly aligned address 250 */ 251 if ((uintptr_t)cb_virt & 0x1f) { 252 device_printf(dev, 253 "DMA address is not 32-bytes aligned: %p\n", 254 (void*)cb_virt); 255 break; 256 } 257 258 err = bus_dmamap_load(sc->sc_dma_tag, ch->dma_map, cb_virt, 259 sizeof(struct bcm_dma_cb), bcm_dmamap_cb, &cb_phys, 260 BUS_DMA_WAITOK); 261 if (err) { 262 device_printf(dev, "cannot load DMA memory\n"); 263 break; 264 } 265 266 bzero(ch, sizeof(struct bcm_dma_ch)); 267 ch->ch = i; 268 ch->cb = cb_virt; 269 ch->vc_cb = cb_phys; 270 ch->intr_func = NULL; 271 ch->intr_arg = NULL; 272 ch->flags = BCM_DMA_CH_UNMAP; 273 274 ch->cb->info = INFO_WAIT_RESP; 275 276 /* reset DMA engine */ 277 bcm_dma_reset(dev, i); 278 } 279 280 /* now use DMA2/DMA3 only */ 281 sc->sc_dma_ch[2].flags = BCM_DMA_CH_FREE; 282 sc->sc_dma_ch[3].flags = BCM_DMA_CH_FREE; 283 284 /* enable DMAs */ 285 mask = 0; 286 287 for (i = 0; i < BCM_DMA_CH_MAX; i++) 288 if (sc->sc_dma_ch[i].flags & BCM_DMA_CH_FREE) 289 mask |= (1 << i); 290 291 bus_write_4(sc->sc_mem, BCM_DMA_ENABLE, mask); 292 293 return (0); 294 } 295 296 /* 297 * Allocate DMA channel for further use, returns channel # or 298 * BCM_DMA_CH_INVALID 299 */ 300 int 301 bcm_dma_allocate(int req_ch) 302 { 303 struct bcm_dma_softc *sc = bcm_dma_sc; 304 int ch = BCM_DMA_CH_INVALID; 305 int i; 306 307 if (req_ch >= BCM_DMA_CH_MAX) 308 return (BCM_DMA_CH_INVALID); 309 310 /* Auto(req_ch < 0) or CH specified */ 311 mtx_lock(&sc->sc_mtx); 312 313 if (req_ch < 0) { 314 for (i = 0; i < BCM_DMA_CH_MAX; i++) { 315 if (sc->sc_dma_ch[i].flags & BCM_DMA_CH_FREE) { 316 ch = i; 317 sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_FREE; 318 sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_USED; 319 break; 320 } 321 } 322 } 323 else { 324 if (sc->sc_dma_ch[req_ch].flags & BCM_DMA_CH_FREE) { 325 ch = req_ch; 326 sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_FREE; 327 sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_USED; 328 } 329 } 330 331 mtx_unlock(&sc->sc_mtx); 332 return (ch); 333 } 334 335 /* 336 * Frees allocated channel. Returns 0 on success, -1 otherwise 337 */ 338 int 339 bcm_dma_free(int ch) 340 { 341 struct bcm_dma_softc *sc = bcm_dma_sc; 342 343 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 344 return (-1); 345 346 mtx_lock(&sc->sc_mtx); 347 if (sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED) { 348 sc->sc_dma_ch[ch].flags |= BCM_DMA_CH_FREE; 349 sc->sc_dma_ch[ch].flags &= ~BCM_DMA_CH_USED; 350 sc->sc_dma_ch[ch].intr_func = NULL; 351 sc->sc_dma_ch[ch].intr_arg = NULL; 352 353 /* reset DMA engine */ 354 bcm_dma_reset(sc->sc_dev, ch); 355 } 356 357 mtx_unlock(&sc->sc_mtx); 358 return (0); 359 } 360 361 /* 362 * Assign handler function for channel interrupt 363 * Returns 0 on success, -1 otherwise 364 */ 365 int 366 bcm_dma_setup_intr(int ch, void (*func)(int, void *), void *arg) 367 { 368 struct bcm_dma_softc *sc = bcm_dma_sc; 369 struct bcm_dma_cb *cb; 370 371 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 372 return (-1); 373 374 if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED)) 375 return (-1); 376 377 sc->sc_dma_ch[ch].intr_func = func; 378 sc->sc_dma_ch[ch].intr_arg = arg; 379 cb = sc->sc_dma_ch[ch].cb; 380 cb->info |= INFO_INT_EN; 381 382 return (0); 383 } 384 385 /* 386 * Setup DMA source parameters 387 * ch - channel number 388 * dreq - hardware DREQ # or BCM_DMA_DREQ_NONE if 389 * source is physical memory 390 * inc_addr - BCM_DMA_INC_ADDR if source address 391 * should be increased after each access or 392 * BCM_DMA_SAME_ADDR if address should remain 393 * the same 394 * width - size of read operation, BCM_DMA_32BIT 395 * for 32bit bursts, BCM_DMA_128BIT for 128 bits 396 * 397 * Returns 0 on success, -1 otherwise 398 */ 399 int 400 bcm_dma_setup_src(int ch, int dreq, int inc_addr, int width) 401 { 402 struct bcm_dma_softc *sc = bcm_dma_sc; 403 uint32_t info; 404 405 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 406 return (-1); 407 408 if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED)) 409 return (-1); 410 411 info = sc->sc_dma_ch[ch].cb->info; 412 info &= ~INFO_PERMAP_MASK; 413 info |= (dreq << INFO_PERMAP_SHIFT) & INFO_PERMAP_MASK; 414 415 if (dreq) 416 info |= INFO_S_DREQ; 417 else 418 info &= ~INFO_S_DREQ; 419 420 if (width == BCM_DMA_128BIT) 421 info |= INFO_S_WIDTH; 422 else 423 info &= ~INFO_S_WIDTH; 424 425 if (inc_addr == BCM_DMA_INC_ADDR) 426 info |= INFO_S_INC; 427 else 428 info &= ~INFO_S_INC; 429 430 sc->sc_dma_ch[ch].cb->info = info; 431 432 return (0); 433 } 434 435 /* 436 * Setup DMA destination parameters 437 * ch - channel number 438 * dreq - hardware DREQ # or BCM_DMA_DREQ_NONE if 439 * destination is physical memory 440 * inc_addr - BCM_DMA_INC_ADDR if source address 441 * should be increased after each access or 442 * BCM_DMA_SAME_ADDR if address should remain 443 * the same 444 * width - size of write operation, BCM_DMA_32BIT 445 * for 32bit bursts, BCM_DMA_128BIT for 128 bits 446 * 447 * Returns 0 on success, -1 otherwise 448 */ 449 int 450 bcm_dma_setup_dst(int ch, int dreq, int inc_addr, int width) 451 { 452 struct bcm_dma_softc *sc = bcm_dma_sc; 453 uint32_t info; 454 455 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 456 return (-1); 457 458 if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED)) 459 return (-1); 460 461 info = sc->sc_dma_ch[ch].cb->info; 462 info &= ~INFO_PERMAP_MASK; 463 info |= (dreq << INFO_PERMAP_SHIFT) & INFO_PERMAP_MASK; 464 465 if (dreq) 466 info |= INFO_D_DREQ; 467 else 468 info &= ~INFO_D_DREQ; 469 470 if (width == BCM_DMA_128BIT) 471 info |= INFO_D_WIDTH; 472 else 473 info &= ~INFO_D_WIDTH; 474 475 if (inc_addr == BCM_DMA_INC_ADDR) 476 info |= INFO_D_INC; 477 else 478 info &= ~INFO_D_INC; 479 480 sc->sc_dma_ch[ch].cb->info = info; 481 482 return (0); 483 } 484 485 #ifdef DEBUG 486 void 487 bcm_dma_cb_dump(struct bcm_dma_cb *cb) 488 { 489 490 printf("DMA CB "); 491 printf("INFO: %8.8x ", cb->info); 492 printf("SRC: %8.8x ", cb->src); 493 printf("DST: %8.8x ", cb->dst); 494 printf("LEN: %8.8x ", cb->len); 495 printf("\n"); 496 printf("STRIDE: %8.8x ", cb->stride); 497 printf("NEXT: %8.8x ", cb->next); 498 printf("RSVD1: %8.8x ", cb->rsvd1); 499 printf("RSVD2: %8.8x ", cb->rsvd2); 500 printf("\n"); 501 } 502 503 void 504 bcm_dma_reg_dump(int ch) 505 { 506 struct bcm_dma_softc *sc = bcm_dma_sc; 507 int i; 508 uint32_t reg; 509 510 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 511 return; 512 513 printf("DMA%d: ", ch); 514 for (i = 0; i < MAX_REG; i++) { 515 reg = bus_read_4(sc->sc_mem, BCM_DMA_CH(ch) + i*4); 516 printf("%8.8x ", reg); 517 } 518 printf("\n"); 519 } 520 #endif 521 522 /* 523 * Start DMA transaction 524 * ch - channel number 525 * src, dst - source and destination address in 526 * ARM physical memory address space. 527 * len - amount of bytes to be transfered 528 * 529 * Returns 0 on success, -1 otherwise 530 */ 531 int 532 bcm_dma_start(int ch, vm_paddr_t src, vm_paddr_t dst, int len) 533 { 534 struct bcm_dma_softc *sc = bcm_dma_sc; 535 struct bcm_dma_cb *cb; 536 537 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 538 return (-1); 539 540 if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED)) 541 return (-1); 542 543 cb = sc->sc_dma_ch[ch].cb; 544 if (BCM2835_ARM_IS_IO(src)) 545 cb->src = IO_TO_VCBUS(src); 546 else 547 cb->src = PHYS_TO_VCBUS(src); 548 if (BCM2835_ARM_IS_IO(dst)) 549 cb->dst = IO_TO_VCBUS(dst); 550 else 551 cb->dst = PHYS_TO_VCBUS(dst); 552 cb->len = len; 553 554 bus_dmamap_sync(sc->sc_dma_tag, 555 sc->sc_dma_ch[ch].dma_map, BUS_DMASYNC_PREWRITE); 556 557 bus_write_4(sc->sc_mem, BCM_DMA_CBADDR(ch), 558 sc->sc_dma_ch[ch].vc_cb); 559 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch), CS_ACTIVE); 560 561 #ifdef DEBUG 562 bcm_dma_cb_dump(sc->sc_dma_ch[ch].cb); 563 bcm_dma_reg_dump(ch); 564 #endif 565 566 return (0); 567 } 568 569 /* 570 * Get length requested for DMA transaction 571 * ch - channel number 572 * 573 * Returns size of transaction, 0 if channel is invalid 574 */ 575 uint32_t 576 bcm_dma_length(int ch) 577 { 578 struct bcm_dma_softc *sc = bcm_dma_sc; 579 struct bcm_dma_cb *cb; 580 581 if (ch < 0 || ch >= BCM_DMA_CH_MAX) 582 return (0); 583 584 if (!(sc->sc_dma_ch[ch].flags & BCM_DMA_CH_USED)) 585 return (0); 586 587 cb = sc->sc_dma_ch[ch].cb; 588 589 return (cb->len); 590 } 591 592 static void 593 bcm_dma_intr(void *arg) 594 { 595 struct bcm_dma_softc *sc = bcm_dma_sc; 596 struct bcm_dma_ch *ch = (struct bcm_dma_ch *)arg; 597 uint32_t cs, debug; 598 599 /* my interrupt? */ 600 cs = bus_read_4(sc->sc_mem, BCM_DMA_CS(ch->ch)); 601 602 if (!(cs & (CS_INT | CS_ERR))) 603 return; 604 605 /* running? */ 606 if (!(ch->flags & BCM_DMA_CH_USED)) { 607 device_printf(sc->sc_dev, 608 "unused DMA intr CH=%d, CS=%x\n", ch->ch, cs); 609 return; 610 } 611 612 if (cs & CS_ERR) { 613 debug = bus_read_4(sc->sc_mem, BCM_DMA_DEBUG(ch->ch)); 614 device_printf(sc->sc_dev, "DMA error %d on CH%d\n", 615 debug & DEBUG_ERROR_MASK, ch->ch); 616 bus_write_4(sc->sc_mem, BCM_DMA_DEBUG(ch->ch), 617 debug & DEBUG_ERROR_MASK); 618 bcm_dma_reset(sc->sc_dev, ch->ch); 619 } 620 621 if (cs & CS_INT) { 622 /* acknowledge interrupt */ 623 bus_write_4(sc->sc_mem, BCM_DMA_CS(ch->ch), 624 CS_INT | CS_END); 625 626 /* Prepare for possible access to len field */ 627 bus_dmamap_sync(sc->sc_dma_tag, ch->dma_map, 628 BUS_DMASYNC_POSTWRITE); 629 630 /* save callback function and argument */ 631 if (ch->intr_func) 632 ch->intr_func(ch->ch, ch->intr_arg); 633 } 634 } 635 636 static int 637 bcm_dma_probe(device_t dev) 638 { 639 640 if (!ofw_bus_is_compatible(dev, "broadcom,bcm2835-dma")) 641 return (ENXIO); 642 643 device_set_desc(dev, "BCM2835 DMA Controller"); 644 return (BUS_PROBE_DEFAULT); 645 } 646 647 static int 648 bcm_dma_attach(device_t dev) 649 { 650 struct bcm_dma_softc *sc = device_get_softc(dev); 651 int rid, err = 0; 652 int i; 653 654 sc->sc_dev = dev; 655 656 if (bcm_dma_sc) 657 return (ENXIO); 658 659 for (i = 0; i < BCM_DMA_CH_MAX; i++) { 660 sc->sc_irq[i] = NULL; 661 sc->sc_intrhand[i] = NULL; 662 } 663 664 /* DMA0 - DMA14 */ 665 rid = 0; 666 sc->sc_mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid, RF_ACTIVE); 667 if (sc->sc_mem == NULL) { 668 device_printf(dev, "could not allocate memory resource\n"); 669 return (ENXIO); 670 } 671 672 /* IRQ DMA0 - DMA11 XXX NOT USE DMA12(spurious?) */ 673 for (rid = 0; rid < BCM_DMA_CH_MAX; rid++) { 674 sc->sc_irq[rid] = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid, 675 RF_ACTIVE); 676 if (sc->sc_irq[rid] == NULL) { 677 device_printf(dev, "cannot allocate interrupt\n"); 678 err = ENXIO; 679 goto fail; 680 } 681 if (bus_setup_intr(dev, sc->sc_irq[rid], INTR_TYPE_MISC | INTR_MPSAFE, 682 NULL, bcm_dma_intr, &sc->sc_dma_ch[rid], 683 &sc->sc_intrhand[rid])) { 684 device_printf(dev, "cannot setup interrupt handler\n"); 685 err = ENXIO; 686 goto fail; 687 } 688 } 689 690 mtx_init(&sc->sc_mtx, "bcmdma", "bcmdma", MTX_DEF); 691 bcm_dma_sc = sc; 692 693 err = bcm_dma_init(dev); 694 if (err) 695 goto fail; 696 697 return (err); 698 699 fail: 700 if (sc->sc_mem) 701 bus_release_resource(dev, SYS_RES_MEMORY, 0, sc->sc_mem); 702 703 for (i = 0; i < BCM_DMA_CH_MAX; i++) { 704 if (sc->sc_intrhand[i]) 705 bus_teardown_intr(dev, sc->sc_irq[i], sc->sc_intrhand[i]); 706 if (sc->sc_irq[i]) 707 bus_release_resource(dev, SYS_RES_IRQ, 0, sc->sc_irq[i]); 708 } 709 710 return (err); 711 } 712 713 static device_method_t bcm_dma_methods[] = { 714 DEVMETHOD(device_probe, bcm_dma_probe), 715 DEVMETHOD(device_attach, bcm_dma_attach), 716 { 0, 0 } 717 }; 718 719 static driver_t bcm_dma_driver = { 720 "bcm_dma", 721 bcm_dma_methods, 722 sizeof(struct bcm_dma_softc), 723 }; 724 725 static devclass_t bcm_dma_devclass; 726 727 DRIVER_MODULE(bcm_dma, simplebus, bcm_dma_driver, bcm_dma_devclass, 0, 0); 728 MODULE_VERSION(bcm_dma, 1); 729