1 /*- 2 * SPDX-License-Identifier: BSD-2-Clause 3 * 4 * Copyright (c) 2012 Oleksandr Tymoshenko <gonzo@freebsd.org> 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions and the following disclaimer. 12 * 2. Redistributions in binary form must reproduce the above copyright 13 * notice, this list of conditions and the following disclaimer in the 14 * documentation and/or other materials provided with the distribution. 15 * 16 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND 17 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 18 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 19 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE 20 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 21 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 22 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 23 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT 24 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY 25 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF 26 * SUCH DAMAGE. 27 * 28 */ 29 #include <sys/cdefs.h> 30 __FBSDID("$FreeBSD$"); 31 32 #include <sys/param.h> 33 #include <sys/systm.h> 34 #include <sys/bus.h> 35 #include <sys/conf.h> 36 #include <sys/kernel.h> 37 #include <sys/lock.h> 38 #include <sys/malloc.h> 39 #include <sys/module.h> 40 #include <sys/mutex.h> 41 #include <sys/rman.h> 42 #include <sys/sysctl.h> 43 #include <sys/taskqueue.h> 44 45 #include <machine/bus.h> 46 47 #include <dev/ofw/ofw_bus.h> 48 #include <dev/ofw/ofw_bus_subr.h> 49 50 #include <dev/mmc/bridge.h> 51 #include <dev/mmc/mmcreg.h> 52 #include <dev/mmc/mmc_fdt_helpers.h> 53 54 #include <dev/sdhci/sdhci.h> 55 56 #include "mmcbr_if.h" 57 #include "sdhci_if.h" 58 59 #include "opt_mmccam.h" 60 61 #include "bcm2835_dma.h" 62 #include <arm/broadcom/bcm2835/bcm2835_mbox_prop.h> 63 #ifdef NOTYET 64 #include <arm/broadcom/bcm2835/bcm2835_clkman.h> 65 #endif 66 #include <arm/broadcom/bcm2835/bcm2835_vcbus.h> 67 68 #define BCM2835_DEFAULT_SDHCI_FREQ 50 69 #define BCM2838_DEFAULT_SDHCI_FREQ 100 70 71 #define BCM_SDHCI_BUFFER_SIZE 512 72 /* 73 * NUM_DMA_SEGS is the number of DMA segments we want to accommodate on average. 74 * We add in a number of segments based on how much we may need to spill into 75 * another segment due to crossing page boundaries. e.g. up to PAGE_SIZE, an 76 * extra page is needed as we can cross a page boundary exactly once. 77 */ 78 #define NUM_DMA_SEGS 1 79 #define NUM_DMA_SPILL_SEGS \ 80 ((((NUM_DMA_SEGS * BCM_SDHCI_BUFFER_SIZE) - 1) / PAGE_SIZE) + 1) 81 #define ALLOCATED_DMA_SEGS (NUM_DMA_SEGS + NUM_DMA_SPILL_SEGS) 82 #define BCM_DMA_MAXSIZE (NUM_DMA_SEGS * BCM_SDHCI_BUFFER_SIZE) 83 84 #define BCM_SDHCI_SLOT_LEFT(slot) \ 85 ((slot)->curcmd->data->len - (slot)->offset) 86 87 #define BCM_SDHCI_SEGSZ_LEFT(slot) \ 88 min(BCM_DMA_MAXSIZE, \ 89 rounddown(BCM_SDHCI_SLOT_LEFT(slot), BCM_SDHCI_BUFFER_SIZE)) 90 91 #define DATA_PENDING_MASK (SDHCI_INT_DATA_AVAIL | SDHCI_INT_SPACE_AVAIL) 92 #define DATA_XFER_MASK (DATA_PENDING_MASK | SDHCI_INT_DATA_END) 93 94 #ifdef DEBUG 95 static int bcm2835_sdhci_debug = 0; 96 97 TUNABLE_INT("hw.bcm2835.sdhci.debug", &bcm2835_sdhci_debug); 98 SYSCTL_INT(_hw_sdhci, OID_AUTO, bcm2835_sdhci_debug, CTLFLAG_RWTUN, 99 &bcm2835_sdhci_debug, 0, "bcm2835 SDHCI debug level"); 100 101 #define dprintf(fmt, args...) \ 102 do { \ 103 if (bcm2835_sdhci_debug) \ 104 printf("%s: " fmt, __func__, ##args); \ 105 } while (0) 106 #else 107 #define dprintf(fmt, args...) 108 #endif 109 110 static int bcm2835_sdhci_hs = 1; 111 static int bcm2835_sdhci_pio_mode = 0; 112 113 struct bcm_mmc_conf { 114 int clock_id; 115 int clock_src; 116 int default_freq; 117 int quirks; 118 int emmc_dreq; 119 }; 120 121 struct bcm_mmc_conf bcm2835_sdhci_conf = { 122 .clock_id = BCM2835_MBOX_CLOCK_ID_EMMC, 123 .clock_src = -1, 124 .default_freq = BCM2835_DEFAULT_SDHCI_FREQ, 125 .quirks = SDHCI_QUIRK_DATA_TIMEOUT_USES_SDCLK | 126 SDHCI_QUIRK_BROKEN_TIMEOUT_VAL | SDHCI_QUIRK_DONT_SET_HISPD_BIT | 127 SDHCI_QUIRK_MISSING_CAPS, 128 .emmc_dreq = BCM_DMA_DREQ_EMMC, 129 }; 130 131 struct bcm_mmc_conf bcm2838_emmc2_conf = { 132 .clock_id = BCM2838_MBOX_CLOCK_ID_EMMC2, 133 .clock_src = -1, 134 .default_freq = BCM2838_DEFAULT_SDHCI_FREQ, 135 .quirks = 0, 136 .emmc_dreq = BCM_DMA_DREQ_NONE, 137 }; 138 139 static struct ofw_compat_data compat_data[] = { 140 {"broadcom,bcm2835-sdhci", (uintptr_t)&bcm2835_sdhci_conf}, 141 {"brcm,bcm2835-sdhci", (uintptr_t)&bcm2835_sdhci_conf}, 142 {"brcm,bcm2835-mmc", (uintptr_t)&bcm2835_sdhci_conf}, 143 {"brcm,bcm2711-emmc2", (uintptr_t)&bcm2838_emmc2_conf}, 144 {"brcm,bcm2838-emmc2", (uintptr_t)&bcm2838_emmc2_conf}, 145 {NULL, 0} 146 }; 147 148 TUNABLE_INT("hw.bcm2835.sdhci.hs", &bcm2835_sdhci_hs); 149 TUNABLE_INT("hw.bcm2835.sdhci.pio_mode", &bcm2835_sdhci_pio_mode); 150 151 struct bcm_sdhci_softc { 152 device_t sc_dev; 153 struct resource * sc_mem_res; 154 struct resource * sc_irq_res; 155 bus_space_tag_t sc_bst; 156 bus_space_handle_t sc_bsh; 157 void * sc_intrhand; 158 struct mmc_request * sc_req; 159 struct sdhci_slot sc_slot; 160 struct mmc_helper sc_mmc_helper; 161 int sc_dma_ch; 162 bus_dma_tag_t sc_dma_tag; 163 bus_dmamap_t sc_dma_map; 164 vm_paddr_t sc_sdhci_buffer_phys; 165 bus_addr_t dmamap_seg_addrs[ALLOCATED_DMA_SEGS]; 166 bus_size_t dmamap_seg_sizes[ALLOCATED_DMA_SEGS]; 167 int dmamap_seg_count; 168 int dmamap_seg_index; 169 int dmamap_status; 170 uint32_t blksz_and_count; 171 uint32_t cmd_and_mode; 172 bool need_update_blk; 173 #ifdef NOTYET 174 device_t clkman; 175 #endif 176 struct bcm_mmc_conf * conf; 177 }; 178 179 static int bcm_sdhci_probe(device_t); 180 static int bcm_sdhci_attach(device_t); 181 static int bcm_sdhci_detach(device_t); 182 static void bcm_sdhci_intr(void *); 183 184 static int bcm_sdhci_get_ro(device_t, device_t); 185 static void bcm_sdhci_dma_intr(int ch, void *arg); 186 static void bcm_sdhci_start_dma(struct sdhci_slot *slot); 187 188 static void 189 bcm_sdhci_dmacb(void *arg, bus_dma_segment_t *segs, int nseg, int err) 190 { 191 struct bcm_sdhci_softc *sc = arg; 192 int i; 193 194 /* Sanity check: we can only ever have one mapping at a time. */ 195 KASSERT(sc->dmamap_seg_count == 0, ("leaked DMA segment")); 196 sc->dmamap_status = err; 197 sc->dmamap_seg_count = nseg; 198 199 /* Note nseg is guaranteed to be zero if err is non-zero. */ 200 for (i = 0; i < nseg; i++) { 201 sc->dmamap_seg_addrs[i] = segs[i].ds_addr; 202 sc->dmamap_seg_sizes[i] = segs[i].ds_len; 203 } 204 } 205 206 static int 207 bcm_sdhci_probe(device_t dev) 208 { 209 210 if (!ofw_bus_status_okay(dev)) 211 return (ENXIO); 212 213 if (ofw_bus_search_compatible(dev, compat_data)->ocd_data == 0) 214 return (ENXIO); 215 216 device_set_desc(dev, "Broadcom 2708 SDHCI controller"); 217 218 return (BUS_PROBE_DEFAULT); 219 } 220 221 static int 222 bcm_sdhci_attach(device_t dev) 223 { 224 struct bcm_sdhci_softc *sc = device_get_softc(dev); 225 int rid, err; 226 phandle_t node; 227 pcell_t cell; 228 u_int default_freq; 229 230 sc->sc_dev = dev; 231 sc->sc_req = NULL; 232 233 sc->conf = (struct bcm_mmc_conf *)ofw_bus_search_compatible(dev, 234 compat_data)->ocd_data; 235 if (sc->conf == 0) 236 return (ENXIO); 237 238 err = bcm2835_mbox_set_power_state(BCM2835_MBOX_POWER_ID_EMMC, TRUE); 239 if (err != 0) { 240 if (bootverbose) 241 device_printf(dev, "Unable to enable the power\n"); 242 return (err); 243 } 244 245 default_freq = 0; 246 err = bcm2835_mbox_get_clock_rate(sc->conf->clock_id, &default_freq); 247 if (err == 0) { 248 /* Convert to MHz */ 249 default_freq /= 1000000; 250 } 251 if (default_freq == 0) { 252 node = ofw_bus_get_node(sc->sc_dev); 253 if ((OF_getencprop(node, "clock-frequency", &cell, 254 sizeof(cell))) > 0) 255 default_freq = cell / 1000000; 256 } 257 if (default_freq == 0) 258 default_freq = sc->conf->default_freq; 259 260 if (bootverbose) 261 device_printf(dev, "SDHCI frequency: %dMHz\n", default_freq); 262 #ifdef NOTYET 263 if (sc->conf->clock_src > 0) { 264 uint32_t f; 265 sc->clkman = devclass_get_device( 266 devclass_find("bcm2835_clkman"), 0); 267 if (sc->clkman == NULL) { 268 device_printf(dev, "cannot find Clock Manager\n"); 269 return (ENXIO); 270 } 271 272 f = bcm2835_clkman_set_frequency(sc->clkman, 273 sc->conf->clock_src, default_freq); 274 if (f == 0) 275 return (EINVAL); 276 277 if (bootverbose) 278 device_printf(dev, "Clock source frequency: %dMHz\n", 279 f); 280 } 281 #endif 282 283 rid = 0; 284 sc->sc_mem_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid, 285 RF_ACTIVE); 286 if (!sc->sc_mem_res) { 287 device_printf(dev, "cannot allocate memory window\n"); 288 err = ENXIO; 289 goto fail; 290 } 291 292 sc->sc_bst = rman_get_bustag(sc->sc_mem_res); 293 sc->sc_bsh = rman_get_bushandle(sc->sc_mem_res); 294 295 rid = 0; 296 sc->sc_irq_res = bus_alloc_resource_any(dev, SYS_RES_IRQ, &rid, 297 RF_ACTIVE | RF_SHAREABLE); 298 if (!sc->sc_irq_res) { 299 device_printf(dev, "cannot allocate interrupt\n"); 300 err = ENXIO; 301 goto fail; 302 } 303 304 if (bus_setup_intr(dev, sc->sc_irq_res, INTR_TYPE_BIO | INTR_MPSAFE, 305 NULL, bcm_sdhci_intr, sc, &sc->sc_intrhand)) { 306 device_printf(dev, "cannot setup interrupt handler\n"); 307 err = ENXIO; 308 goto fail; 309 } 310 311 if (!bcm2835_sdhci_pio_mode) 312 sc->sc_slot.opt = SDHCI_PLATFORM_TRANSFER; 313 314 sc->sc_slot.caps = SDHCI_CAN_VDD_330 | SDHCI_CAN_VDD_180; 315 if (bcm2835_sdhci_hs) 316 sc->sc_slot.caps |= SDHCI_CAN_DO_HISPD; 317 sc->sc_slot.caps |= (default_freq << SDHCI_CLOCK_BASE_SHIFT); 318 sc->sc_slot.quirks = sc->conf->quirks; 319 320 sdhci_init_slot(dev, &sc->sc_slot, 0); 321 mmc_fdt_parse(dev, 0, &sc->sc_mmc_helper, &sc->sc_slot.host); 322 323 sc->sc_dma_ch = bcm_dma_allocate(BCM_DMA_CH_ANY); 324 if (sc->sc_dma_ch == BCM_DMA_CH_INVALID) 325 goto fail; 326 327 err = bcm_dma_setup_intr(sc->sc_dma_ch, bcm_sdhci_dma_intr, sc); 328 if (err != 0) { 329 device_printf(dev, 330 "cannot setup dma interrupt handler\n"); 331 err = ENXIO; 332 goto fail; 333 } 334 335 /* Allocate bus_dma resources. */ 336 err = bus_dma_tag_create(bus_get_dma_tag(dev), 337 1, 0, bcm283x_dmabus_peripheral_lowaddr(), 338 BUS_SPACE_MAXADDR, NULL, NULL, 339 BCM_DMA_MAXSIZE, ALLOCATED_DMA_SEGS, BCM_SDHCI_BUFFER_SIZE, 340 BUS_DMA_ALLOCNOW, NULL, NULL, 341 &sc->sc_dma_tag); 342 343 if (err) { 344 device_printf(dev, "failed allocate DMA tag"); 345 goto fail; 346 } 347 348 err = bus_dmamap_create(sc->sc_dma_tag, 0, &sc->sc_dma_map); 349 if (err) { 350 device_printf(dev, "bus_dmamap_create failed\n"); 351 goto fail; 352 } 353 354 /* FIXME: Fix along with other BUS_SPACE_PHYSADDR instances */ 355 sc->sc_sdhci_buffer_phys = rman_get_start(sc->sc_mem_res) + 356 SDHCI_BUFFER; 357 358 bus_generic_probe(dev); 359 bus_generic_attach(dev); 360 361 sdhci_start_slot(&sc->sc_slot); 362 363 /* Seed our copies. */ 364 sc->blksz_and_count = SDHCI_READ_4(dev, &sc->sc_slot, SDHCI_BLOCK_SIZE); 365 sc->cmd_and_mode = SDHCI_READ_4(dev, &sc->sc_slot, SDHCI_TRANSFER_MODE); 366 367 return (0); 368 369 fail: 370 if (sc->sc_intrhand) 371 bus_teardown_intr(dev, sc->sc_irq_res, sc->sc_intrhand); 372 if (sc->sc_irq_res) 373 bus_release_resource(dev, SYS_RES_IRQ, 0, sc->sc_irq_res); 374 if (sc->sc_mem_res) 375 bus_release_resource(dev, SYS_RES_MEMORY, 0, sc->sc_mem_res); 376 377 return (err); 378 } 379 380 static int 381 bcm_sdhci_detach(device_t dev) 382 { 383 384 return (EBUSY); 385 } 386 387 static void 388 bcm_sdhci_intr(void *arg) 389 { 390 struct bcm_sdhci_softc *sc = arg; 391 392 sdhci_generic_intr(&sc->sc_slot); 393 } 394 395 static int 396 bcm_sdhci_update_ios(device_t bus, device_t child) 397 { 398 struct bcm_sdhci_softc *sc; 399 struct mmc_ios *ios; 400 int rv; 401 402 sc = device_get_softc(bus); 403 ios = &sc->sc_slot.host.ios; 404 405 if (ios->power_mode == power_up) { 406 if (sc->sc_mmc_helper.vmmc_supply) 407 regulator_enable(sc->sc_mmc_helper.vmmc_supply); 408 if (sc->sc_mmc_helper.vqmmc_supply) 409 regulator_enable(sc->sc_mmc_helper.vqmmc_supply); 410 } 411 412 rv = sdhci_generic_update_ios(bus, child); 413 if (rv != 0) 414 return (rv); 415 416 if (ios->power_mode == power_off) { 417 if (sc->sc_mmc_helper.vmmc_supply) 418 regulator_disable(sc->sc_mmc_helper.vmmc_supply); 419 if (sc->sc_mmc_helper.vqmmc_supply) 420 regulator_disable(sc->sc_mmc_helper.vqmmc_supply); 421 } 422 423 return (0); 424 } 425 426 static int 427 bcm_sdhci_get_ro(device_t bus, device_t child) 428 { 429 430 return (0); 431 } 432 433 static inline uint32_t 434 RD4(struct bcm_sdhci_softc *sc, bus_size_t off) 435 { 436 uint32_t val = bus_space_read_4(sc->sc_bst, sc->sc_bsh, off); 437 return val; 438 } 439 440 static inline void 441 WR4(struct bcm_sdhci_softc *sc, bus_size_t off, uint32_t val) 442 { 443 444 bus_space_write_4(sc->sc_bst, sc->sc_bsh, off, val); 445 /* 446 * The Arasan HC has a bug where it may lose the content of 447 * consecutive writes to registers that are within two SD-card 448 * clock cycles of each other (a clock domain crossing problem). 449 */ 450 if (sc->sc_slot.clock > 0) 451 DELAY(((2 * 1000000) / sc->sc_slot.clock) + 1); 452 } 453 454 static uint8_t 455 bcm_sdhci_read_1(device_t dev, struct sdhci_slot *slot, bus_size_t off) 456 { 457 struct bcm_sdhci_softc *sc = device_get_softc(dev); 458 uint32_t val = RD4(sc, off & ~3); 459 460 return ((val >> (off & 3)*8) & 0xff); 461 } 462 463 static uint16_t 464 bcm_sdhci_read_2(device_t dev, struct sdhci_slot *slot, bus_size_t off) 465 { 466 struct bcm_sdhci_softc *sc = device_get_softc(dev); 467 uint32_t val32; 468 469 /* 470 * Standard 32-bit handling of command and transfer mode, as 471 * well as block size and count. 472 */ 473 if ((off == SDHCI_BLOCK_SIZE || off == SDHCI_BLOCK_COUNT) && 474 sc->need_update_blk) 475 val32 = sc->blksz_and_count; 476 else if (off == SDHCI_TRANSFER_MODE || off == SDHCI_COMMAND_FLAGS) 477 val32 = sc->cmd_and_mode; 478 else 479 val32 = RD4(sc, off & ~3); 480 481 return ((val32 >> (off & 3)*8) & 0xffff); 482 } 483 484 static uint32_t 485 bcm_sdhci_read_4(device_t dev, struct sdhci_slot *slot, bus_size_t off) 486 { 487 struct bcm_sdhci_softc *sc = device_get_softc(dev); 488 489 return RD4(sc, off); 490 } 491 492 static void 493 bcm_sdhci_read_multi_4(device_t dev, struct sdhci_slot *slot, bus_size_t off, 494 uint32_t *data, bus_size_t count) 495 { 496 struct bcm_sdhci_softc *sc = device_get_softc(dev); 497 498 bus_space_read_multi_4(sc->sc_bst, sc->sc_bsh, off, data, count); 499 } 500 501 static void 502 bcm_sdhci_write_1(device_t dev, struct sdhci_slot *slot, bus_size_t off, 503 uint8_t val) 504 { 505 struct bcm_sdhci_softc *sc = device_get_softc(dev); 506 uint32_t val32 = RD4(sc, off & ~3); 507 val32 &= ~(0xff << (off & 3)*8); 508 val32 |= (val << (off & 3)*8); 509 WR4(sc, off & ~3, val32); 510 } 511 512 static void 513 bcm_sdhci_write_2(device_t dev, struct sdhci_slot *slot, bus_size_t off, 514 uint16_t val) 515 { 516 struct bcm_sdhci_softc *sc = device_get_softc(dev); 517 uint32_t val32; 518 519 /* 520 * If we have a queued up 16bit value for blk size or count, use and 521 * update the saved value rather than doing any real register access. 522 * If we did not touch either since the last write, then read from 523 * register as at least block count can change. 524 * Similarly, if we are about to issue a command, always use the saved 525 * value for transfer mode as we can never write that without issuing 526 * a command. 527 */ 528 if ((off == SDHCI_BLOCK_SIZE || off == SDHCI_BLOCK_COUNT) && 529 sc->need_update_blk) 530 val32 = sc->blksz_and_count; 531 else if (off == SDHCI_COMMAND_FLAGS) 532 val32 = sc->cmd_and_mode; 533 else 534 val32 = RD4(sc, off & ~3); 535 536 val32 &= ~(0xffff << (off & 3)*8); 537 val32 |= (val << (off & 3)*8); 538 539 if (off == SDHCI_TRANSFER_MODE) 540 sc->cmd_and_mode = val32; 541 else if (off == SDHCI_BLOCK_SIZE || off == SDHCI_BLOCK_COUNT) { 542 sc->blksz_and_count = val32; 543 sc->need_update_blk = true; 544 } else { 545 if (off == SDHCI_COMMAND_FLAGS) { 546 /* If we saved blk writes, do them now before cmd. */ 547 if (sc->need_update_blk) { 548 WR4(sc, SDHCI_BLOCK_SIZE, sc->blksz_and_count); 549 sc->need_update_blk = false; 550 } 551 /* Always save cmd and mode registers. */ 552 sc->cmd_and_mode = val32; 553 } 554 WR4(sc, off & ~3, val32); 555 } 556 } 557 558 static void 559 bcm_sdhci_write_4(device_t dev, struct sdhci_slot *slot, bus_size_t off, 560 uint32_t val) 561 { 562 struct bcm_sdhci_softc *sc = device_get_softc(dev); 563 WR4(sc, off, val); 564 } 565 566 static void 567 bcm_sdhci_write_multi_4(device_t dev, struct sdhci_slot *slot, bus_size_t off, 568 uint32_t *data, bus_size_t count) 569 { 570 struct bcm_sdhci_softc *sc = device_get_softc(dev); 571 572 bus_space_write_multi_4(sc->sc_bst, sc->sc_bsh, off, data, count); 573 } 574 575 static void 576 bcm_sdhci_start_dma_seg(struct bcm_sdhci_softc *sc) 577 { 578 struct sdhci_slot *slot; 579 vm_paddr_t pdst, psrc; 580 int err __diagused, idx, len, sync_op, width; 581 582 slot = &sc->sc_slot; 583 mtx_assert(&slot->mtx, MA_OWNED); 584 idx = sc->dmamap_seg_index++; 585 len = sc->dmamap_seg_sizes[idx]; 586 slot->offset += len; 587 width = (len & 0xf ? BCM_DMA_32BIT : BCM_DMA_128BIT); 588 589 if (slot->curcmd->data->flags & MMC_DATA_READ) { 590 /* 591 * Peripherals on the AXI bus do not need DREQ pacing for reads 592 * from the ARM core, so we can safely set this to NONE. 593 */ 594 bcm_dma_setup_src(sc->sc_dma_ch, BCM_DMA_DREQ_NONE, 595 BCM_DMA_SAME_ADDR, BCM_DMA_32BIT); 596 bcm_dma_setup_dst(sc->sc_dma_ch, BCM_DMA_DREQ_NONE, 597 BCM_DMA_INC_ADDR, width); 598 psrc = sc->sc_sdhci_buffer_phys; 599 pdst = sc->dmamap_seg_addrs[idx]; 600 sync_op = BUS_DMASYNC_PREREAD; 601 } else { 602 /* 603 * The ordering here is important, because the last write to 604 * dst/src in the dma control block writes the real dreq value. 605 */ 606 bcm_dma_setup_src(sc->sc_dma_ch, BCM_DMA_DREQ_NONE, 607 BCM_DMA_INC_ADDR, width); 608 bcm_dma_setup_dst(sc->sc_dma_ch, sc->conf->emmc_dreq, 609 BCM_DMA_SAME_ADDR, BCM_DMA_32BIT); 610 psrc = sc->dmamap_seg_addrs[idx]; 611 pdst = sc->sc_sdhci_buffer_phys; 612 sync_op = BUS_DMASYNC_PREWRITE; 613 } 614 615 /* 616 * When starting a new DMA operation do the busdma sync operation, and 617 * disable SDCHI data interrrupts because we'll be driven by DMA 618 * interrupts (or SDHCI error interrupts) until the IO is done. 619 */ 620 if (idx == 0) { 621 bus_dmamap_sync(sc->sc_dma_tag, sc->sc_dma_map, sync_op); 622 623 slot->intmask &= ~DATA_XFER_MASK; 624 bcm_sdhci_write_4(sc->sc_dev, slot, SDHCI_SIGNAL_ENABLE, 625 slot->intmask); 626 } 627 628 /* 629 * Start the DMA transfer. Only programming errors (like failing to 630 * allocate a channel) cause a non-zero return from bcm_dma_start(). 631 */ 632 err = bcm_dma_start(sc->sc_dma_ch, psrc, pdst, len); 633 KASSERT((err == 0), ("bcm2835_sdhci: failed DMA start")); 634 } 635 636 static void 637 bcm_sdhci_dma_exit(struct bcm_sdhci_softc *sc) 638 { 639 struct sdhci_slot *slot = &sc->sc_slot; 640 641 mtx_assert(&slot->mtx, MA_OWNED); 642 643 /* Re-enable interrupts */ 644 slot->intmask |= DATA_XFER_MASK; 645 bcm_sdhci_write_4(slot->bus, slot, SDHCI_SIGNAL_ENABLE, 646 slot->intmask); 647 } 648 649 static void 650 bcm_sdhci_dma_unload(struct bcm_sdhci_softc *sc) 651 { 652 struct sdhci_slot *slot = &sc->sc_slot; 653 654 if (sc->dmamap_seg_count == 0) 655 return; 656 if ((slot->curcmd->data->flags & MMC_DATA_READ) != 0) 657 bus_dmamap_sync(sc->sc_dma_tag, sc->sc_dma_map, 658 BUS_DMASYNC_POSTREAD); 659 else 660 bus_dmamap_sync(sc->sc_dma_tag, sc->sc_dma_map, 661 BUS_DMASYNC_POSTWRITE); 662 bus_dmamap_unload(sc->sc_dma_tag, sc->sc_dma_map); 663 664 sc->dmamap_seg_count = 0; 665 sc->dmamap_seg_index = 0; 666 } 667 668 static void 669 bcm_sdhci_dma_intr(int ch, void *arg) 670 { 671 struct bcm_sdhci_softc *sc = (struct bcm_sdhci_softc *)arg; 672 struct sdhci_slot *slot = &sc->sc_slot; 673 uint32_t reg; 674 675 mtx_lock(&slot->mtx); 676 if (slot->curcmd == NULL) 677 goto out; 678 /* 679 * If there are more segments for the current dma, start the next one. 680 * Otherwise unload the dma map and decide what to do next based on the 681 * status of the sdhci controller and whether there's more data left. 682 */ 683 if (sc->dmamap_seg_index < sc->dmamap_seg_count) { 684 bcm_sdhci_start_dma_seg(sc); 685 goto out; 686 } 687 688 bcm_sdhci_dma_unload(sc); 689 690 /* 691 * If we had no further segments pending, we need to determine how to 692 * proceed next. If the 'data/space pending' bit is already set and we 693 * can continue via DMA, do so. Otherwise, re-enable interrupts and 694 * return. 695 */ 696 reg = bcm_sdhci_read_4(slot->bus, slot, SDHCI_INT_STATUS) & 697 DATA_XFER_MASK; 698 if ((reg & DATA_PENDING_MASK) != 0 && 699 BCM_SDHCI_SEGSZ_LEFT(slot) >= BCM_SDHCI_BUFFER_SIZE) { 700 /* ACK any pending interrupts */ 701 bcm_sdhci_write_4(slot->bus, slot, SDHCI_INT_STATUS, 702 DATA_PENDING_MASK); 703 704 bcm_sdhci_start_dma(slot); 705 if (slot->curcmd->error != 0) { 706 /* We won't recover from this error for this command. */ 707 bcm_sdhci_dma_unload(sc); 708 bcm_sdhci_dma_exit(sc); 709 sdhci_finish_data(slot); 710 } 711 } else if ((reg & SDHCI_INT_DATA_END) != 0) { 712 bcm_sdhci_dma_exit(sc); 713 bcm_sdhci_write_4(slot->bus, slot, SDHCI_INT_STATUS, 714 reg); 715 slot->flags &= ~PLATFORM_DATA_STARTED; 716 sdhci_finish_data(slot); 717 } else { 718 bcm_sdhci_dma_exit(sc); 719 } 720 out: 721 mtx_unlock(&slot->mtx); 722 } 723 724 static void 725 bcm_sdhci_start_dma(struct sdhci_slot *slot) 726 { 727 struct bcm_sdhci_softc *sc = device_get_softc(slot->bus); 728 uint8_t *buf; 729 size_t left; 730 731 mtx_assert(&slot->mtx, MA_OWNED); 732 733 left = BCM_SDHCI_SEGSZ_LEFT(slot); 734 buf = (uint8_t *)slot->curcmd->data->data + slot->offset; 735 KASSERT(left != 0, 736 ("%s: DMA handling incorrectly indicated", __func__)); 737 738 /* 739 * No need to check segment count here; if we've not yet unloaded 740 * previous segments, we'll catch that in bcm_sdhci_dmacb. 741 */ 742 if (bus_dmamap_load(sc->sc_dma_tag, sc->sc_dma_map, buf, left, 743 bcm_sdhci_dmacb, sc, BUS_DMA_NOWAIT) != 0 || 744 sc->dmamap_status != 0) { 745 slot->curcmd->error = MMC_ERR_NO_MEMORY; 746 return; 747 } 748 749 /* DMA start */ 750 bcm_sdhci_start_dma_seg(sc); 751 } 752 753 static int 754 bcm_sdhci_will_handle_transfer(device_t dev, struct sdhci_slot *slot) 755 { 756 #ifdef INVARIANTS 757 struct bcm_sdhci_softc *sc = device_get_softc(slot->bus); 758 #endif 759 760 /* 761 * We don't want to perform DMA in this context -- interrupts are 762 * disabled, and a transaction may already be in progress. 763 */ 764 if (dumping) 765 return (0); 766 767 /* 768 * This indicates that we somehow let a data interrupt slip by into the 769 * SDHCI framework, when it should not have. This really needs to be 770 * caught and fixed ASAP, as it really shouldn't happen. 771 */ 772 KASSERT(sc->dmamap_seg_count == 0, 773 ("data pending interrupt pushed through SDHCI framework")); 774 775 /* 776 * Do not use DMA for transfers less than our block size. Checking 777 * alignment serves little benefit, as we round transfer sizes down to 778 * a multiple of the block size and push the transfer back to 779 * SDHCI-driven PIO once we're below the block size. 780 */ 781 if (BCM_SDHCI_SEGSZ_LEFT(slot) < BCM_DMA_BLOCK_SIZE) 782 return (0); 783 784 return (1); 785 } 786 787 static void 788 bcm_sdhci_start_transfer(device_t dev, struct sdhci_slot *slot, 789 uint32_t *intmask) 790 { 791 792 /* DMA transfer FIFO 1KB */ 793 bcm_sdhci_start_dma(slot); 794 } 795 796 static void 797 bcm_sdhci_finish_transfer(device_t dev, struct sdhci_slot *slot) 798 { 799 struct bcm_sdhci_softc *sc = device_get_softc(slot->bus); 800 801 /* 802 * Clean up. Interrupts are clearly enabled, because we received an 803 * SDHCI_INT_DATA_END to get this far -- just make sure we don't leave 804 * anything laying around. 805 */ 806 if (sc->dmamap_seg_count != 0) { 807 /* 808 * Our segment math should have worked out such that we would 809 * never finish the transfer without having used up all of the 810 * segments. If we haven't, that means we must have erroneously 811 * regressed to SDHCI-driven PIO to finish the operation and 812 * this is certainly caused by developer-error. 813 */ 814 bcm_sdhci_dma_unload(sc); 815 } 816 817 sdhci_finish_data(slot); 818 } 819 820 static device_method_t bcm_sdhci_methods[] = { 821 /* Device interface */ 822 DEVMETHOD(device_probe, bcm_sdhci_probe), 823 DEVMETHOD(device_attach, bcm_sdhci_attach), 824 DEVMETHOD(device_detach, bcm_sdhci_detach), 825 826 /* Bus interface */ 827 DEVMETHOD(bus_read_ivar, sdhci_generic_read_ivar), 828 DEVMETHOD(bus_write_ivar, sdhci_generic_write_ivar), 829 DEVMETHOD(bus_add_child, bus_generic_add_child), 830 831 /* MMC bridge interface */ 832 DEVMETHOD(mmcbr_update_ios, bcm_sdhci_update_ios), 833 DEVMETHOD(mmcbr_request, sdhci_generic_request), 834 DEVMETHOD(mmcbr_get_ro, bcm_sdhci_get_ro), 835 DEVMETHOD(mmcbr_acquire_host, sdhci_generic_acquire_host), 836 DEVMETHOD(mmcbr_release_host, sdhci_generic_release_host), 837 838 /* Platform transfer methods */ 839 DEVMETHOD(sdhci_platform_will_handle, bcm_sdhci_will_handle_transfer), 840 DEVMETHOD(sdhci_platform_start_transfer, bcm_sdhci_start_transfer), 841 DEVMETHOD(sdhci_platform_finish_transfer, bcm_sdhci_finish_transfer), 842 /* SDHCI registers accessors */ 843 DEVMETHOD(sdhci_read_1, bcm_sdhci_read_1), 844 DEVMETHOD(sdhci_read_2, bcm_sdhci_read_2), 845 DEVMETHOD(sdhci_read_4, bcm_sdhci_read_4), 846 DEVMETHOD(sdhci_read_multi_4, bcm_sdhci_read_multi_4), 847 DEVMETHOD(sdhci_write_1, bcm_sdhci_write_1), 848 DEVMETHOD(sdhci_write_2, bcm_sdhci_write_2), 849 DEVMETHOD(sdhci_write_4, bcm_sdhci_write_4), 850 DEVMETHOD(sdhci_write_multi_4, bcm_sdhci_write_multi_4), 851 852 DEVMETHOD_END 853 }; 854 855 static driver_t bcm_sdhci_driver = { 856 "sdhci_bcm", 857 bcm_sdhci_methods, 858 sizeof(struct bcm_sdhci_softc), 859 }; 860 861 DRIVER_MODULE(sdhci_bcm, simplebus, bcm_sdhci_driver, NULL, NULL); 862 #ifdef NOTYET 863 MODULE_DEPEND(sdhci_bcm, bcm2835_clkman, 1, 1, 1); 864 #endif 865 SDHCI_DEPEND(sdhci_bcm); 866 #ifndef MMCCAM 867 MMC_DECLARE_BRIDGE(sdhci_bcm); 868 #endif 869