1 /*-
2 * Copyright (c) 2015-2016 Landon Fuller <landon@landonf.org>
3 * Copyright (c) 2017 The FreeBSD Foundation
4 * All rights reserved.
5 *
6 * Portions of this software were developed by Landon Fuller
7 * under sponsorship from the FreeBSD Foundation.
8 *
9 * Redistribution and use in source and binary forms, with or without
10 * modification, are permitted provided that the following conditions
11 * are met:
12 * 1. Redistributions of source code must retain the above copyright
13 * notice, this list of conditions and the following disclaimer,
14 * without modification.
15 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
16 * similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
17 * redistribution must be conditioned upon including a substantially
18 * similar Disclaimer requirement for further binary redistribution.
19 *
20 * NO WARRANTY
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
24 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
25 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
26 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
27 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
28 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
29 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
30 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
31 * THE POSSIBILITY OF SUCH DAMAGES.
32 */
33
34 #include <sys/param.h>
35 #include <sys/kernel.h>
36 #include <sys/limits.h>
37
38 #include "bhndb_private.h"
39 #include "bhndbvar.h"
40
41 static int bhndb_dma_tag_create(device_t dev, bus_dma_tag_t parent_dmat,
42 const struct bhnd_dma_translation *translation,
43 bus_dma_tag_t *dmat);
44
45 /**
46 * Attach a BHND bridge device to @p parent.
47 *
48 * @param parent A parent PCI device.
49 * @param[out] bhndb On success, the probed and attached bhndb bridge device.
50 * @param unit The device unit number, or -1 to select the next available unit
51 * number.
52 *
53 * @retval 0 success
54 * @retval non-zero Failed to attach the bhndb device.
55 */
56 int
bhndb_attach_bridge(device_t parent,device_t * bhndb,int unit)57 bhndb_attach_bridge(device_t parent, device_t *bhndb, int unit)
58 {
59 int error;
60
61 *bhndb = device_add_child(parent, "bhndb", unit);
62 if (*bhndb == NULL)
63 return (ENXIO);
64
65 if (!(error = device_probe_and_attach(*bhndb)))
66 return (0);
67
68 if ((device_delete_child(parent, *bhndb)))
69 device_printf(parent, "failed to detach bhndb child\n");
70
71 return (error);
72 }
73
74 /*
75 * Call BHNDB_SUSPEND_RESOURCE() for all resources in @p rl.
76 */
77 static void
bhndb_do_suspend_resources(device_t dev,struct resource_list * rl)78 bhndb_do_suspend_resources(device_t dev, struct resource_list *rl)
79 {
80 struct resource_list_entry *rle;
81
82 /* Suspend all child resources. */
83 STAILQ_FOREACH(rle, rl, link) {
84 /* Skip non-allocated resources */
85 if (rle->res == NULL)
86 continue;
87
88 BHNDB_SUSPEND_RESOURCE(device_get_parent(dev), dev, rle->type,
89 rle->res);
90 }
91 }
92
93 /**
94 * Helper function for implementing BUS_RESUME_CHILD() on bridged
95 * bhnd(4) buses.
96 *
97 * This implementation of BUS_RESUME_CHILD() uses BUS_GET_RESOURCE_LIST()
98 * to find the child's resources and call BHNDB_SUSPEND_RESOURCE() for all
99 * child resources, ensuring that the device's allocated bridge resources
100 * will be available to other devices during bus resumption.
101 *
102 * Before suspending any resources, @p child is suspended by
103 * calling bhnd_generic_suspend_child().
104 *
105 * If @p child is not a direct child of @p dev, suspension is delegated to
106 * the @p dev parent.
107 */
108 int
bhnd_generic_br_suspend_child(device_t dev,device_t child)109 bhnd_generic_br_suspend_child(device_t dev, device_t child)
110 {
111 struct resource_list *rl;
112 int error;
113
114 if (device_get_parent(child) != dev)
115 BUS_SUSPEND_CHILD(device_get_parent(dev), child);
116
117 if (device_is_suspended(child))
118 return (EBUSY);
119
120 /* Suspend the child device */
121 if ((error = bhnd_generic_suspend_child(dev, child)))
122 return (error);
123
124 /* Fetch the resource list. If none, there's nothing else to do */
125 rl = BUS_GET_RESOURCE_LIST(device_get_parent(child), child);
126 if (rl == NULL)
127 return (0);
128
129 /* Suspend all child resources. */
130 bhndb_do_suspend_resources(dev, rl);
131
132 return (0);
133 }
134
135 /**
136 * Helper function for implementing BUS_RESUME_CHILD() on bridged
137 * bhnd(4) bus devices.
138 *
139 * This implementation of BUS_RESUME_CHILD() uses BUS_GET_RESOURCE_LIST()
140 * to find the child's resources and call BHNDB_RESUME_RESOURCE() for all
141 * child resources, before delegating to bhnd_generic_resume_child().
142 *
143 * If resource resumption fails, @p child will not be resumed.
144 *
145 * If @p child is not a direct child of @p dev, suspension is delegated to
146 * the @p dev parent.
147 */
148 int
bhnd_generic_br_resume_child(device_t dev,device_t child)149 bhnd_generic_br_resume_child(device_t dev, device_t child)
150 {
151 struct resource_list *rl;
152 struct resource_list_entry *rle;
153 int error;
154
155 if (device_get_parent(child) != dev)
156 BUS_RESUME_CHILD(device_get_parent(dev), child);
157
158 if (!device_is_suspended(child))
159 return (EBUSY);
160
161 /* Fetch the resource list. If none, there's nothing else to do */
162 rl = BUS_GET_RESOURCE_LIST(device_get_parent(child), child);
163 if (rl == NULL)
164 return (bhnd_generic_resume_child(dev, child));
165
166 /* Resume all resources */
167 STAILQ_FOREACH(rle, rl, link) {
168 /* Skip non-allocated resources */
169 if (rle->res == NULL)
170 continue;
171
172 error = BHNDB_RESUME_RESOURCE(device_get_parent(dev), dev,
173 rle->type, rle->res);
174 if (error) {
175 /* Put all resources back into a suspend state */
176 bhndb_do_suspend_resources(dev, rl);
177 return (error);
178 }
179 }
180
181 /* Now that all resources are resumed, resume child */
182 if ((error = bhnd_generic_resume_child(dev, child))) {
183 /* Put all resources back into a suspend state */
184 bhndb_do_suspend_resources(dev, rl);
185 }
186
187 return (error);
188 }
189
190 /**
191 * Find a host resource of @p type that maps the given range.
192 *
193 * @param hr The resource state to search.
194 * @param type The resource type to search for (see SYS_RES_*).
195 * @param start The start address of the range to search for.
196 * @param count The size of the range to search for.
197 *
198 * @retval resource the host resource containing the requested range.
199 * @retval NULL if no resource containing the requested range can be found.
200 */
201 struct resource *
bhndb_host_resource_for_range(struct bhndb_host_resources * hr,int type,rman_res_t start,rman_res_t count)202 bhndb_host_resource_for_range(struct bhndb_host_resources *hr, int type,
203 rman_res_t start, rman_res_t count)
204 {
205 for (u_int i = 0; hr->resource_specs[i].type != -1; i++) {
206 struct resource *r = hr->resources[i];
207
208 if (hr->resource_specs[i].type != type)
209 continue;
210
211 /* Verify range */
212 if (rman_get_start(r) > start)
213 continue;
214
215 if (rman_get_end(r) < (start + count - 1))
216 continue;
217
218 return (r);
219 }
220
221 return (NULL);
222 }
223
224 /**
225 * Find a host resource of that matches the given register window definition.
226 *
227 * @param hr The resource state to search.
228 * @param win A register window definition.
229 *
230 * @retval resource the host resource corresponding to @p win.
231 * @retval NULL if no resource corresponding to @p win can be found.
232 */
233 struct resource *
bhndb_host_resource_for_regwin(struct bhndb_host_resources * hr,const struct bhndb_regwin * win)234 bhndb_host_resource_for_regwin(struct bhndb_host_resources *hr,
235 const struct bhndb_regwin *win)
236 {
237 const struct resource_spec *rspecs;
238
239 rspecs = hr->resource_specs;
240 for (u_int i = 0; rspecs[i].type != -1; i++) {
241 if (win->res.type != rspecs[i].type)
242 continue;
243
244 if (win->res.rid != rspecs[i].rid)
245 continue;
246
247 /* Found declared resource */
248 return (hr->resources[i]);
249 }
250
251 device_printf(hr->owner, "missing regwin resource spec "
252 "(type=%d, rid=%d)\n", win->res.type, win->res.rid);
253
254 return (NULL);
255 }
256
257 /**
258 * Allocate and initialize a new resource state structure.
259 *
260 * @param dev The bridge device.
261 * @param parent_dev The parent device from which host resources should be
262 * allocated.
263 * @param cfg The hardware configuration to be used.
264 */
265 struct bhndb_resources *
bhndb_alloc_resources(device_t dev,device_t parent_dev,const struct bhndb_hwcfg * cfg)266 bhndb_alloc_resources(device_t dev, device_t parent_dev,
267 const struct bhndb_hwcfg *cfg)
268 {
269 struct bhndb_resources *r;
270 const struct bhndb_regwin *win;
271 bus_size_t last_window_size;
272 int rnid;
273 int error;
274 bool free_ht_mem, free_br_mem, free_br_irq;
275
276 free_ht_mem = false;
277 free_br_mem = false;
278 free_br_irq = false;
279
280 r = malloc(sizeof(*r), M_BHND, M_NOWAIT|M_ZERO);
281 if (r == NULL)
282 return (NULL);
283
284 /* Basic initialization */
285 r->dev = dev;
286 r->cfg = cfg;
287 r->res = NULL;
288 r->min_prio = BHNDB_PRIORITY_NONE;
289 STAILQ_INIT(&r->bus_regions);
290 STAILQ_INIT(&r->bus_intrs);
291
292 mtx_init(&r->dw_steal_mtx, device_get_nameunit(dev),
293 "bhndb dwa_steal lock", MTX_SPIN);
294
295 /* Initialize host address space resource manager. */
296 r->ht_mem_rman.rm_start = 0;
297 r->ht_mem_rman.rm_end = ~0;
298 r->ht_mem_rman.rm_type = RMAN_ARRAY;
299 r->ht_mem_rman.rm_descr = "BHNDB host memory";
300 if ((error = rman_init(&r->ht_mem_rman))) {
301 device_printf(r->dev, "could not initialize ht_mem_rman\n");
302 goto failed;
303 }
304 free_ht_mem = true;
305
306 /* Initialize resource manager for the bridged address space. */
307 r->br_mem_rman.rm_start = 0;
308 r->br_mem_rman.rm_end = BUS_SPACE_MAXADDR_32BIT;
309 r->br_mem_rman.rm_type = RMAN_ARRAY;
310 r->br_mem_rman.rm_descr = "BHNDB bridged memory";
311
312 if ((error = rman_init(&r->br_mem_rman))) {
313 device_printf(r->dev, "could not initialize br_mem_rman\n");
314 goto failed;
315 }
316 free_br_mem = true;
317
318 error = rman_manage_region(&r->br_mem_rman, 0, BUS_SPACE_MAXADDR_32BIT);
319 if (error) {
320 device_printf(r->dev, "could not configure br_mem_rman\n");
321 goto failed;
322 }
323
324 /* Initialize resource manager for the bridged interrupt controller. */
325 r->br_irq_rman.rm_start = 0;
326 r->br_irq_rman.rm_end = RM_MAX_END;
327 r->br_irq_rman.rm_type = RMAN_ARRAY;
328 r->br_irq_rman.rm_descr = "BHNDB bridged interrupts";
329
330 if ((error = rman_init(&r->br_irq_rman))) {
331 device_printf(r->dev, "could not initialize br_irq_rman\n");
332 goto failed;
333 }
334 free_br_irq = true;
335
336 error = rman_manage_region(&r->br_irq_rman, 0, RM_MAX_END);
337 if (error) {
338 device_printf(r->dev, "could not configure br_irq_rman\n");
339 goto failed;
340 }
341
342 /* Fetch the dynamic regwin count and verify that it does not exceed
343 * what is representable via our freelist bitstring. */
344 r->dwa_count = bhndb_regwin_count(cfg->register_windows,
345 BHNDB_REGWIN_T_DYN);
346 if (r->dwa_count >= INT_MAX) {
347 device_printf(r->dev, "max dynamic regwin count exceeded\n");
348 goto failed;
349 }
350
351 /* Allocate the dynamic window allocation table. */
352 r->dw_alloc = malloc(sizeof(r->dw_alloc[0]) * r->dwa_count, M_BHND,
353 M_NOWAIT);
354 if (r->dw_alloc == NULL)
355 goto failed;
356
357 /* Allocate the dynamic window allocation freelist */
358 r->dwa_freelist = bit_alloc(r->dwa_count, M_BHND, M_NOWAIT);
359 if (r->dwa_freelist == NULL)
360 goto failed;
361
362 /* Initialize the dynamic window table */
363 rnid = 0;
364 last_window_size = 0;
365 for (win = cfg->register_windows;
366 win->win_type != BHNDB_REGWIN_T_INVALID; win++)
367 {
368 struct bhndb_dw_alloc *dwa;
369
370 /* Skip non-DYN windows */
371 if (win->win_type != BHNDB_REGWIN_T_DYN)
372 continue;
373
374 /* Validate the window size */
375 if (win->win_size == 0) {
376 device_printf(r->dev, "ignoring zero-length dynamic "
377 "register window\n");
378 continue;
379 } else if (last_window_size == 0) {
380 last_window_size = win->win_size;
381 } else if (last_window_size != win->win_size) {
382 /*
383 * No existing hardware should trigger this.
384 *
385 * If you run into this in the future, the dynamic
386 * window allocator and the resource priority system
387 * will need to be extended to support multiple register
388 * window allocation pools.
389 */
390 device_printf(r->dev, "devices that vend multiple "
391 "dynamic register window sizes are not currently "
392 "supported\n");
393 goto failed;
394 }
395
396 dwa = &r->dw_alloc[rnid];
397 dwa->win = win;
398 dwa->parent_res = NULL;
399 dwa->rnid = rnid;
400 dwa->target = 0x0;
401
402 LIST_INIT(&dwa->refs);
403 rnid++;
404 }
405
406 /* Allocate host resources */
407 error = bhndb_alloc_host_resources(&r->res, dev, parent_dev, r->cfg);
408 if (error) {
409 device_printf(r->dev,
410 "could not allocate host resources on %s: %d\n",
411 device_get_nameunit(parent_dev), error);
412 goto failed;
413 }
414
415 /* Populate (and validate) parent resource references for all
416 * dynamic windows */
417 for (size_t i = 0; i < r->dwa_count; i++) {
418 struct bhndb_dw_alloc *dwa;
419 const struct bhndb_regwin *win;
420
421 dwa = &r->dw_alloc[i];
422 win = dwa->win;
423
424 /* Find and validate corresponding resource. */
425 dwa->parent_res = bhndb_host_resource_for_regwin(r->res, win);
426 if (dwa->parent_res == NULL) {
427 device_printf(r->dev, "no host resource found for %u "
428 "register window with offset %#jx and "
429 "size %#jx\n",
430 win->win_type,
431 (uintmax_t)win->win_offset,
432 (uintmax_t)win->win_size);
433
434 error = ENXIO;
435 goto failed;
436 }
437
438 if (rman_get_size(dwa->parent_res) < win->win_offset +
439 win->win_size)
440 {
441 device_printf(r->dev, "resource %d too small for "
442 "register window with offset %llx and size %llx\n",
443 rman_get_rid(dwa->parent_res),
444 (unsigned long long) win->win_offset,
445 (unsigned long long) win->win_size);
446
447 error = EINVAL;
448 goto failed;
449 }
450 }
451
452 /* Add allocated memory resources to our host memory resource manager */
453 for (u_int i = 0; r->res->resource_specs[i].type != -1; i++) {
454 struct resource *res;
455
456 /* skip non-memory resources */
457 if (r->res->resource_specs[i].type != SYS_RES_MEMORY)
458 continue;
459
460 /* add host resource to set of managed regions */
461 res = r->res->resources[i];
462 error = rman_manage_region(&r->ht_mem_rman,
463 rman_get_start(res), rman_get_end(res));
464 if (error) {
465 device_printf(r->dev,
466 "could not register host memory region with "
467 "ht_mem_rman: %d\n", error);
468 goto failed;
469 }
470 }
471
472 return (r);
473
474 failed:
475 if (free_ht_mem)
476 rman_fini(&r->ht_mem_rman);
477
478 if (free_br_mem)
479 rman_fini(&r->br_mem_rman);
480
481 if (free_br_irq)
482 rman_fini(&r->br_irq_rman);
483
484 if (r->dw_alloc != NULL)
485 free(r->dw_alloc, M_BHND);
486
487 if (r->dwa_freelist != NULL)
488 free(r->dwa_freelist, M_BHND);
489
490 if (r->res != NULL)
491 bhndb_release_host_resources(r->res);
492
493 mtx_destroy(&r->dw_steal_mtx);
494
495 free(r, M_BHND);
496
497 return (NULL);
498 }
499
500 /**
501 * Create a new DMA tag for the given @p translation.
502 *
503 * @param dev The bridge device.
504 * @param parent_dmat The parent DMA tag, or NULL if none.
505 * @param translation The DMA translation for which a DMA tag will
506 * be created.
507 * @param[out] dmat On success, the newly created DMA tag.
508 *
509 * @retval 0 success
510 * @retval non-zero if creating the new DMA tag otherwise fails, a regular
511 * unix error code will be returned.
512 */
513 static int
bhndb_dma_tag_create(device_t dev,bus_dma_tag_t parent_dmat,const struct bhnd_dma_translation * translation,bus_dma_tag_t * dmat)514 bhndb_dma_tag_create(device_t dev, bus_dma_tag_t parent_dmat,
515 const struct bhnd_dma_translation *translation, bus_dma_tag_t *dmat)
516 {
517 bus_dma_tag_t translation_tag;
518 bhnd_addr_t dt_mask;
519 bus_addr_t lowaddr, highaddr;
520 bus_size_t maxsegsz;
521 int error;
522
523 highaddr = BUS_SPACE_MAXADDR;
524 maxsegsz = BUS_SPACE_MAXSIZE;
525
526 /* Determine full addressable mask */
527 dt_mask = (translation->addr_mask | translation->addrext_mask);
528 KASSERT(dt_mask != 0, ("DMA addr_mask invalid: %#jx",
529 (uintmax_t)dt_mask));
530
531 /* (addr_mask|addrext_mask) is our maximum supported address */
532 lowaddr = MIN(dt_mask, BUS_SPACE_MAXADDR);
533
534 /* Constrain to translation window size */
535 if (translation->addr_mask < maxsegsz)
536 maxsegsz = translation->addr_mask;
537
538 /* Create our DMA tag */
539 error = bus_dma_tag_create(parent_dmat,
540 1, 0, /* alignment, boundary */
541 lowaddr, highaddr,
542 NULL, NULL, /* filter, filterarg */
543 BUS_SPACE_MAXSIZE, 0, /* maxsize, nsegments */
544 maxsegsz, 0, /* maxsegsize, flags */
545 NULL, NULL, /* lockfunc, lockarg */
546 &translation_tag);
547 if (error) {
548 device_printf(dev, "failed to create bridge DMA tag: %d\n",
549 error);
550 return (error);
551 }
552
553 *dmat = translation_tag;
554 return (0);
555 }
556
557 /**
558 * Deallocate the given bridge resource structure and any associated resources.
559 *
560 * @param br Resource state to be deallocated.
561 */
562 void
bhndb_free_resources(struct bhndb_resources * br)563 bhndb_free_resources(struct bhndb_resources *br)
564 {
565 struct bhndb_region *region, *r_next;
566 struct bhndb_dw_alloc *dwa;
567 struct bhndb_dw_rentry *dwr, *dwr_next;
568 struct bhndb_intr_handler *ih;
569 bool leaked_regions, leaked_intrs;
570
571 leaked_regions = false;
572 leaked_intrs = false;
573
574 /* No window regions may still be held */
575 if (!bhndb_dw_all_free(br)) {
576 for (int i = 0; i < br->dwa_count; i++) {
577 dwa = &br->dw_alloc[i];
578
579 /* Skip free dynamic windows */
580 if (bhndb_dw_is_free(br, dwa))
581 continue;
582
583 device_printf(br->dev,
584 "leaked dynamic register window %d\n", dwa->rnid);
585 leaked_regions = true;
586 }
587 }
588
589 /* There should be no interrupt handlers still registered */
590 STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
591 device_printf(br->dev, "interrupt handler leaked %p\n",
592 ih->ih_cookiep);
593 }
594
595 if (leaked_intrs || leaked_regions) {
596 panic("leaked%s%s", leaked_intrs ? " active interrupts" : "",
597 leaked_regions ? " active register windows" : "");
598 }
599
600 /* Release host resources allocated through our parent. */
601 if (br->res != NULL)
602 bhndb_release_host_resources(br->res);
603
604 /* Clean up resource reservations */
605 for (size_t i = 0; i < br->dwa_count; i++) {
606 dwa = &br->dw_alloc[i];
607
608 LIST_FOREACH_SAFE(dwr, &dwa->refs, dw_link, dwr_next) {
609 LIST_REMOVE(dwr, dw_link);
610 free(dwr, M_BHND);
611 }
612 }
613
614 /* Release bus regions */
615 STAILQ_FOREACH_SAFE(region, &br->bus_regions, link, r_next) {
616 STAILQ_REMOVE(&br->bus_regions, region, bhndb_region, link);
617 free(region, M_BHND);
618 }
619
620 /* Release our resource managers */
621 rman_fini(&br->ht_mem_rman);
622 rman_fini(&br->br_mem_rman);
623 rman_fini(&br->br_irq_rman);
624
625 free(br->dw_alloc, M_BHND);
626 free(br->dwa_freelist, M_BHND);
627
628 mtx_destroy(&br->dw_steal_mtx);
629
630 free(br, M_BHND);
631 }
632
633 /**
634 * Allocate host bus resources defined by @p hwcfg.
635 *
636 * On success, the caller assumes ownership of the allocated host resources,
637 * which must be freed via bhndb_release_host_resources().
638 *
639 * @param[out] resources On success, the allocated host resources.
640 * @param dev The bridge device.
641 * @param parent_dev The parent device from which host resources
642 * should be allocated (e.g. via
643 * bus_alloc_resources()).
644 * @param hwcfg The hardware configuration defining the host
645 * resources to be allocated
646 */
647 int
bhndb_alloc_host_resources(struct bhndb_host_resources ** resources,device_t dev,device_t parent_dev,const struct bhndb_hwcfg * hwcfg)648 bhndb_alloc_host_resources(struct bhndb_host_resources **resources,
649 device_t dev, device_t parent_dev, const struct bhndb_hwcfg *hwcfg)
650 {
651 struct bhndb_host_resources *hr;
652 const struct bhnd_dma_translation *dt;
653 bus_dma_tag_t parent_dmat;
654 size_t nres, ndt;
655 int error;
656
657 parent_dmat = bus_get_dma_tag(parent_dev);
658
659 hr = malloc(sizeof(*hr), M_BHND, M_WAITOK);
660 hr->owner = parent_dev;
661 hr->cfg = hwcfg;
662 hr->resource_specs = NULL;
663 hr->resources = NULL;
664 hr->dma_tags = NULL;
665 hr->num_dma_tags = 0;
666
667 /* Determine our bridge resource count from the hardware config. */
668 nres = 0;
669 for (size_t i = 0; hwcfg->resource_specs[i].type != -1; i++)
670 nres++;
671
672 /* Determine the total count and validate our DMA translation table. */
673 ndt = 0;
674 for (dt = hwcfg->dma_translations; dt != NULL &&
675 !BHND_DMA_IS_TRANSLATION_TABLE_END(dt); dt++)
676 {
677 /* Validate the defined translation */
678 if ((dt->base_addr & dt->addr_mask) != 0) {
679 device_printf(dev, "invalid DMA translation; base "
680 "address %#jx overlaps address mask %#jx",
681 (uintmax_t)dt->base_addr, (uintmax_t)dt->addr_mask);
682
683 error = EINVAL;
684 goto failed;
685 }
686
687 if ((dt->addrext_mask & dt->addr_mask) != 0) {
688 device_printf(dev, "invalid DMA translation; addrext "
689 "mask %#jx overlaps address mask %#jx",
690 (uintmax_t)dt->addrext_mask,
691 (uintmax_t)dt->addr_mask);
692
693 error = EINVAL;
694 goto failed;
695 }
696
697 /* Increment our entry count */
698 ndt++;
699 }
700
701 /* Allocate our DMA tags */
702 hr->dma_tags = malloc(sizeof(*hr->dma_tags) * ndt, M_BHND,
703 M_WAITOK|M_ZERO);
704 for (size_t i = 0; i < ndt; i++) {
705 error = bhndb_dma_tag_create(dev, parent_dmat,
706 &hwcfg->dma_translations[i], &hr->dma_tags[i]);
707 if (error)
708 goto failed;
709
710 hr->num_dma_tags++;
711 }
712
713 /* Allocate space for a non-const copy of our resource_spec
714 * table; this will be updated with the RIDs assigned by
715 * bus_alloc_resources. */
716 hr->resource_specs = malloc(sizeof(hr->resource_specs[0]) * (nres + 1),
717 M_BHND, M_WAITOK);
718
719 /* Initialize and terminate the table */
720 for (size_t i = 0; i < nres; i++)
721 hr->resource_specs[i] = hwcfg->resource_specs[i];
722
723 hr->resource_specs[nres].type = -1;
724
725 /* Allocate space for our resource references */
726 hr->resources = malloc(sizeof(hr->resources[0]) * nres, M_BHND,
727 M_WAITOK);
728
729 /* Allocate host resources */
730 error = bus_alloc_resources(hr->owner, hr->resource_specs,
731 hr->resources);
732 if (error) {
733 device_printf(dev, "could not allocate bridge resources via "
734 "%s: %d\n", device_get_nameunit(parent_dev), error);
735 goto failed;
736 }
737
738 *resources = hr;
739 return (0);
740
741 failed:
742 if (hr->resource_specs != NULL)
743 free(hr->resource_specs, M_BHND);
744
745 if (hr->resources != NULL)
746 free(hr->resources, M_BHND);
747
748 for (size_t i = 0; i < hr->num_dma_tags; i++)
749 bus_dma_tag_destroy(hr->dma_tags[i]);
750
751 if (hr->dma_tags != NULL)
752 free(hr->dma_tags, M_BHND);
753
754 free(hr, M_BHND);
755
756 return (error);
757 }
758
759 /**
760 * Deallocate a set of bridge host resources.
761 *
762 * @param hr The resources to be freed.
763 */
764 void
bhndb_release_host_resources(struct bhndb_host_resources * hr)765 bhndb_release_host_resources(struct bhndb_host_resources *hr)
766 {
767 bus_release_resources(hr->owner, hr->resource_specs, hr->resources);
768
769 for (size_t i = 0; i < hr->num_dma_tags; i++)
770 bus_dma_tag_destroy(hr->dma_tags[i]);
771
772 free(hr->resources, M_BHND);
773 free(hr->resource_specs, M_BHND);
774 free(hr->dma_tags, M_BHND);
775 free(hr, M_BHND);
776 }
777
778 /**
779 * Search @p cores for the core serving as the bhnd host bridge.
780 *
781 * This function uses a heuristic valid on all known PCI/PCIe/PCMCIA-bridged
782 * bhnd(4) devices to determine the hostb core:
783 *
784 * - The core must have a Broadcom vendor ID.
785 * - The core devclass must match the bridge type.
786 * - The core must be the first device on the bus with the bridged device
787 * class.
788 *
789 * @param cores The core table to search.
790 * @param ncores The number of cores in @p cores.
791 * @param bridge_devclass The expected device class of the bridge core.
792 * @param[out] core If found, the matching host bridge core info.
793 *
794 * @retval 0 success
795 * @retval ENOENT not found
796 */
797 int
bhndb_find_hostb_core(struct bhnd_core_info * cores,u_int ncores,bhnd_devclass_t bridge_devclass,struct bhnd_core_info * core)798 bhndb_find_hostb_core(struct bhnd_core_info *cores, u_int ncores,
799 bhnd_devclass_t bridge_devclass, struct bhnd_core_info *core)
800 {
801 struct bhnd_core_match md;
802 struct bhnd_core_info *match;
803 u_int match_core_idx;
804
805 /* Set up a match descriptor for the required device class. */
806 md = (struct bhnd_core_match) {
807 BHND_MATCH_CORE_CLASS(bridge_devclass),
808 BHND_MATCH_CORE_UNIT(0)
809 };
810
811 /* Find the matching core with the lowest core index */
812 match = NULL;
813 match_core_idx = UINT_MAX;
814
815 for (u_int i = 0; i < ncores; i++) {
816 if (!bhnd_core_matches(&cores[i], &md))
817 continue;
818
819 /* Lower core indices take precedence */
820 if (match != NULL && match_core_idx < match->core_idx)
821 continue;
822
823 match = &cores[i];
824 match_core_idx = match->core_idx;
825 }
826
827 if (match == NULL)
828 return (ENOENT);
829
830 *core = *match;
831 return (0);
832 }
833
834 /**
835 * Allocate a host interrupt source and its backing SYS_RES_IRQ host resource.
836 *
837 * @param owner The device to be used to allocate a SYS_RES_IRQ
838 * resource with @p rid.
839 * @param rid The resource ID of the IRQ to be allocated.
840 * @param start The start value to be passed to bus_alloc_resource().
841 * @param end The end value to be passed to bus_alloc_resource().
842 * @param count The count to be passed to bus_alloc_resource().
843 * @param flags The flags to be passed to bus_alloc_resource().
844 *
845 * @retval non-NULL success
846 * @retval NULL if allocation fails.
847 */
848 struct bhndb_intr_isrc *
bhndb_alloc_intr_isrc(device_t owner,int rid,rman_res_t start,rman_res_t end,rman_res_t count,u_int flags)849 bhndb_alloc_intr_isrc(device_t owner, int rid, rman_res_t start, rman_res_t end,
850 rman_res_t count, u_int flags)
851 {
852 struct bhndb_intr_isrc *isrc;
853
854 isrc = malloc(sizeof(*isrc), M_BHND, M_NOWAIT);
855 if (isrc == NULL)
856 return (NULL);
857
858 isrc->is_owner = owner;
859 isrc->is_rid = rid;
860 isrc->is_res = bus_alloc_resource(owner, SYS_RES_IRQ, &isrc->is_rid,
861 start, end, count, flags);
862 if (isrc->is_res == NULL) {
863 free(isrc, M_BHND);
864 return (NULL);
865 }
866
867 return (isrc);
868 }
869
870 /**
871 * Free a host interrupt source and its backing host resource.
872 *
873 * @param isrc The interrupt source to be freed.
874 */
875 void
bhndb_free_intr_isrc(struct bhndb_intr_isrc * isrc)876 bhndb_free_intr_isrc(struct bhndb_intr_isrc *isrc)
877 {
878 bus_release_resource(isrc->is_owner, SYS_RES_IRQ, isrc->is_rid,
879 isrc->is_res);
880 free(isrc, M_BHND);
881 }
882
883 /**
884 * Allocate and initialize a new interrupt handler entry.
885 *
886 * @param owner The child device that owns this entry.
887 * @param r The child's interrupt resource.
888 * @param isrc The isrc mapped for this entry.
889 *
890 * @retval non-NULL success
891 * @retval NULL if allocation fails.
892 */
893 struct bhndb_intr_handler *
bhndb_alloc_intr_handler(device_t owner,struct resource * r,struct bhndb_intr_isrc * isrc)894 bhndb_alloc_intr_handler(device_t owner, struct resource *r,
895 struct bhndb_intr_isrc *isrc)
896 {
897 struct bhndb_intr_handler *ih;
898
899 ih = malloc(sizeof(*ih), M_BHND, M_NOWAIT | M_ZERO);
900 ih->ih_owner = owner;
901 ih->ih_res = r;
902 ih->ih_isrc = isrc;
903 ih->ih_cookiep = NULL;
904 ih->ih_active = false;
905
906 return (ih);
907 }
908
909 /**
910 * Free an interrupt handler entry.
911 *
912 * @param br The resource state owning @p ih.
913 * @param ih The interrupt handler entry to be removed.
914 */
915 void
bhndb_free_intr_handler(struct bhndb_intr_handler * ih)916 bhndb_free_intr_handler(struct bhndb_intr_handler *ih)
917 {
918 KASSERT(!ih->ih_active, ("free of active interrupt handler %p",
919 ih->ih_cookiep));
920
921 free(ih, M_BHND);
922 }
923
924 /**
925 * Add an active interrupt handler to the given resource state.
926 *
927 * @param br The resource state to be modified.
928 * @param ih The interrupt handler entry to be added.
929 */
930 void
bhndb_register_intr_handler(struct bhndb_resources * br,struct bhndb_intr_handler * ih)931 bhndb_register_intr_handler(struct bhndb_resources *br,
932 struct bhndb_intr_handler *ih)
933 {
934 KASSERT(!ih->ih_active, ("duplicate registration of interrupt "
935 "handler %p", ih->ih_cookiep));
936 KASSERT(ih->ih_cookiep != NULL, ("missing cookiep"));
937
938 ih->ih_active = true;
939 STAILQ_INSERT_HEAD(&br->bus_intrs, ih, ih_link);
940 }
941
942 /**
943 * Remove an interrupt handler from the given resource state.
944 *
945 * @param br The resource state containing @p ih.
946 * @param ih The interrupt handler entry to be removed.
947 */
948 void
bhndb_deregister_intr_handler(struct bhndb_resources * br,struct bhndb_intr_handler * ih)949 bhndb_deregister_intr_handler(struct bhndb_resources *br,
950 struct bhndb_intr_handler *ih)
951 {
952 KASSERT(ih->ih_active, ("duplicate deregistration of interrupt "
953 "handler %p", ih->ih_cookiep));
954
955 KASSERT(bhndb_find_intr_handler(br, ih) == ih,
956 ("unknown interrupt handler %p", ih));
957
958 STAILQ_REMOVE(&br->bus_intrs, ih, bhndb_intr_handler, ih_link);
959 ih->ih_active = false;
960 }
961
962 /**
963 * Return the interrupt handler entry corresponding to @p cookiep, or NULL
964 * if no entry is found.
965 *
966 * @param br The resource state to search for the given @p cookiep.
967 * @param cookiep The interrupt handler's bus-assigned cookiep value.
968 */
969 struct bhndb_intr_handler *
bhndb_find_intr_handler(struct bhndb_resources * br,void * cookiep)970 bhndb_find_intr_handler(struct bhndb_resources *br, void *cookiep)
971 {
972 struct bhndb_intr_handler *ih;
973
974 STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
975 if (ih == cookiep)
976 return (ih);
977 }
978
979 /* Not found */
980 return (NULL);
981 }
982
983 /**
984 * Find the maximum start and end limits of the bridged resource @p r.
985 *
986 * If the resource is not currently mapped by the bridge, ENOENT will be
987 * returned.
988 *
989 * @param br The resource state to search.
990 * @param r The resource to search for in @p br.
991 * @param[out] start On success, the minimum supported start address.
992 * @param[out] end On success, the maximum supported end address.
993 *
994 * @retval 0 success
995 * @retval ENOENT no active mapping found for @p r of @p type
996 */
997 int
bhndb_find_resource_limits(struct bhndb_resources * br,struct resource * r,rman_res_t * start,rman_res_t * end)998 bhndb_find_resource_limits(struct bhndb_resources *br,
999 struct resource *r, rman_res_t *start, rman_res_t *end)
1000 {
1001 struct bhndb_dw_alloc *dynamic;
1002 struct bhndb_region *sregion;
1003 struct bhndb_intr_handler *ih;
1004
1005 switch (rman_get_type(r)) {
1006 case SYS_RES_IRQ:
1007 /* Is this one of ours? */
1008 STAILQ_FOREACH(ih, &br->bus_intrs, ih_link) {
1009 if (ih->ih_res == r)
1010 continue;
1011
1012 /* We don't support adjusting IRQ resource limits */
1013 *start = rman_get_start(r);
1014 *end = rman_get_end(r);
1015 return (0);
1016 }
1017
1018 /* Not found */
1019 return (ENOENT);
1020
1021 case SYS_RES_MEMORY: {
1022 /* Check for an enclosing dynamic register window */
1023 if ((dynamic = bhndb_dw_find_resource(br, r))) {
1024 *start = dynamic->target;
1025 *end = dynamic->target + dynamic->win->win_size - 1;
1026 return (0);
1027 }
1028
1029 /* Check for a static region */
1030 sregion = bhndb_find_resource_region(br, rman_get_start(r),
1031 rman_get_size(r));
1032 if (sregion != NULL && sregion->static_regwin != NULL) {
1033 *start = sregion->addr;
1034 *end = sregion->addr + sregion->size - 1;
1035
1036 return (0);
1037 }
1038
1039 /* Not found */
1040 return (ENOENT);
1041 }
1042
1043 default:
1044 device_printf(br->dev, "unknown resource type: %d\n",
1045 rman_get_type(r));
1046 return (ENOENT);
1047 }
1048 }
1049
1050 /**
1051 * Add a bus region entry to @p r for the given base @p addr and @p size.
1052 *
1053 * @param br The resource state to which the bus region entry will be added.
1054 * @param addr The base address of this region.
1055 * @param size The size of this region.
1056 * @param priority The resource priority to be assigned to allocations
1057 * made within this bus region.
1058 * @param alloc_flags resource allocation flags (@see bhndb_alloc_flags)
1059 * @param static_regwin If available, a static register window mapping this
1060 * bus region entry. If not available, NULL.
1061 *
1062 * @retval 0 success
1063 * @retval non-zero if adding the bus region fails.
1064 */
1065 int
bhndb_add_resource_region(struct bhndb_resources * br,bhnd_addr_t addr,bhnd_size_t size,bhndb_priority_t priority,uint32_t alloc_flags,const struct bhndb_regwin * static_regwin)1066 bhndb_add_resource_region(struct bhndb_resources *br, bhnd_addr_t addr,
1067 bhnd_size_t size, bhndb_priority_t priority, uint32_t alloc_flags,
1068 const struct bhndb_regwin *static_regwin)
1069 {
1070 struct bhndb_region *reg;
1071
1072 /* Insert in the bus resource list */
1073 reg = malloc(sizeof(*reg), M_BHND, M_NOWAIT);
1074 if (reg == NULL)
1075 return (ENOMEM);
1076
1077 *reg = (struct bhndb_region) {
1078 .addr = addr,
1079 .size = size,
1080 .priority = priority,
1081 .alloc_flags = alloc_flags,
1082 .static_regwin = static_regwin
1083 };
1084
1085 STAILQ_INSERT_HEAD(&br->bus_regions, reg, link);
1086
1087 return (0);
1088 }
1089
1090 /**
1091 * Return true if a mapping of @p size bytes at @p addr is provided by either
1092 * one contiguous bus region, or by multiple discontiguous regions.
1093 *
1094 * @param br The resource state to query.
1095 * @param addr The requested starting address.
1096 * @param size The requested size.
1097 */
1098 bool
bhndb_has_static_region_mapping(struct bhndb_resources * br,bhnd_addr_t addr,bhnd_size_t size)1099 bhndb_has_static_region_mapping(struct bhndb_resources *br,
1100 bhnd_addr_t addr, bhnd_size_t size)
1101 {
1102 struct bhndb_region *region;
1103 bhnd_addr_t r_addr;
1104
1105 r_addr = addr;
1106 while ((region = bhndb_find_resource_region(br, r_addr, 1)) != NULL) {
1107 /* Must be backed by a static register window */
1108 if (region->static_regwin == NULL)
1109 return (false);
1110
1111 /* Adjust the search offset */
1112 r_addr += region->size;
1113
1114 /* Have we traversed a complete (if discontiguous) mapping? */
1115 if (r_addr == addr + size)
1116 return (true);
1117 }
1118
1119 /* No complete mapping found */
1120 return (false);
1121 }
1122
1123 /**
1124 * Find the bus region that maps @p size bytes at @p addr.
1125 *
1126 * @param br The resource state to search.
1127 * @param addr The requested starting address.
1128 * @param size The requested size.
1129 *
1130 * @retval bhndb_region A region that fully contains the requested range.
1131 * @retval NULL If no mapping region can be found.
1132 */
1133 struct bhndb_region *
bhndb_find_resource_region(struct bhndb_resources * br,bhnd_addr_t addr,bhnd_size_t size)1134 bhndb_find_resource_region(struct bhndb_resources *br, bhnd_addr_t addr,
1135 bhnd_size_t size)
1136 {
1137 struct bhndb_region *region;
1138
1139 STAILQ_FOREACH(region, &br->bus_regions, link) {
1140 /* Request must fit within the region's mapping */
1141 if (addr < region->addr)
1142 continue;
1143
1144 if (addr + size > region->addr + region->size)
1145 continue;
1146
1147 return (region);
1148 }
1149
1150 /* Not found */
1151 return (NULL);
1152 }
1153
1154 /**
1155 * Find the entry matching @p r in @p dwa's references, if any.
1156 *
1157 * @param dwa The dynamic window allocation to search
1158 * @param r The resource to search for in @p dwa.
1159 */
1160 static struct bhndb_dw_rentry *
bhndb_dw_find_resource_entry(struct bhndb_dw_alloc * dwa,struct resource * r)1161 bhndb_dw_find_resource_entry(struct bhndb_dw_alloc *dwa, struct resource *r)
1162 {
1163 struct bhndb_dw_rentry *rentry;
1164
1165 LIST_FOREACH(rentry, &dwa->refs, dw_link) {
1166 struct resource *dw_res = rentry->dw_res;
1167
1168 /* Match dev/rid/addr/size */
1169 if (rman_get_device(dw_res) != rman_get_device(r) ||
1170 rman_get_rid(dw_res) != rman_get_rid(r) ||
1171 rman_get_start(dw_res) != rman_get_start(r) ||
1172 rman_get_size(dw_res) != rman_get_size(r))
1173 {
1174 continue;
1175 }
1176
1177 /* Matching allocation found */
1178 return (rentry);
1179 }
1180
1181 return (NULL);
1182 }
1183
1184 /**
1185 * Find the dynamic region allocated for @p r, if any.
1186 *
1187 * @param br The resource state to search.
1188 * @param r The resource to search for.
1189 *
1190 * @retval bhndb_dw_alloc The allocation record for @p r.
1191 * @retval NULL if no dynamic window is allocated for @p r.
1192 */
1193 struct bhndb_dw_alloc *
bhndb_dw_find_resource(struct bhndb_resources * br,struct resource * r)1194 bhndb_dw_find_resource(struct bhndb_resources *br, struct resource *r)
1195 {
1196 struct bhndb_dw_alloc *dwa;
1197
1198 for (size_t i = 0; i < br->dwa_count; i++) {
1199 dwa = &br->dw_alloc[i];
1200
1201 /* Skip free dynamic windows */
1202 if (bhndb_dw_is_free(br, dwa))
1203 continue;
1204
1205 /* Matching allocation found? */
1206 if (bhndb_dw_find_resource_entry(dwa, r) != NULL)
1207 return (dwa);
1208 }
1209
1210 return (NULL);
1211 }
1212
1213 /**
1214 * Find an existing dynamic window mapping @p size bytes
1215 * at @p addr. The window may or may not be free.
1216 *
1217 * @param br The resource state to search.
1218 * @param addr The requested starting address.
1219 * @param size The requested size.
1220 *
1221 * @retval bhndb_dw_alloc A window allocation that fully contains the requested
1222 * range.
1223 * @retval NULL If no mapping region can be found.
1224 */
1225 struct bhndb_dw_alloc *
bhndb_dw_find_mapping(struct bhndb_resources * br,bhnd_addr_t addr,bhnd_size_t size)1226 bhndb_dw_find_mapping(struct bhndb_resources *br, bhnd_addr_t addr,
1227 bhnd_size_t size)
1228 {
1229 struct bhndb_dw_alloc *dwr;
1230 const struct bhndb_regwin *win;
1231
1232 /* Search for an existing dynamic mapping of this address range. */
1233 for (size_t i = 0; i < br->dwa_count; i++) {
1234 dwr = &br->dw_alloc[i];
1235 win = dwr->win;
1236
1237 /* Verify the range */
1238 if (addr < dwr->target)
1239 continue;
1240
1241 if (addr + size > dwr->target + win->win_size)
1242 continue;
1243
1244 /* Found a usable mapping */
1245 return (dwr);
1246 }
1247
1248 /* not found */
1249 return (NULL);
1250 }
1251
1252 /**
1253 * Retain a reference to @p dwa for use by @p res.
1254 *
1255 * @param br The resource state owning @p dwa.
1256 * @param dwa The allocation record to be retained.
1257 * @param res The resource that will own a reference to @p dwa.
1258 *
1259 * @retval 0 success
1260 * @retval ENOMEM Failed to allocate a new reference structure.
1261 */
1262 int
bhndb_dw_retain(struct bhndb_resources * br,struct bhndb_dw_alloc * dwa,struct resource * res)1263 bhndb_dw_retain(struct bhndb_resources *br, struct bhndb_dw_alloc *dwa,
1264 struct resource *res)
1265 {
1266 struct bhndb_dw_rentry *rentry;
1267
1268 KASSERT(bhndb_dw_find_resource_entry(dwa, res) == NULL,
1269 ("double-retain of dynamic window for same resource"));
1270
1271 /* Insert a reference entry; we use M_NOWAIT to allow use from
1272 * within a non-sleepable lock */
1273 rentry = malloc(sizeof(*rentry), M_BHND, M_NOWAIT);
1274 if (rentry == NULL)
1275 return (ENOMEM);
1276
1277 rentry->dw_res = res;
1278 LIST_INSERT_HEAD(&dwa->refs, rentry, dw_link);
1279
1280 /* Update the free list */
1281 bit_set(br->dwa_freelist, dwa->rnid);
1282
1283 return (0);
1284 }
1285
1286 /**
1287 * Release a reference to @p dwa previously retained by @p res. If the
1288 * reference count of @p dwa reaches zero, it will be added to the
1289 * free list.
1290 *
1291 * @param br The resource state owning @p dwa.
1292 * @param dwa The allocation record to be released.
1293 * @param res The resource that currently owns a reference to @p dwa.
1294 */
1295 void
bhndb_dw_release(struct bhndb_resources * br,struct bhndb_dw_alloc * dwa,struct resource * r)1296 bhndb_dw_release(struct bhndb_resources *br, struct bhndb_dw_alloc *dwa,
1297 struct resource *r)
1298 {
1299 struct bhndb_dw_rentry *rentry;
1300
1301 /* Find the rentry */
1302 rentry = bhndb_dw_find_resource_entry(dwa, r);
1303 KASSERT(rentry != NULL, ("over release of resource entry"));
1304
1305 LIST_REMOVE(rentry, dw_link);
1306 free(rentry, M_BHND);
1307
1308 /* If this was the last reference, update the free list */
1309 if (LIST_EMPTY(&dwa->refs))
1310 bit_clear(br->dwa_freelist, dwa->rnid);
1311 }
1312
1313 /**
1314 * Attempt to set (or reset) the target address of @p dwa to map @p size bytes
1315 * at @p addr.
1316 *
1317 * This will apply any necessary window alignment and verify that
1318 * the window is capable of mapping the requested range prior to modifying
1319 * therecord.
1320 *
1321 * @param dev The device on which to issue the BHNDB_SET_WINDOW_ADDR() request.
1322 * @param br The resource state owning @p dwa.
1323 * @param dwa The allocation record to be configured.
1324 * @param addr The address to be mapped via @p dwa.
1325 * @param size The number of bytes to be mapped at @p addr.
1326 *
1327 * @retval 0 success
1328 * @retval non-zero no usable register window available.
1329 */
1330 int
bhndb_dw_set_addr(device_t dev,struct bhndb_resources * br,struct bhndb_dw_alloc * dwa,bus_addr_t addr,bus_size_t size)1331 bhndb_dw_set_addr(device_t dev, struct bhndb_resources *br,
1332 struct bhndb_dw_alloc *dwa, bus_addr_t addr, bus_size_t size)
1333 {
1334 const struct bhndb_regwin *rw;
1335 bus_addr_t offset;
1336 int error;
1337
1338 rw = dwa->win;
1339
1340 KASSERT(bhndb_dw_is_free(br, dwa) || mtx_owned(&br->dw_steal_mtx),
1341 ("attempting to set the target address on an in-use window"));
1342
1343 /* Page-align the target address */
1344 offset = addr % rw->win_size;
1345 dwa->target = addr - offset;
1346
1347 /* Verify that the window is large enough for the full target */
1348 if (rw->win_size - offset < size)
1349 return (ENOMEM);
1350
1351 /* Update the window target */
1352 error = BHNDB_SET_WINDOW_ADDR(dev, dwa->win, dwa->target);
1353 if (error) {
1354 dwa->target = 0x0;
1355 return (error);
1356 }
1357
1358 return (0);
1359 }
1360
1361 /**
1362 * Steal an in-use allocation record from @p br, returning the record's current
1363 * target in @p saved on success.
1364 *
1365 * This function acquires a mutex and disables interrupts; callers should
1366 * avoid holding a stolen window longer than required to issue an I/O
1367 * request.
1368 *
1369 * A successful call to bhndb_dw_steal() must be balanced with a call to
1370 * bhndb_dw_return_stolen().
1371 *
1372 * @param br The resource state from which a window should be stolen.
1373 * @param saved The stolen window's saved target address.
1374 *
1375 * @retval non-NULL success
1376 * @retval NULL no dynamic window regions are defined.
1377 */
1378 struct bhndb_dw_alloc *
bhndb_dw_steal(struct bhndb_resources * br,bus_addr_t * saved)1379 bhndb_dw_steal(struct bhndb_resources *br, bus_addr_t *saved)
1380 {
1381 struct bhndb_dw_alloc *dw_stolen;
1382
1383 KASSERT(bhndb_dw_next_free(br) == NULL,
1384 ("attempting to steal an in-use window while free windows remain"));
1385
1386 /* Nothing to steal from? */
1387 if (br->dwa_count == 0)
1388 return (NULL);
1389
1390 /*
1391 * Acquire our steal spinlock; this will be released in
1392 * bhndb_dw_return_stolen().
1393 *
1394 * Acquiring also disables interrupts, which is required when one is
1395 * stealing an in-use existing register window.
1396 */
1397 mtx_lock_spin(&br->dw_steal_mtx);
1398
1399 dw_stolen = &br->dw_alloc[0];
1400 *saved = dw_stolen->target;
1401 return (dw_stolen);
1402 }
1403
1404 /**
1405 * Return an allocation record previously stolen using bhndb_dw_steal().
1406 *
1407 * @param dev The device on which to issue a BHNDB_SET_WINDOW_ADDR() request.
1408 * @param br The resource state owning @p dwa.
1409 * @param dwa The allocation record to be returned.
1410 * @param saved The original target address provided by bhndb_dw_steal().
1411 */
1412 void
bhndb_dw_return_stolen(device_t dev,struct bhndb_resources * br,struct bhndb_dw_alloc * dwa,bus_addr_t saved)1413 bhndb_dw_return_stolen(device_t dev, struct bhndb_resources *br,
1414 struct bhndb_dw_alloc *dwa, bus_addr_t saved)
1415 {
1416 int error;
1417
1418 mtx_assert(&br->dw_steal_mtx, MA_OWNED);
1419
1420 error = bhndb_dw_set_addr(dev, br, dwa, saved, 0);
1421 if (error) {
1422 panic("failed to restore register window target %#jx: %d\n",
1423 (uintmax_t)saved, error);
1424 }
1425
1426 mtx_unlock_spin(&br->dw_steal_mtx);
1427 }
1428
1429 /**
1430 * Return the count of @p type register windows in @p table.
1431 *
1432 * @param table The table to search.
1433 * @param type The required window type, or BHNDB_REGWIN_T_INVALID to
1434 * count all register window types.
1435 */
1436 size_t
bhndb_regwin_count(const struct bhndb_regwin * table,bhndb_regwin_type_t type)1437 bhndb_regwin_count(const struct bhndb_regwin *table,
1438 bhndb_regwin_type_t type)
1439 {
1440 const struct bhndb_regwin *rw;
1441 size_t count;
1442
1443 count = 0;
1444 for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++) {
1445 if (type == BHNDB_REGWIN_T_INVALID || rw->win_type == type)
1446 count++;
1447 }
1448
1449 return (count);
1450 }
1451
1452 /**
1453 * Search @p table for the first window with the given @p type.
1454 *
1455 * @param table The table to search.
1456 * @param type The required window type.
1457 * @param min_size The minimum window size.
1458 *
1459 * @retval bhndb_regwin The first matching window.
1460 * @retval NULL If no window of the requested type could be found.
1461 */
1462 const struct bhndb_regwin *
bhndb_regwin_find_type(const struct bhndb_regwin * table,bhndb_regwin_type_t type,bus_size_t min_size)1463 bhndb_regwin_find_type(const struct bhndb_regwin *table,
1464 bhndb_regwin_type_t type, bus_size_t min_size)
1465 {
1466 const struct bhndb_regwin *rw;
1467
1468 for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++)
1469 {
1470 if (rw->win_type == type && rw->win_size >= min_size)
1471 return (rw);
1472 }
1473
1474 return (NULL);
1475 }
1476
1477 /**
1478 * Search @p windows for the first matching core window.
1479 *
1480 * @param table The table to search.
1481 * @param class The required core class.
1482 * @param unit The required core unit, or -1.
1483 * @param port_type The required port type.
1484 * @param port The required port.
1485 * @param region The required region.
1486 * @param offset The required readable core register block offset.
1487 * @param min_size The required minimum readable size at @p offset.
1488 *
1489 * @retval bhndb_regwin The first matching window.
1490 * @retval NULL If no matching window was found.
1491 */
1492 const struct bhndb_regwin *
bhndb_regwin_find_core(const struct bhndb_regwin * table,bhnd_devclass_t class,int unit,bhnd_port_type port_type,u_int port,u_int region,bus_size_t offset,bus_size_t min_size)1493 bhndb_regwin_find_core(const struct bhndb_regwin *table, bhnd_devclass_t class,
1494 int unit, bhnd_port_type port_type, u_int port, u_int region,
1495 bus_size_t offset, bus_size_t min_size)
1496 {
1497 const struct bhndb_regwin *rw;
1498
1499 for (rw = table; rw->win_type != BHNDB_REGWIN_T_INVALID; rw++)
1500 {
1501 bus_size_t rw_offset;
1502
1503 /* Match on core, port, and region attributes */
1504 if (rw->win_type != BHNDB_REGWIN_T_CORE)
1505 continue;
1506
1507 if (rw->d.core.class != class)
1508 continue;
1509
1510 if (unit != -1 && rw->d.core.unit != unit)
1511 continue;
1512
1513 if (rw->d.core.port_type != port_type)
1514 continue;
1515
1516 if (rw->d.core.port != port)
1517 continue;
1518
1519 if (rw->d.core.region != region)
1520 continue;
1521
1522 /* Verify that the requested range is mapped within
1523 * this register window */
1524 if (rw->d.core.offset > offset)
1525 continue;
1526
1527 rw_offset = offset - rw->d.core.offset;
1528
1529 if (rw->win_size < rw_offset)
1530 continue;
1531
1532 if (rw->win_size - rw_offset < min_size)
1533 continue;
1534
1535 return (rw);
1536 }
1537
1538 return (NULL);
1539 }
1540
1541 /**
1542 * Search @p windows for the best available window of at least @p min_size.
1543 *
1544 * Search order:
1545 * - BHND_REGWIN_T_CORE
1546 * - BHND_REGWIN_T_DYN
1547 *
1548 * @param table The table to search.
1549 * @param class The required core class.
1550 * @param unit The required core unit, or -1.
1551 * @param port_type The required port type.
1552 * @param port The required port.
1553 * @param region The required region.
1554 * @param offset The required readable core register block offset.
1555 * @param min_size The required minimum readable size at @p offset.
1556 *
1557 * @retval bhndb_regwin The first matching window.
1558 * @retval NULL If no matching window was found.
1559 */
1560 const struct bhndb_regwin *
bhndb_regwin_find_best(const struct bhndb_regwin * table,bhnd_devclass_t class,int unit,bhnd_port_type port_type,u_int port,u_int region,bus_size_t offset,bus_size_t min_size)1561 bhndb_regwin_find_best(const struct bhndb_regwin *table,
1562 bhnd_devclass_t class, int unit, bhnd_port_type port_type, u_int port,
1563 u_int region, bus_size_t offset, bus_size_t min_size)
1564 {
1565 const struct bhndb_regwin *rw;
1566
1567 /* Prefer a fixed core mapping */
1568 rw = bhndb_regwin_find_core(table, class, unit, port_type,
1569 port, region, offset, min_size);
1570 if (rw != NULL)
1571 return (rw);
1572
1573 /* Fall back on a generic dynamic window */
1574 return (bhndb_regwin_find_type(table, BHNDB_REGWIN_T_DYN, min_size));
1575 }
1576
1577 /**
1578 * Return true if @p regw defines a BHNDB_REGWIN_T_CORE register window
1579 * that matches against @p core.
1580 *
1581 * @param regw A register window to match against.
1582 * @param core The bhnd(4) core info to match against @p regw.
1583 */
1584 bool
bhndb_regwin_match_core(const struct bhndb_regwin * regw,struct bhnd_core_info * core)1585 bhndb_regwin_match_core(const struct bhndb_regwin *regw,
1586 struct bhnd_core_info *core)
1587 {
1588 /* Only core windows are supported */
1589 if (regw->win_type != BHNDB_REGWIN_T_CORE)
1590 return (false);
1591
1592 /* Device class must match */
1593 if (bhnd_core_class(core) != regw->d.core.class)
1594 return (false);
1595
1596 /* Device unit must match */
1597 if (core->unit != regw->d.core.unit)
1598 return (false);
1599
1600 /* Matches */
1601 return (true);
1602 }
1603
1604 /**
1605 * Search for a core resource priority descriptor in @p table that matches
1606 * @p core.
1607 *
1608 * @param table The table to search.
1609 * @param core The core to match against @p table.
1610 */
1611 const struct bhndb_hw_priority *
bhndb_hw_priority_find_core(const struct bhndb_hw_priority * table,struct bhnd_core_info * core)1612 bhndb_hw_priority_find_core(const struct bhndb_hw_priority *table,
1613 struct bhnd_core_info *core)
1614 {
1615 const struct bhndb_hw_priority *hp;
1616
1617 for (hp = table; hp->ports != NULL; hp++) {
1618 if (bhnd_core_matches(core, &hp->match))
1619 return (hp);
1620 }
1621
1622 /* not found */
1623 return (NULL);
1624 }
1625
1626 /**
1627 * Search for a port resource priority descriptor in @p table.
1628 *
1629 * @param table The table to search.
1630 * @param core The core to match against @p table.
1631 * @param port_type The required port type.
1632 * @param port The required port.
1633 * @param region The required region.
1634 */
1635 const struct bhndb_port_priority *
bhndb_hw_priorty_find_port(const struct bhndb_hw_priority * table,struct bhnd_core_info * core,bhnd_port_type port_type,u_int port,u_int region)1636 bhndb_hw_priorty_find_port(const struct bhndb_hw_priority *table,
1637 struct bhnd_core_info *core, bhnd_port_type port_type, u_int port,
1638 u_int region)
1639 {
1640 const struct bhndb_hw_priority *hp;
1641
1642 if ((hp = bhndb_hw_priority_find_core(table, core)) == NULL)
1643 return (NULL);
1644
1645 for (u_int i = 0; i < hp->num_ports; i++) {
1646 const struct bhndb_port_priority *pp = &hp->ports[i];
1647
1648 if (pp->type != port_type)
1649 continue;
1650
1651 if (pp->port != port)
1652 continue;
1653
1654 if (pp->region != region)
1655 continue;
1656
1657 return (pp);
1658 }
1659
1660 /* not found */
1661 return (NULL);
1662 }
1663