xref: /illumos-gate/usr/src/uts/common/io/busra.c (revision 4df55fde49134f9735f84011f23a767c75e393c7)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 
26 #if defined(DEBUG)
27 #define	BUSRA_DEBUG
28 #endif
29 
30 /*
31  * This module provides a set of resource management interfaces
32  * to manage bus resources globally in the system.
33  *
34  * The bus nexus drivers are typically responsible to setup resource
35  * maps for the bus resources available for a bus instance. However
36  * this module also provides resource setup functions for PCI bus
37  * (used by both SPARC and X86 platforms) and ISA bus instances (used
38  * only for X86 platforms).
39  */
40 
41 #include <sys/types.h>
42 #include <sys/systm.h>
43 #include <sys/ddi.h>
44 #include <sys/sunddi.h>
45 #include <sys/sunndi.h>
46 #include <sys/ddi_impldefs.h>
47 #include <sys/ndi_impldefs.h>
48 #include <sys/kmem.h>
49 #include <sys/pctypes.h>
50 #include <sys/modctl.h>
51 #include <sys/debug.h>
52 #include <sys/spl.h>
53 #include <sys/pci.h>
54 #include <sys/autoconf.h>
55 
56 #if defined(BUSRA_DEBUG)
57 int busra_debug = 0;
58 #define	DEBUGPRT \
59 	if (busra_debug) cmn_err
60 
61 #else
62 #define	DEBUGPRT \
63 	if (0) cmn_err
64 #endif
65 
66 
67 /*
68  * global mutex that protects the global list of resource maps.
69  */
70 kmutex_t ra_lock;
71 
72 /*
73  * basic resource element
74  */
75 struct ra_resource {
76 	struct ra_resource *ra_next;
77 	uint64_t	ra_base;
78 	uint64_t 	ra_len;
79 };
80 
81 /*
82  * link list element for the list of dips (and their resource ranges)
83  * for a particular resource type.
84  * ra_rangeset points to the list of resources available
85  * for this type and this dip.
86  */
87 struct ra_dip_type  {
88 	struct ra_dip_type *ra_next;
89 	struct ra_resource  *ra_rangeset;
90 	dev_info_t *ra_dip;
91 };
92 
93 
94 /*
95  * link list element for list of types resources. Each element
96  * has all resources for a particular type.
97  */
98 struct ra_type_map {
99 	struct ra_type_map *ra_next;
100 	struct ra_dip_type *ra_dip_list;
101 	char *type;
102 };
103 
104 
105 /*
106  * place holder to keep the head of the whole global list.
107  * the address of the first typemap would be stored in it.
108  */
109 static struct ra_type_map	*ra_map_list_head = NULL;
110 
111 
112 /*
113  * This is the loadable module wrapper.
114  * It is essentially boilerplate so isn't documented
115  */
116 extern struct mod_ops mod_miscops;
117 
118 #ifdef BUSRA_DEBUG
119 void ra_dump_all();
120 #endif
121 
122 /* internal function prototypes */
123 static struct ra_dip_type *find_dip_map_resources(dev_info_t *dip, char *type,
124     struct ra_dip_type ***backdip, struct ra_type_map ***backtype,
125     uint32_t flag);
126 static int isnot_pow2(uint64_t value);
127 static int claim_pci_busnum(dev_info_t *dip, void *arg);
128 static int ra_map_exist(dev_info_t *dip, char *type);
129 
130 static int pci_get_available_prop(dev_info_t *dip, uint64_t base,
131     uint64_t len, char *busra_type);
132 static int pci_put_available_prop(dev_info_t *dip, uint64_t base,
133     uint64_t len, char *busra_type);
134 static uint32_t pci_type_ra2pci(char *type);
135 static boolean_t is_pcie_fabric(dev_info_t *dip);
136 
137 #define	PCI_ADDR_TYPE_MASK	(PCI_REG_ADDR_M | PCI_REG_PF_M)
138 #define	PCI_ADDR_TYPE_INVAL	0xffffffff
139 
140 #define	RA_INSERT(prev, el) \
141 	el->ra_next = *prev; \
142 	*prev = el;
143 
144 #define	RA_REMOVE(prev, el) \
145 	*prev = el->ra_next;
146 
147 
148 static struct modlmisc modlmisc = {
149 	&mod_miscops,		/* Type of module. This one is a module */
150 	"Bus Resource Allocator (BUSRA)",	/* Name of the module. */
151 };
152 
153 static struct modlinkage modlinkage = {
154 	MODREV_1, (void *)&modlmisc, NULL
155 };
156 
157 int
158 _init()
159 {
160 	int	ret;
161 
162 	mutex_init(&ra_lock, NULL, MUTEX_DRIVER,
163 	    (void *)(intptr_t)__ipltospl(SPL7 - 1));
164 	if ((ret = mod_install(&modlinkage)) != 0) {
165 		mutex_destroy(&ra_lock);
166 	}
167 	return (ret);
168 }
169 
170 int
171 _fini()
172 {
173 	int	ret;
174 
175 	mutex_enter(&ra_lock);
176 
177 	if (ra_map_list_head != NULL) {
178 		mutex_exit(&ra_lock);
179 		return (EBUSY);
180 	}
181 
182 	ret = mod_remove(&modlinkage);
183 
184 	mutex_exit(&ra_lock);
185 
186 	if (ret == 0)
187 		mutex_destroy(&ra_lock);
188 
189 	return (ret);
190 }
191 
192 int
193 _info(struct modinfo *modinfop)
194 
195 {
196 	return (mod_info(&modlinkage, modinfop));
197 }
198 
199 /*
200  * set up an empty resource map for a given type and dip
201  */
202 int
203 ndi_ra_map_setup(dev_info_t *dip, char *type)
204 {
205 	struct ra_type_map  *typemapp;
206 	struct ra_dip_type  *dipmap;
207 	struct ra_dip_type  **backdip;
208 	struct ra_type_map  **backtype;
209 
210 
211 	mutex_enter(&ra_lock);
212 
213 	dipmap = find_dip_map_resources(dip, type, &backdip, &backtype, 0);
214 
215 	if (dipmap == NULL) {
216 		if (backtype == NULL) {
217 			typemapp = (struct ra_type_map *)
218 			    kmem_zalloc(sizeof (*typemapp), KM_SLEEP);
219 			typemapp->type = (char *)kmem_zalloc(strlen(type) + 1,
220 			    KM_SLEEP);
221 			(void) strcpy(typemapp->type, type);
222 			RA_INSERT(&ra_map_list_head, typemapp);
223 		} else {
224 			typemapp = *backtype;
225 		}
226 		if (backdip == NULL) {
227 			/* allocate and insert in list of dips for this type */
228 			dipmap = (struct ra_dip_type *)
229 			    kmem_zalloc(sizeof (*dipmap), KM_SLEEP);
230 			dipmap->ra_dip = dip;
231 			RA_INSERT(&typemapp->ra_dip_list, dipmap);
232 		}
233 	}
234 
235 	mutex_exit(&ra_lock);
236 	return (NDI_SUCCESS);
237 }
238 
239 /*
240  * destroys a resource map for a given dip and type
241  */
242 int
243 ndi_ra_map_destroy(dev_info_t *dip, char *type)
244 {
245 	struct ra_dip_type	*dipmap;
246 	struct ra_dip_type	**backdip;
247 	struct ra_type_map  	**backtype, *typemap;
248 	struct ra_resource	*range;
249 
250 	mutex_enter(&ra_lock);
251 	dipmap = find_dip_map_resources(dip, type, &backdip, &backtype, 0);
252 
253 	if (dipmap == NULL) {
254 		mutex_exit(&ra_lock);
255 		return (NDI_FAILURE);
256 	}
257 
258 	/*
259 	 * destroy all resources for this dip
260 	 * remove dip from type list
261 	 */
262 	ASSERT((backdip != NULL) && (backtype != NULL));
263 	while (dipmap->ra_rangeset != NULL) {
264 		range = dipmap->ra_rangeset;
265 		RA_REMOVE(&dipmap->ra_rangeset, range);
266 		kmem_free((caddr_t)range, sizeof (*range));
267 	}
268 	/* remove from dip list */
269 	RA_REMOVE(backdip, dipmap);
270 	kmem_free((caddr_t)dipmap, sizeof (*dipmap));
271 	if ((*backtype)->ra_dip_list == NULL) {
272 		/*
273 		 * This was the last dip with this resource type.
274 		 * Remove the type from the global list.
275 		 */
276 		typemap = *backtype;
277 		RA_REMOVE(backtype, (*backtype));
278 		kmem_free((caddr_t)typemap->type, strlen(typemap->type) + 1);
279 		kmem_free((caddr_t)typemap, sizeof (*typemap));
280 	}
281 
282 	mutex_exit(&ra_lock);
283 	return (NDI_SUCCESS);
284 }
285 
286 static int
287 ra_map_exist(dev_info_t *dip, char *type)
288 {
289 	struct ra_dip_type  **backdip;
290 	struct ra_type_map  **backtype;
291 
292 	mutex_enter(&ra_lock);
293 	if (find_dip_map_resources(dip, type, &backdip, &backtype, 0) == NULL) {
294 		mutex_exit(&ra_lock);
295 		return (NDI_FAILURE);
296 	}
297 
298 	mutex_exit(&ra_lock);
299 	return (NDI_SUCCESS);
300 }
301 /*
302  * Find a dip map for the specified type, if NDI_RA_PASS will go up on dev tree
303  * if found, backdip and backtype will be updated to point to the previous
304  * dip in the list and previous type for this dip in the list.
305  * If no such type at all in the resource list both backdip and backtype
306  * will be null. If the type found but no dip, back dip will be null.
307  */
308 
309 static struct ra_dip_type *
310 find_dip_map_resources(dev_info_t *dip, char *type,
311     struct ra_dip_type ***backdip, struct ra_type_map ***backtype,
312     uint32_t flag)
313 {
314 	struct ra_type_map **prevmap;
315 	struct ra_dip_type *dipmap, **prevdip;
316 
317 	ASSERT(mutex_owned(&ra_lock));
318 	prevdip = NULL;
319 	dipmap = NULL;
320 	prevmap = &ra_map_list_head;
321 
322 	while (*prevmap) {
323 		if (strcmp((*prevmap)->type, type) == 0)
324 			break;
325 		prevmap = &(*prevmap)->ra_next;
326 	}
327 
328 	if (*prevmap) {
329 		for (; dip != NULL; dip = ddi_get_parent(dip)) {
330 			prevdip = &(*prevmap)->ra_dip_list;
331 			dipmap = *prevdip;
332 
333 			while (dipmap) {
334 				if (dipmap->ra_dip == dip)
335 					break;
336 				prevdip =  &dipmap->ra_next;
337 				dipmap = dipmap->ra_next;
338 			}
339 
340 			if (dipmap != NULL) {
341 				/* found it */
342 				break;
343 			}
344 
345 			if (!(flag & NDI_RA_PASS)) {
346 				break;
347 			}
348 		}
349 	}
350 
351 	*backtype = (*prevmap == NULL) ?  NULL: prevmap;
352 	*backdip = (dipmap == NULL) ?  NULL: prevdip;
353 
354 	return (dipmap);
355 }
356 
357 int
358 ndi_ra_free(dev_info_t *dip, uint64_t base, uint64_t len, char *type,
359     uint32_t flag)
360 {
361 	struct ra_dip_type *dipmap;
362 	struct ra_resource *newmap, *overlapmap, *oldmap = NULL;
363 	struct ra_resource  *mapp, **backp;
364 	uint64_t newend, mapend;
365 	struct ra_dip_type **backdip;
366 	struct ra_type_map **backtype;
367 
368 	if (len == 0) {
369 		return (NDI_SUCCESS);
370 	}
371 
372 	mutex_enter(&ra_lock);
373 
374 	if ((dipmap = find_dip_map_resources(dip, type, &backdip, &backtype,
375 	    flag)) == NULL) {
376 		mutex_exit(&ra_lock);
377 		return (NDI_FAILURE);
378 	}
379 
380 	mapp = dipmap->ra_rangeset;
381 	backp = &dipmap->ra_rangeset;
382 
383 	/* now find where range lies and fix things up */
384 	newend = base + len;
385 	for (; mapp != NULL; backp = &(mapp->ra_next), mapp = mapp->ra_next) {
386 		mapend = mapp->ra_base + mapp->ra_len;
387 
388 		/* check for overlap first */
389 		if ((base <= mapp->ra_base && newend > mapp->ra_base) ||
390 		    (base > mapp->ra_base && base < mapend)) {
391 			/* overlap with mapp */
392 			overlapmap = mapp;
393 			goto overlap;
394 		} else if ((base == mapend && mapp->ra_next) &&
395 		    (newend > mapp->ra_next->ra_base)) {
396 			/* overlap with mapp->ra_next */
397 			overlapmap = mapp->ra_next;
398 			goto overlap;
399 		}
400 
401 		if (newend == mapp->ra_base) {
402 			/* simple - on front */
403 			mapp->ra_base = base;
404 			mapp->ra_len += len;
405 			/*
406 			 * don't need to check if it merges with
407 			 * previous since that would match on on end
408 			 */
409 			break;
410 		} else if (base == mapend) {
411 			/* simple - on end */
412 			mapp->ra_len += len;
413 			if (mapp->ra_next &&
414 			    (newend == mapp->ra_next->ra_base)) {
415 				/* merge with next node */
416 				oldmap = mapp->ra_next;
417 				mapp->ra_len += oldmap->ra_len;
418 				RA_REMOVE(&mapp->ra_next, oldmap);
419 				kmem_free((caddr_t)oldmap, sizeof (*oldmap));
420 			}
421 			break;
422 		} else if (base < mapp->ra_base) {
423 			/* somewhere in between so just an insert */
424 			newmap = (struct ra_resource *)
425 			    kmem_zalloc(sizeof (*newmap), KM_SLEEP);
426 			newmap->ra_base = base;
427 			newmap->ra_len = len;
428 			RA_INSERT(backp, newmap);
429 			break;
430 		}
431 	}
432 	if (mapp == NULL) {
433 		/* stick on end */
434 		newmap = (struct ra_resource *)
435 		    kmem_zalloc(sizeof (*newmap), KM_SLEEP);
436 		newmap->ra_base = base;
437 		newmap->ra_len = len;
438 		RA_INSERT(backp, newmap);
439 	}
440 
441 	mutex_exit(&ra_lock);
442 
443 	/*
444 	 * Update dip's "available" property, adding this piece of
445 	 * resource to the pool.
446 	 */
447 	(void) pci_put_available_prop(dipmap->ra_dip, base, len, type);
448 done:
449 	return (NDI_SUCCESS);
450 
451 overlap:
452 	/*
453 	 * Bad free may happen on some x86 platforms with BIOS exporting
454 	 * incorrect resource maps. The system is otherwise functioning
455 	 * normally. We send such messages to syslog only.
456 	 */
457 	cmn_err(CE_NOTE, "!ndi_ra_free: bad free, dip %p, resource type %s \n",
458 	    (void *)dip, type);
459 	cmn_err(CE_NOTE, "!ndi_ra_free: freeing base 0x%" PRIx64 ", len 0x%"
460 	    PRIX64 " overlaps with existing resource base 0x%" PRIx64
461 	    ", len 0x%" PRIx64 "\n", base, len, overlapmap->ra_base,
462 	    overlapmap->ra_len);
463 
464 	mutex_exit(&ra_lock);
465 	return (NDI_FAILURE);
466 }
467 
468 /* check to see if value is power of 2 or not. */
469 static int
470 isnot_pow2(uint64_t value)
471 {
472 	uint32_t low;
473 	uint32_t hi;
474 
475 	low = value & 0xffffffff;
476 	hi = value >> 32;
477 
478 	/*
479 	 * ddi_ffs and ddi_fls gets long values, so in 32bit environment
480 	 * won't work correctly for 64bit values
481 	 */
482 	if ((ddi_ffs(low) == ddi_fls(low)) &&
483 	    (ddi_ffs(hi) == ddi_fls(hi)))
484 		return (0);
485 	return (1);
486 }
487 
488 static  void
489 adjust_link(struct ra_resource **backp, struct ra_resource *mapp,
490 	    uint64_t base, uint64_t len)
491 {
492 	struct ra_resource *newmap;
493 	uint64_t newlen;
494 
495 	if (base != mapp->ra_base) {
496 		/* in the middle or end */
497 		newlen = base - mapp->ra_base;
498 		if ((mapp->ra_len - newlen) == len) {
499 			/* on the end */
500 			mapp->ra_len = newlen;
501 		} else {
502 			/* in the middle */
503 			newmap = (struct ra_resource *)
504 			    kmem_zalloc(sizeof (*newmap), KM_SLEEP);
505 			newmap->ra_base = base + len;
506 			newmap->ra_len = mapp->ra_len - (len + newlen);
507 			mapp->ra_len = newlen;
508 			RA_INSERT(&(mapp->ra_next), newmap);
509 		}
510 	} else {
511 		/* at the beginning */
512 		mapp->ra_base += len;
513 		mapp->ra_len -= len;
514 		if (mapp->ra_len == 0) {
515 			/* remove the whole node */
516 			RA_REMOVE(backp, mapp);
517 			kmem_free((caddr_t)mapp, sizeof (*mapp));
518 		}
519 	}
520 }
521 
522 int
523 ndi_ra_alloc(dev_info_t *dip, ndi_ra_request_t *req, uint64_t *retbasep,
524     uint64_t *retlenp, char *type, uint32_t flag)
525 {
526 	struct ra_dip_type *dipmap;
527 	struct ra_resource *mapp, **backp, **backlargestp;
528 	uint64_t mask = 0;
529 	uint64_t len, remlen, largestbase, largestlen;
530 	uint64_t base, oldbase, lower, upper;
531 	struct ra_dip_type  **backdip;
532 	struct ra_type_map  **backtype;
533 	int  rval = NDI_FAILURE;
534 
535 
536 	len = req->ra_len;
537 
538 	if (req->ra_flags & NDI_RA_ALIGN_SIZE) {
539 		if (isnot_pow2(req->ra_len)) {
540 			DEBUGPRT(CE_WARN, "ndi_ra_alloc: bad length(pow2) 0x%"
541 			    PRIx64, req->ra_len);
542 			*retbasep = 0;
543 			*retlenp = 0;
544 			return (NDI_FAILURE);
545 		}
546 	}
547 
548 	mask = (req->ra_flags & NDI_RA_ALIGN_SIZE) ? (len - 1) :
549 	    req->ra_align_mask;
550 
551 
552 	mutex_enter(&ra_lock);
553 	dipmap = find_dip_map_resources(dip, type, &backdip, &backtype, flag);
554 	if ((dipmap == NULL) || ((mapp = dipmap->ra_rangeset) == NULL)) {
555 		mutex_exit(&ra_lock);
556 		DEBUGPRT(CE_CONT, "ndi_ra_alloc no map found for this type\n");
557 		return (NDI_FAILURE);
558 	}
559 
560 	DEBUGPRT(CE_CONT, "ndi_ra_alloc: mapp = %p len=%" PRIx64 ", mask=%"
561 	    PRIx64 "\n", (void *)mapp, len, mask);
562 
563 	backp = &(dipmap->ra_rangeset);
564 	backlargestp = NULL;
565 	largestbase = 0;
566 	largestlen = 0;
567 
568 	lower = 0;
569 	upper = ~(uint64_t)0;
570 
571 	if (req->ra_flags & NDI_RA_ALLOC_BOUNDED) {
572 		/* bounded so skip to first possible */
573 		lower = req->ra_boundbase;
574 		upper = req->ra_boundlen + lower;
575 		if ((upper == 0) || (upper < req->ra_boundlen))
576 			upper = ~(uint64_t)0;
577 		DEBUGPRT(CE_CONT, "ndi_ra_alloc: ra_len = %" PRIx64 ", len = %"
578 		    PRIx64 " ra_base=%" PRIx64 ", mask=%" PRIx64
579 		    "\n", mapp->ra_len, len, mapp->ra_base, mask);
580 		for (; mapp != NULL && (mapp->ra_base + mapp->ra_len) < lower;
581 		    backp = &(mapp->ra_next), mapp = mapp->ra_next) {
582 			if (((mapp->ra_len + mapp->ra_base) == 0) ||
583 			    ((mapp->ra_len + mapp->ra_base) < mapp->ra_len))
584 				/*
585 				 * This elements end goes beyond max uint64_t.
586 				 * potential candidate, check end against lower
587 				 * would not be precise.
588 				 */
589 				break;
590 
591 			DEBUGPRT(CE_CONT, " ra_len = %" PRIx64 ", ra_base=%"
592 			    PRIx64 "\n", mapp->ra_len, mapp->ra_base);
593 			}
594 
595 	}
596 
597 	if (!(req->ra_flags & NDI_RA_ALLOC_SPECIFIED)) {
598 		/* first fit - not user specified */
599 		DEBUGPRT(CE_CONT, "ndi_ra_alloc(unspecified request)"
600 		    "lower=%" PRIx64 ", upper=%" PRIx64 "\n", lower, upper);
601 		for (; mapp != NULL && mapp->ra_base <= upper;
602 		    backp = &(mapp->ra_next), mapp = mapp->ra_next) {
603 
604 			DEBUGPRT(CE_CONT, "ndi_ra_alloc: ra_len = %" PRIx64
605 			    ", len = %" PRIx64 "", mapp->ra_len, len);
606 			base = mapp->ra_base;
607 			if (base < lower) {
608 				base = lower;
609 				DEBUGPRT(CE_CONT, "\tbase=%" PRIx64
610 				    ", ra_base=%" PRIx64 ", mask=%" PRIx64,
611 				    base, mapp->ra_base, mask);
612 			}
613 
614 			if ((base & mask) != 0) {
615 				oldbase = base;
616 				/*
617 				 * failed a critical constraint
618 				 * adjust and see if it still fits
619 				 */
620 				base = base & ~mask;
621 				base += (mask + 1);
622 				DEBUGPRT(CE_CONT, "\tnew base=%" PRIx64 "\n",
623 				    base);
624 
625 				/*
626 				 * Check to see if the new base is past
627 				 * the end of the resource.
628 				 */
629 				if (base >= (oldbase + mapp->ra_len + 1)) {
630 					continue;
631 				}
632 			}
633 
634 			if (req->ra_flags & NDI_RA_ALLOC_PARTIAL_OK) {
635 				if ((upper - mapp->ra_base)  <  mapp->ra_len)
636 					remlen = upper - base;
637 				else
638 					remlen = mapp->ra_len -
639 					    (base - mapp->ra_base);
640 
641 				if ((backlargestp == NULL) ||
642 				    (largestlen < remlen)) {
643 
644 					backlargestp = backp;
645 					largestbase = base;
646 					largestlen = remlen;
647 				}
648 			}
649 
650 			if (mapp->ra_len >= len) {
651 				/* a candidate -- apply constraints */
652 				if ((len > (mapp->ra_len -
653 				    (base - mapp->ra_base))) ||
654 				    ((len - 1 + base) > upper)) {
655 					continue;
656 				}
657 
658 				/* we have a fit */
659 
660 				DEBUGPRT(CE_CONT, "\thave a fit\n");
661 
662 				adjust_link(backp, mapp, base, len);
663 				rval = NDI_SUCCESS;
664 				break;
665 
666 			}
667 		}
668 	} else {
669 		/* want an exact value/fit */
670 		base = req->ra_addr;
671 		len = req->ra_len;
672 		for (; mapp != NULL && mapp->ra_base <= upper;
673 		    backp = &(mapp->ra_next), mapp = mapp->ra_next) {
674 			if (base >= mapp->ra_base &&
675 			    ((base - mapp->ra_base) < mapp->ra_len)) {
676 				/*
677 				 * This is the node with he requested base in
678 				 * its range
679 				 */
680 				if ((len > mapp->ra_len) ||
681 				    (base - mapp->ra_base >
682 				    mapp->ra_len - len)) {
683 					/* length requirement not satisfied */
684 					if (req->ra_flags &
685 					    NDI_RA_ALLOC_PARTIAL_OK) {
686 						if ((upper - mapp->ra_base)
687 						    < mapp->ra_len)
688 							remlen = upper - base;
689 						else
690 							remlen =
691 							    mapp->ra_len -
692 							    (base -
693 							    mapp->ra_base);
694 					}
695 					backlargestp = backp;
696 					largestbase = base;
697 					largestlen = remlen;
698 					base = 0;
699 				} else {
700 					/* We have a match */
701 					adjust_link(backp, mapp, base, len);
702 					rval = NDI_SUCCESS;
703 				}
704 				break;
705 			}
706 		}
707 	}
708 
709 	if ((rval != NDI_SUCCESS) &&
710 	    (req->ra_flags & NDI_RA_ALLOC_PARTIAL_OK) &&
711 	    (backlargestp != NULL)) {
712 		adjust_link(backlargestp, *backlargestp, largestbase,
713 		    largestlen);
714 
715 		base = largestbase;
716 		len = largestlen;
717 		rval = NDI_RA_PARTIAL_REQ;
718 	}
719 
720 	mutex_exit(&ra_lock);
721 
722 	if (rval == NDI_FAILURE) {
723 		*retbasep = 0;
724 		*retlenp = 0;
725 	} else {
726 		*retbasep = base;
727 		*retlenp = len;
728 	}
729 
730 	/*
731 	 * Update dip's "available" property, substract this piece of
732 	 * resource from the pool.
733 	 */
734 	if ((rval == NDI_SUCCESS) || (rval == NDI_RA_PARTIAL_REQ))
735 		(void) pci_get_available_prop(dipmap->ra_dip,
736 		    *retbasep, *retlenp, type);
737 
738 	return (rval);
739 }
740 
741 /*
742  * isa_resource_setup
743  *	check for /used-resources and initialize
744  *	based on info there.  If no /used-resources,
745  *	fail.
746  */
747 int
748 isa_resource_setup()
749 {
750 	dev_info_t *used, *usedpdip;
751 	/*
752 	 * note that at this time bootconf creates 32 bit properties for
753 	 * io-space and device-memory
754 	 */
755 	struct iorange {
756 		uint32_t	base;
757 		uint32_t	len;
758 	} *iorange;
759 	struct memrange {
760 		uint32_t	base;
761 		uint32_t	len;
762 	} *memrange;
763 	uint32_t *irq;
764 	int proplen;
765 	int i, len;
766 	int maxrange;
767 	ndi_ra_request_t req;
768 	uint64_t retbase;
769 	uint64_t retlen;
770 
771 	used = ddi_find_devinfo("used-resources", -1, 0);
772 	if (used == NULL) {
773 		DEBUGPRT(CE_CONT,
774 		    "isa_resource_setup: used-resources not found");
775 		return (NDI_FAILURE);
776 	}
777 
778 	/*
779 	 * initialize to all resources being present
780 	 * and then remove the ones in use.
781 	 */
782 
783 	usedpdip = ddi_root_node();
784 
785 	DEBUGPRT(CE_CONT, "isa_resource_setup: used = %p usedpdip = %p\n",
786 	    (void *)used, (void *)usedpdip);
787 
788 	if (ndi_ra_map_setup(usedpdip, NDI_RA_TYPE_IO) == NDI_FAILURE) {
789 		return (NDI_FAILURE);
790 	}
791 
792 	/* initialize io space, highest end base is 0xffff */
793 	/* note that length is highest addr + 1 since starts from 0 */
794 
795 	(void) ndi_ra_free(usedpdip, 0, 0xffff + 1,  NDI_RA_TYPE_IO, 0);
796 
797 	if (ddi_getlongprop(DDI_DEV_T_ANY, used, DDI_PROP_DONTPASS,
798 	    "io-space", (caddr_t)&iorange, &proplen) == DDI_SUCCESS) {
799 		maxrange = proplen / sizeof (struct iorange);
800 		/* remove the "used" I/O resources */
801 		for (i = 0; i < maxrange; i++) {
802 			bzero((caddr_t)&req, sizeof (req));
803 			req.ra_addr =  (uint64_t)iorange[i].base;
804 			req.ra_len = (uint64_t)iorange[i].len;
805 			req.ra_flags = NDI_RA_ALLOC_SPECIFIED;
806 			(void) ndi_ra_alloc(usedpdip, &req, &retbase, &retlen,
807 			    NDI_RA_TYPE_IO, 0);
808 		}
809 
810 		kmem_free((caddr_t)iorange, proplen);
811 	}
812 
813 	if (ndi_ra_map_setup(usedpdip, NDI_RA_TYPE_MEM) == NDI_FAILURE) {
814 		return (NDI_FAILURE);
815 	}
816 	/* initialize memory space where highest end base is 0xffffffff */
817 	/* note that length is highest addr + 1 since starts from 0 */
818 	(void) ndi_ra_free(usedpdip, 0, ((uint64_t)((uint32_t)~0)) + 1,
819 	    NDI_RA_TYPE_MEM, 0);
820 
821 	if (ddi_getlongprop(DDI_DEV_T_ANY, used, DDI_PROP_DONTPASS,
822 	    "device-memory", (caddr_t)&memrange, &proplen) == DDI_SUCCESS) {
823 		maxrange = proplen / sizeof (struct memrange);
824 		/* remove the "used" memory resources */
825 		for (i = 0; i < maxrange; i++) {
826 			bzero((caddr_t)&req, sizeof (req));
827 			req.ra_addr = (uint64_t)memrange[i].base;
828 			req.ra_len = (uint64_t)memrange[i].len;
829 			req.ra_flags = NDI_RA_ALLOC_SPECIFIED;
830 			(void) ndi_ra_alloc(usedpdip, &req, &retbase, &retlen,
831 			    NDI_RA_TYPE_MEM, 0);
832 		}
833 
834 		kmem_free((caddr_t)memrange, proplen);
835 	}
836 
837 	if (ndi_ra_map_setup(usedpdip, NDI_RA_TYPE_INTR) == NDI_FAILURE) {
838 		return (NDI_FAILURE);
839 	}
840 
841 	/* initialize the interrupt space */
842 	(void) ndi_ra_free(usedpdip, 0, 16, NDI_RA_TYPE_INTR, 0);
843 
844 #if defined(__i386) || defined(__amd64)
845 	bzero(&req, sizeof (req));
846 	req.ra_addr = 2;	/* 2 == 9 so never allow */
847 	req.ra_len = 1;
848 	req.ra_flags = NDI_RA_ALLOC_SPECIFIED;
849 	(void) ndi_ra_alloc(usedpdip, &req, &retbase, &retlen,
850 	    NDI_RA_TYPE_INTR, 0);
851 #endif
852 
853 	if (ddi_getlongprop(DDI_DEV_T_ANY, used, DDI_PROP_DONTPASS,
854 	    "interrupts", (caddr_t)&irq, &proplen) == DDI_SUCCESS) {
855 		/* Initialize available interrupts by negating the used */
856 		len = (proplen / sizeof (uint32_t));
857 		for (i = 0; i < len; i++) {
858 			bzero((caddr_t)&req, sizeof (req));
859 			req.ra_addr = (uint64_t)irq[i];
860 			req.ra_len = 1;
861 			req.ra_flags = NDI_RA_ALLOC_SPECIFIED;
862 			(void) ndi_ra_alloc(usedpdip, &req, &retbase, &retlen,
863 			    NDI_RA_TYPE_INTR, 0);
864 		}
865 		kmem_free((caddr_t)irq, proplen);
866 	}
867 
868 #ifdef BUSRA_DEBUG
869 	if (busra_debug) {
870 		(void) ra_dump_all(NULL, usedpdip);
871 	}
872 #endif
873 	return (NDI_SUCCESS);
874 
875 }
876 
877 #ifdef BUSRA_DEBUG
878 void
879 ra_dump_all(char *type, dev_info_t *dip)
880 {
881 
882 	struct ra_type_map *typemap;
883 	struct ra_dip_type *dipmap;
884 	struct ra_resource *res;
885 
886 	typemap =  (struct ra_type_map *)ra_map_list_head;
887 
888 	for (; typemap != NULL; typemap = typemap->ra_next) {
889 		if (type != NULL) {
890 			if (strcmp(typemap->type, type) != 0)
891 				continue;
892 		}
893 		cmn_err(CE_CONT, "type is %s\n", typemap->type);
894 		for (dipmap = typemap->ra_dip_list; dipmap != NULL;
895 		    dipmap = dipmap->ra_next) {
896 			if (dip != NULL) {
897 				if ((dipmap->ra_dip) != dip)
898 					continue;
899 			}
900 			cmn_err(CE_CONT, "  dip is %p\n",
901 			    (void *)dipmap->ra_dip);
902 			for (res = dipmap->ra_rangeset; res != NULL;
903 			    res = res->ra_next) {
904 				cmn_err(CE_CONT, "\t  range is %" PRIx64
905 				    " %" PRIx64 "\n", res->ra_base,
906 				    res->ra_len);
907 			}
908 			if (dip != NULL)
909 				break;
910 		}
911 		if (type != NULL)
912 			break;
913 	}
914 }
915 #endif
916 
917 struct bus_range {	/* 1275 "bus-range" property definition */
918 	uint32_t lo;
919 	uint32_t hi;
920 } pci_bus_range;
921 
922 struct busnum_ctrl {
923 	int	rv;
924 	dev_info_t *dip;
925 	struct	bus_range *range;
926 };
927 
928 
929 /*
930  * Setup resource map for the pci bus node based on the "available"
931  * property and "bus-range" property.
932  */
933 int
934 pci_resource_setup(dev_info_t *dip)
935 {
936 	pci_regspec_t *regs;
937 	int rlen, rcount, i;
938 	char bus_type[16] = "(unknown)";
939 	int len;
940 	struct busnum_ctrl ctrl;
941 	int circular_count;
942 	int rval = NDI_SUCCESS;
943 
944 	/*
945 	 * If this is a pci bus node then look for "available" property
946 	 * to find the available resources on this bus.
947 	 */
948 	len = sizeof (bus_type);
949 	if (ddi_prop_op(DDI_DEV_T_ANY, dip, PROP_LEN_AND_VAL_BUF,
950 	    DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "device_type",
951 	    (caddr_t)&bus_type, &len) != DDI_SUCCESS)
952 		return (NDI_FAILURE);
953 
954 	/* it is not a pci/pci-ex bus type */
955 	if ((strcmp(bus_type, "pci") != 0) && (strcmp(bus_type, "pciex") != 0))
956 		return (NDI_FAILURE);
957 
958 	/*
959 	 * The pci-hotplug project addresses adding the call
960 	 * to pci_resource_setup from pci nexus driver.
961 	 * However that project would initially be only for x86,
962 	 * so for sparc pcmcia-pci support we still need to call
963 	 * pci_resource_setup in pcic driver. Once all pci nexus drivers
964 	 * are updated to call pci_resource_setup this portion of the
965 	 * code would really become an assert to make sure this
966 	 * function is not called for the same dip twice.
967 	 */
968 	/*
969 	 * Another user for the check below is hotplug PCI/PCIe bridges.
970 	 *
971 	 * For PCI/PCIE devices under a PCIE hierarchy, ndi_ra_alloc/free
972 	 * will update the devinfo node's "available" property, to reflect
973 	 * the fact that a piece of resource has been removed/added to
974 	 * a devinfo node.
975 	 * During probe of a new PCI bridge in the hotplug case, PCI
976 	 * configurator firstly allocates maximum MEM/IO from its parent,
977 	 * then calls ndi_ra_free() to use these resources to setup busra
978 	 * pool for the new bridge, as well as adding these resources to
979 	 * the "available" property of the new devinfo node. Then configu-
980 	 * rator will attach driver for the bridge before probing its
981 	 * children, and the bridge driver will then initialize its hotplug
982 	 * contollers (if it supports hotplug) and HPC driver will call
983 	 * this function to setup the busra pool, but the resource pool
984 	 * has already been setup at the first of pcicfg_probe_bridge(),
985 	 * thus we need the check below to return directly in this case.
986 	 * Otherwise the ndi_ra_free() below will see overlapping resources.
987 	 */
988 	{
989 		if (ra_map_exist(dip, NDI_RA_TYPE_MEM) == NDI_SUCCESS) {
990 			return (NDI_FAILURE);
991 		}
992 	}
993 
994 
995 	/*
996 	 * Create empty resource maps first.
997 	 *
998 	 * NOTE: If all the allocated resources are already assigned to
999 	 * device(s) in the hot plug slot then "available" property may not
1000 	 * be present. But, subsequent hot plug operation may unconfigure
1001 	 * the device in the slot and try to free up it's resources. So,
1002 	 * at the minimum we should create empty maps here.
1003 	 */
1004 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_MEM) == NDI_FAILURE) {
1005 		return (NDI_FAILURE);
1006 	}
1007 
1008 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_IO) == NDI_FAILURE) {
1009 		return (NDI_FAILURE);
1010 	}
1011 
1012 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_PCI_BUSNUM) == NDI_FAILURE) {
1013 		return (NDI_FAILURE);
1014 	}
1015 
1016 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_PCI_PREFETCH_MEM) ==
1017 	    NDI_FAILURE) {
1018 		return (NDI_FAILURE);
1019 	}
1020 
1021 	/* read the "available" property if it is available */
1022 	if (ddi_getlongprop(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS,
1023 	    "available", (caddr_t)&regs, &rlen) == DDI_SUCCESS) {
1024 		/*
1025 		 * Remove "available" property as the entries will be
1026 		 * re-created in ndi_ra_free() below, note prom based
1027 		 * property will not be removed. But in ndi_ra_free()
1028 		 * we'll be creating non prom based property entries.
1029 		 */
1030 		(void) ndi_prop_remove(DDI_DEV_T_NONE, dip, "available");
1031 		/*
1032 		 * create the available resource list for both memory and
1033 		 * io space
1034 		 */
1035 		rcount = rlen / sizeof (pci_regspec_t);
1036 		for (i = 0; i < rcount; i++) {
1037 			switch (PCI_REG_ADDR_G(regs[i].pci_phys_hi)) {
1038 			case PCI_REG_ADDR_G(PCI_ADDR_MEM32):
1039 				(void) ndi_ra_free(dip,
1040 				    (uint64_t)regs[i].pci_phys_low,
1041 				    (uint64_t)regs[i].pci_size_low,
1042 				    (regs[i].pci_phys_hi & PCI_REG_PF_M) ?
1043 				    NDI_RA_TYPE_PCI_PREFETCH_MEM :
1044 				    NDI_RA_TYPE_MEM,
1045 				    0);
1046 				break;
1047 			case PCI_REG_ADDR_G(PCI_ADDR_MEM64):
1048 				(void) ndi_ra_free(dip,
1049 				    ((uint64_t)(regs[i].pci_phys_mid) << 32) |
1050 				    ((uint64_t)(regs[i].pci_phys_low)),
1051 				    ((uint64_t)(regs[i].pci_size_hi) << 32) |
1052 				    ((uint64_t)(regs[i].pci_size_low)),
1053 				    (regs[i].pci_phys_hi & PCI_REG_PF_M) ?
1054 				    NDI_RA_TYPE_PCI_PREFETCH_MEM :
1055 				    NDI_RA_TYPE_MEM,
1056 				    0);
1057 				break;
1058 			case PCI_REG_ADDR_G(PCI_ADDR_IO):
1059 				(void) ndi_ra_free(dip,
1060 				    (uint64_t)regs[i].pci_phys_low,
1061 				    (uint64_t)regs[i].pci_size_low,
1062 				    NDI_RA_TYPE_IO,
1063 				    0);
1064 				break;
1065 			case PCI_REG_ADDR_G(PCI_ADDR_CONFIG):
1066 				break;
1067 			default:
1068 				cmn_err(CE_WARN,
1069 				    "pci_resource_setup: bad addr type: %x\n",
1070 				    PCI_REG_ADDR_G(regs[i].pci_phys_hi));
1071 				break;
1072 			}
1073 		}
1074 		kmem_free(regs, rlen);
1075 	}
1076 
1077 	/*
1078 	 * update resource map for available bus numbers if the node
1079 	 * has available-bus-range or bus-range property.
1080 	 */
1081 	len = sizeof (struct bus_range);
1082 	if (ddi_getlongprop_buf(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS,
1083 	    "available-bus-range", (caddr_t)&pci_bus_range, &len) ==
1084 	    DDI_SUCCESS) {
1085 		/*
1086 		 * Add bus numbers in the range to the free list.
1087 		 */
1088 		(void) ndi_ra_free(dip, (uint64_t)pci_bus_range.lo,
1089 		    (uint64_t)pci_bus_range.hi - (uint64_t)pci_bus_range.lo +
1090 		    1, NDI_RA_TYPE_PCI_BUSNUM, 0);
1091 	} else {
1092 		/*
1093 		 * We don't have an available-bus-range property. If, instead,
1094 		 * we have a bus-range property we add all the bus numbers
1095 		 * in that range to the free list but we must then scan
1096 		 * for pci-pci bridges on this bus to find out the if there
1097 		 * are any of those bus numbers already in use. If so, we can
1098 		 * reclaim them.
1099 		 */
1100 		len = sizeof (struct bus_range);
1101 		if (ddi_getlongprop_buf(DDI_DEV_T_ANY, dip,
1102 		    DDI_PROP_DONTPASS, "bus-range", (caddr_t)&pci_bus_range,
1103 		    &len) == DDI_SUCCESS) {
1104 			if (pci_bus_range.lo != pci_bus_range.hi) {
1105 				/*
1106 				 * Add bus numbers other than the secondary
1107 				 * bus number to the free list.
1108 				 */
1109 				(void) ndi_ra_free(dip,
1110 				    (uint64_t)pci_bus_range.lo + 1,
1111 				    (uint64_t)pci_bus_range.hi -
1112 				    (uint64_t)pci_bus_range.lo,
1113 				    NDI_RA_TYPE_PCI_BUSNUM, 0);
1114 
1115 				/* scan for pci-pci bridges */
1116 				ctrl.rv = DDI_SUCCESS;
1117 				ctrl.dip = dip;
1118 				ctrl.range = &pci_bus_range;
1119 				ndi_devi_enter(dip, &circular_count);
1120 				ddi_walk_devs(ddi_get_child(dip),
1121 				    claim_pci_busnum, (void *)&ctrl);
1122 				ndi_devi_exit(dip, circular_count);
1123 				if (ctrl.rv != DDI_SUCCESS) {
1124 					/* failed to create the map */
1125 					(void) ndi_ra_map_destroy(dip,
1126 					    NDI_RA_TYPE_PCI_BUSNUM);
1127 					rval = NDI_FAILURE;
1128 				}
1129 			}
1130 		}
1131 	}
1132 
1133 #ifdef BUSRA_DEBUG
1134 	if (busra_debug) {
1135 		(void) ra_dump_all(NULL, dip);
1136 	}
1137 #endif
1138 
1139 	return (rval);
1140 }
1141 
1142 /*
1143  * If the device is a PCI bus device (i.e bus-range property exists) then
1144  * claim the bus numbers used by the device from the specified bus
1145  * resource map.
1146  */
1147 static int
1148 claim_pci_busnum(dev_info_t *dip, void *arg)
1149 {
1150 	struct bus_range pci_bus_range;
1151 	struct busnum_ctrl *ctrl;
1152 	ndi_ra_request_t req;
1153 	char bus_type[16] = "(unknown)";
1154 	int len;
1155 	uint64_t base;
1156 	uint64_t retlen;
1157 
1158 	ctrl = (struct busnum_ctrl *)arg;
1159 
1160 	/* check if this is a PCI bus node */
1161 	len = sizeof (bus_type);
1162 	if (ddi_prop_op(DDI_DEV_T_ANY, dip, PROP_LEN_AND_VAL_BUF,
1163 	    DDI_PROP_CANSLEEP | DDI_PROP_DONTPASS, "device_type",
1164 	    (caddr_t)&bus_type, &len) != DDI_SUCCESS)
1165 		return (DDI_WALK_PRUNECHILD);
1166 
1167 	/* it is not a pci/pci-ex bus type */
1168 	if ((strcmp(bus_type, "pci") != 0) && (strcmp(bus_type, "pciex") != 0))
1169 		return (DDI_WALK_PRUNECHILD);
1170 
1171 	/* look for the bus-range property */
1172 	len = sizeof (struct bus_range);
1173 	if (ddi_getlongprop_buf(DDI_DEV_T_ANY, dip, DDI_PROP_DONTPASS,
1174 	    "bus-range", (caddr_t)&pci_bus_range, &len) == DDI_SUCCESS) {
1175 		if ((pci_bus_range.lo >= ctrl->range->lo) &&
1176 		    (pci_bus_range.hi <= ctrl->range->hi)) {
1177 
1178 			/* claim the bus range from the bus resource map */
1179 			bzero((caddr_t)&req, sizeof (req));
1180 			req.ra_addr = (uint64_t)pci_bus_range.lo;
1181 			req.ra_flags |= NDI_RA_ALLOC_SPECIFIED;
1182 			req.ra_len = (uint64_t)pci_bus_range.hi -
1183 			    (uint64_t)pci_bus_range.lo + 1;
1184 			if (ndi_ra_alloc(ctrl->dip, &req, &base, &retlen,
1185 			    NDI_RA_TYPE_PCI_BUSNUM, 0) == NDI_SUCCESS)
1186 				return (DDI_WALK_PRUNECHILD);
1187 		}
1188 	}
1189 
1190 	/*
1191 	 * Error return.
1192 	 */
1193 	ctrl->rv = DDI_FAILURE;
1194 	return (DDI_WALK_TERMINATE);
1195 }
1196 
1197 void
1198 pci_resource_destroy(dev_info_t *dip)
1199 {
1200 	(void) ndi_ra_map_destroy(dip, NDI_RA_TYPE_IO);
1201 
1202 	(void) ndi_ra_map_destroy(dip, NDI_RA_TYPE_MEM);
1203 
1204 	(void) ndi_ra_map_destroy(dip, NDI_RA_TYPE_PCI_BUSNUM);
1205 
1206 	(void) ndi_ra_map_destroy(dip, NDI_RA_TYPE_PCI_PREFETCH_MEM);
1207 }
1208 
1209 
1210 int
1211 pci_resource_setup_avail(dev_info_t *dip, pci_regspec_t *avail_p, int entries)
1212 {
1213 	int i;
1214 
1215 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_MEM) == NDI_FAILURE)
1216 		return (NDI_FAILURE);
1217 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_IO) == NDI_FAILURE)
1218 		return (NDI_FAILURE);
1219 	if (ndi_ra_map_setup(dip, NDI_RA_TYPE_PCI_PREFETCH_MEM) == NDI_FAILURE)
1220 		return (NDI_FAILURE);
1221 
1222 	/* for each entry in the PCI "available" property */
1223 	for (i = 0; i < entries; i++, avail_p++) {
1224 		if (avail_p->pci_phys_hi == -1u)
1225 			goto err;
1226 
1227 		switch (PCI_REG_ADDR_G(avail_p->pci_phys_hi)) {
1228 		case PCI_REG_ADDR_G(PCI_ADDR_MEM32): {
1229 			(void) ndi_ra_free(dip, (uint64_t)avail_p->pci_phys_low,
1230 			    (uint64_t)avail_p->pci_size_low,
1231 			    (avail_p->pci_phys_hi & PCI_REG_PF_M) ?
1232 			    NDI_RA_TYPE_PCI_PREFETCH_MEM : NDI_RA_TYPE_MEM,
1233 			    0);
1234 			}
1235 			break;
1236 		case PCI_REG_ADDR_G(PCI_ADDR_IO):
1237 			(void) ndi_ra_free(dip, (uint64_t)avail_p->pci_phys_low,
1238 			    (uint64_t)avail_p->pci_size_low, NDI_RA_TYPE_IO, 0);
1239 			break;
1240 		default:
1241 			goto err;
1242 		}
1243 	}
1244 #ifdef BUSRA_DEBUG
1245 	if (busra_debug) {
1246 		(void) ra_dump_all(NULL, dip);
1247 	}
1248 #endif
1249 	return (NDI_SUCCESS);
1250 
1251 err:
1252 	cmn_err(CE_WARN, "pci_resource_setup_avail: bad entry[%d]=%x\n",
1253 	    i, avail_p->pci_phys_hi);
1254 	return (NDI_FAILURE);
1255 }
1256 
1257 /*
1258  * Return true if the devinfo node resides on PCI or PCI Express bus,
1259  * sitting in a PCI Express hierarchy.
1260  */
1261 static boolean_t
1262 is_pcie_fabric(dev_info_t *dip)
1263 {
1264 	dev_info_t *root = ddi_root_node();
1265 	dev_info_t *pdip;
1266 	boolean_t found = B_FALSE;
1267 	char *bus;
1268 
1269 	/*
1270 	 * Is this pci/pcie ?
1271 	 */
1272 	if (ddi_prop_lookup_string(DDI_DEV_T_ANY, dip,
1273 	    DDI_PROP_DONTPASS, "device_type", &bus) !=
1274 	    DDI_PROP_SUCCESS) {
1275 		DEBUGPRT(CE_WARN, "is_pcie_fabric: cannot find "
1276 		    "\"device_type\" property for dip %p\n", (void *)dip);
1277 		return (B_FALSE);
1278 	}
1279 
1280 	if (strcmp(bus, "pciex") == 0) {
1281 		/* pcie bus, done */
1282 		ddi_prop_free(bus);
1283 		return (B_TRUE);
1284 	} else if (strcmp(bus, "pci") == 0) {
1285 		/*
1286 		 * pci bus, fall through to check if it resides in
1287 		 * a pcie hierarchy.
1288 		 */
1289 		ddi_prop_free(bus);
1290 	} else {
1291 		/* other bus, return failure */
1292 		ddi_prop_free(bus);
1293 		return (B_FALSE);
1294 	}
1295 
1296 	/*
1297 	 * Does this device reside in a pcie fabric ?
1298 	 */
1299 	for (pdip = ddi_get_parent(dip); pdip && (pdip != root) &&
1300 	    !found; pdip = ddi_get_parent(pdip)) {
1301 		if (ddi_prop_lookup_string(DDI_DEV_T_ANY, pdip,
1302 		    DDI_PROP_DONTPASS, "device_type", &bus) !=
1303 		    DDI_PROP_SUCCESS)
1304 			break;
1305 
1306 		if (strcmp(bus, "pciex") == 0)
1307 			found = B_TRUE;
1308 
1309 		ddi_prop_free(bus);
1310 	}
1311 
1312 	return (found);
1313 }
1314 
1315 /*
1316  * Remove a piece of IO/MEM resource from "available" property of 'dip'.
1317  */
1318 static int
1319 pci_get_available_prop(dev_info_t *dip, uint64_t base, uint64_t len,
1320     char *busra_type)
1321 {
1322 	pci_regspec_t	*regs, *newregs;
1323 	uint_t		status;
1324 	int		rlen, rcount;
1325 	int		i, j, k;
1326 	uint64_t	dlen;
1327 	boolean_t	found = B_FALSE;
1328 	uint32_t	type;
1329 
1330 	/* check if we're manipulating MEM/IO resource */
1331 	if ((type = pci_type_ra2pci(busra_type)) == PCI_ADDR_TYPE_INVAL)
1332 		return (DDI_SUCCESS);
1333 
1334 	/* check if dip is a pci/pcie device resides in a pcie fabric */
1335 	if (!is_pcie_fabric(dip))
1336 		return (DDI_SUCCESS);
1337 
1338 	status = ddi_getlongprop(DDI_DEV_T_ANY, dip,
1339 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
1340 	    "available", (caddr_t)&regs, &rlen);
1341 
1342 	ASSERT(status == DDI_SUCCESS);
1343 	if (status != DDI_SUCCESS)
1344 		return (status);
1345 
1346 	/*
1347 	 * The updated "available" property will at most have one more entry
1348 	 * than existing one (when the requested range is in the middle of
1349 	 * the matched property entry)
1350 	 */
1351 	newregs = kmem_alloc(rlen + sizeof (pci_regspec_t), KM_SLEEP);
1352 
1353 	rcount = rlen / sizeof (pci_regspec_t);
1354 	for (i = 0, j = 0; i < rcount; i++) {
1355 		if (type == (regs[i].pci_phys_hi & PCI_ADDR_TYPE_MASK)) {
1356 			uint64_t range_base, range_len;
1357 
1358 			range_base = ((uint64_t)(regs[i].pci_phys_mid) << 32) |
1359 			    ((uint64_t)(regs[i].pci_phys_low));
1360 			range_len = ((uint64_t)(regs[i].pci_size_hi) << 32) |
1361 			    ((uint64_t)(regs[i].pci_size_low));
1362 
1363 			if ((base < range_base) ||
1364 			    (base + len > range_base + range_len)) {
1365 				/*
1366 				 * not a match, copy the entry
1367 				 */
1368 				goto copy_entry;
1369 			}
1370 
1371 			/*
1372 			 * range_base	base	base+len	range_base
1373 			 *					+range_len
1374 			 *   +------------+-----------+----------+
1375 			 *   |		  |///////////|		 |
1376 			 *   +------------+-----------+----------+
1377 			 */
1378 			/*
1379 			 * Found a match, remove the range out of this entry.
1380 			 */
1381 			found = B_TRUE;
1382 
1383 			dlen = base - range_base;
1384 			if (dlen != 0) {
1385 				newregs[j].pci_phys_hi = regs[i].pci_phys_hi;
1386 				newregs[j].pci_phys_mid =
1387 				    (uint32_t)(range_base >> 32);
1388 				newregs[j].pci_phys_low =
1389 				    (uint32_t)(range_base);
1390 				newregs[j].pci_size_hi = (uint32_t)(dlen >> 32);
1391 				newregs[j].pci_size_low = (uint32_t)dlen;
1392 				j++;
1393 			}
1394 
1395 			dlen = (range_base + range_len) - (base + len);
1396 			if (dlen != 0) {
1397 				newregs[j].pci_phys_hi = regs[i].pci_phys_hi;
1398 				newregs[j].pci_phys_mid =
1399 				    (uint32_t)((base + len)>> 32);
1400 				newregs[j].pci_phys_low =
1401 				    (uint32_t)(base + len);
1402 				newregs[j].pci_size_hi = (uint32_t)(dlen >> 32);
1403 				newregs[j].pci_size_low = (uint32_t)dlen;
1404 				j++;
1405 			}
1406 
1407 			/*
1408 			 * We've allocated the resource from the matched
1409 			 * entry, almost finished but still need to copy
1410 			 * the rest entries from the original property
1411 			 * array.
1412 			 */
1413 			for (k = i + 1; k < rcount; k++) {
1414 				newregs[j] = regs[k];
1415 				j++;
1416 			}
1417 
1418 			goto done;
1419 
1420 		} else {
1421 copy_entry:
1422 			newregs[j] = regs[i];
1423 			j++;
1424 		}
1425 	}
1426 
1427 done:
1428 	/*
1429 	 * This should not fail so assert it. For non-debug kernel we don't
1430 	 * want to panic thus only logging a warning message.
1431 	 */
1432 	ASSERT(found == B_TRUE);
1433 	if (!found) {
1434 		cmn_err(CE_WARN, "pci_get_available_prop: failed to remove "
1435 		    "resource from dip %p : base 0x%" PRIx64 ", len 0x%" PRIX64
1436 		    ", type 0x%x\n", (void *)dip, base, len, type);
1437 		kmem_free(newregs, rlen + sizeof (pci_regspec_t));
1438 		kmem_free(regs, rlen);
1439 
1440 		return (DDI_FAILURE);
1441 	}
1442 
1443 	/*
1444 	 * Found the resources from parent, update the "available"
1445 	 * property.
1446 	 */
1447 	if (j == 0) {
1448 		/* all the resources are consumed, remove the property */
1449 		(void) ndi_prop_remove(DDI_DEV_T_NONE, dip, "available");
1450 	} else {
1451 		/*
1452 		 * There are still resource available in the parent dip,
1453 		 * update with the remaining resources.
1454 		 */
1455 		(void) ndi_prop_update_int_array(DDI_DEV_T_NONE, dip,
1456 		    "available", (int *)newregs,
1457 		    (j * sizeof (pci_regspec_t)) / sizeof (int));
1458 	}
1459 
1460 	kmem_free(newregs, rlen + sizeof (pci_regspec_t));
1461 	kmem_free(regs, rlen);
1462 
1463 	return (DDI_SUCCESS);
1464 }
1465 
1466 /*
1467  * Add a piece of IO/MEM resource to "available" property of 'dip'.
1468  */
1469 static int
1470 pci_put_available_prop(dev_info_t *dip, uint64_t base, uint64_t len,
1471     char *busra_type)
1472 {
1473 	pci_regspec_t	*regs, *newregs;
1474 	uint_t		status;
1475 	int		rlen, rcount;
1476 	int		i, j, k;
1477 	int		matched = 0;
1478 	uint64_t	orig_base = base;
1479 	uint64_t	orig_len = len;
1480 	uint32_t	type;
1481 
1482 	/* check if we're manipulating MEM/IO resource */
1483 	if ((type = pci_type_ra2pci(busra_type)) == PCI_ADDR_TYPE_INVAL)
1484 		return (DDI_SUCCESS);
1485 
1486 	/* check if dip is a pci/pcie device resides in a pcie fabric */
1487 	if (!is_pcie_fabric(dip))
1488 		return (DDI_SUCCESS);
1489 
1490 	status = ddi_getlongprop(DDI_DEV_T_ANY, dip,
1491 	    DDI_PROP_DONTPASS | DDI_PROP_NOTPROM,
1492 	    "available", (caddr_t)&regs, &rlen);
1493 
1494 	switch (status) {
1495 		case DDI_PROP_NOT_FOUND:
1496 			goto not_found;
1497 
1498 		case DDI_PROP_SUCCESS:
1499 			break;
1500 
1501 		default:
1502 			return (status);
1503 	}
1504 
1505 	/*
1506 	 * The "available" property exist on the node, try to put this
1507 	 * resource back, merge if there are adjacent resources.
1508 	 *
1509 	 * The updated "available" property will at most have one more entry
1510 	 * than existing one (when there is no adjacent entries thus the new
1511 	 * resource is appended at the end)
1512 	 */
1513 	newregs = kmem_alloc(rlen + sizeof (pci_regspec_t), KM_SLEEP);
1514 
1515 	rcount = rlen / sizeof (pci_regspec_t);
1516 	for (i = 0, j = 0; i < rcount; i++) {
1517 		if (type == (regs[i].pci_phys_hi & PCI_ADDR_TYPE_MASK)) {
1518 			uint64_t range_base, range_len;
1519 
1520 			range_base = ((uint64_t)(regs[i].pci_phys_mid) << 32) |
1521 			    ((uint64_t)(regs[i].pci_phys_low));
1522 			range_len = ((uint64_t)(regs[i].pci_size_hi) << 32) |
1523 			    ((uint64_t)(regs[i].pci_size_low));
1524 
1525 			if ((base + len < range_base) ||
1526 			    (base > range_base + range_len)) {
1527 				/*
1528 				 * Not adjacent, copy the entry and contiue
1529 				 */
1530 				goto copy_entry;
1531 			}
1532 
1533 			/*
1534 			 * Adjacent or overlap?
1535 			 *
1536 			 * Should not have overlapping resources so assert it.
1537 			 * For non-debug kernel we don't want to panic thus
1538 			 * only logging a warning message.
1539 			 */
1540 #if 0
1541 			ASSERT((base + len == range_base) ||
1542 			    (base == range_base + range_len));
1543 #endif
1544 			if ((base + len != range_base) &&
1545 			    (base != range_base + range_len)) {
1546 				cmn_err(CE_WARN, "pci_put_available_prop: "
1547 				    "failed to add resource to dip %p : "
1548 				    "base 0x%" PRIx64 ", len 0x%" PRIx64 " "
1549 				    "overlaps with existing resource "
1550 				    "base 0x%" PRIx64 ", len 0x%" PRIx64 "\n",
1551 				    (void *)dip, orig_base, orig_len,
1552 				    range_base, range_len);
1553 
1554 				goto failure;
1555 			}
1556 
1557 			/*
1558 			 * On the left:
1559 			 *
1560 			 * base		range_base
1561 			 *   +-------------+-------------+
1562 			 *   |/////////////|		 |
1563 			 *   +-------------+-------------+
1564 			 *	len		range_len
1565 			 *
1566 			 * On the right:
1567 			 *
1568 			 * range_base	 base
1569 			 *   +-------------+-------------+
1570 			 *   |		   |/////////////|
1571 			 *   +-------------+-------------+
1572 			 *	range_len	len
1573 			 */
1574 			/*
1575 			 * There are at most two piece of resources adjacent
1576 			 * with this resource, assert it.
1577 			 */
1578 			ASSERT(matched < 2);
1579 
1580 			if (!(matched < 2)) {
1581 				cmn_err(CE_WARN, "pci_put_available_prop: "
1582 				    "failed to add resource to dip %p : "
1583 				    "base 0x%" PRIx64 ", len 0x%" PRIx64 " "
1584 				    "found overlaps in existing resources\n",
1585 				    (void *)dip, orig_base, orig_len);
1586 
1587 				goto failure;
1588 			}
1589 
1590 			/* setup base & len to refer to the merged range */
1591 			len += range_len;
1592 			if (base == range_base + range_len)
1593 				base = range_base;
1594 
1595 			if (matched == 0) {
1596 				/*
1597 				 * One adjacent entry, add this resource in
1598 				 */
1599 				newregs[j].pci_phys_hi = regs[i].pci_phys_hi;
1600 				newregs[j].pci_phys_mid =
1601 				    (uint32_t)(base >> 32);
1602 				newregs[j].pci_phys_low = (uint32_t)(base);
1603 				newregs[j].pci_size_hi = (uint32_t)(len >> 32);
1604 				newregs[j].pci_size_low = (uint32_t)len;
1605 
1606 				matched = 1;
1607 				k = j;
1608 				j++;
1609 			} else { /* matched == 1 */
1610 				/*
1611 				 * Two adjacent entries, merge them together
1612 				 */
1613 				newregs[k].pci_phys_hi = regs[i].pci_phys_hi;
1614 				newregs[k].pci_phys_mid =
1615 				    (uint32_t)(base >> 32);
1616 				newregs[k].pci_phys_low = (uint32_t)(base);
1617 				newregs[k].pci_size_hi = (uint32_t)(len >> 32);
1618 				newregs[k].pci_size_low = (uint32_t)len;
1619 
1620 				matched = 2;
1621 			}
1622 		} else {
1623 copy_entry:
1624 			newregs[j] = regs[i];
1625 			j++;
1626 		}
1627 	}
1628 
1629 	if (matched == 0) {
1630 		/* No adjacent entries, append at end */
1631 		ASSERT(j == rcount);
1632 
1633 		/*
1634 		 * According to page 15 of 1275 spec, bit "n" of "available"
1635 		 * should be set to 1.
1636 		 */
1637 		newregs[j].pci_phys_hi = type;
1638 		newregs[j].pci_phys_hi |= PCI_REG_REL_M;
1639 
1640 		newregs[j].pci_phys_mid = (uint32_t)(base >> 32);
1641 		newregs[j].pci_phys_low = (uint32_t)base;
1642 		newregs[j].pci_size_hi = (uint32_t)(len >> 32);
1643 		newregs[j].pci_size_low = (uint32_t)len;
1644 
1645 		j++;
1646 	}
1647 
1648 	(void) ndi_prop_update_int_array(DDI_DEV_T_NONE, dip,
1649 	    "available", (int *)newregs,
1650 	    (j * sizeof (pci_regspec_t)) / sizeof (int));
1651 
1652 	kmem_free(newregs, rlen + sizeof (pci_regspec_t));
1653 	kmem_free(regs, rlen);
1654 	return (DDI_SUCCESS);
1655 
1656 not_found:
1657 	/*
1658 	 * There is no "available" property on the parent node, create it.
1659 	 */
1660 	newregs = kmem_alloc(sizeof (pci_regspec_t), KM_SLEEP);
1661 
1662 	/*
1663 	 * According to page 15 of 1275 spec, bit "n" of "available" should
1664 	 * be set to 1.
1665 	 */
1666 	newregs[0].pci_phys_hi = type;
1667 	newregs[0].pci_phys_hi |= PCI_REG_REL_M;
1668 
1669 	newregs[0].pci_phys_mid = (uint32_t)(base >> 32);
1670 	newregs[0].pci_phys_low = (uint32_t)base;
1671 	newregs[0].pci_size_hi = (uint32_t)(len >> 32);
1672 	newregs[0].pci_size_low = (uint32_t)len;
1673 
1674 	(void) ndi_prop_update_int_array(DDI_DEV_T_NONE, dip,
1675 	    "available", (int *)newregs,
1676 	    sizeof (pci_regspec_t) / sizeof (int));
1677 	kmem_free(newregs, sizeof (pci_regspec_t));
1678 	return (DDI_SUCCESS);
1679 
1680 failure:
1681 	kmem_free(newregs, rlen + sizeof (pci_regspec_t));
1682 	kmem_free(regs, rlen);
1683 	return (DDI_FAILURE);
1684 }
1685 
1686 static uint32_t
1687 pci_type_ra2pci(char *type)
1688 {
1689 	uint32_t	pci_type = PCI_ADDR_TYPE_INVAL;
1690 
1691 	/*
1692 	 * No 64 bit mem support for now
1693 	 */
1694 	if (strcmp(type, NDI_RA_TYPE_IO) == 0) {
1695 		pci_type = PCI_ADDR_IO;
1696 
1697 	} else if (strcmp(type, NDI_RA_TYPE_MEM) == 0) {
1698 		pci_type = PCI_ADDR_MEM32;
1699 
1700 	} else if (strcmp(type, NDI_RA_TYPE_PCI_PREFETCH_MEM)  == 0) {
1701 		pci_type = PCI_ADDR_MEM32;
1702 		pci_type |= PCI_REG_PF_M;
1703 	}
1704 
1705 	return (pci_type);
1706 }
1707