xref: /illumos-gate/usr/src/uts/common/io/bofi.c (revision 43449cdcd0600512dd862537f2cf014140dd0844)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 /*
22  * Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
23  * Use is subject to license terms.
24  */
25 /*
26  * Copyright 2012 Garrett D'Amore <garrett@damore.org>.  All rights reserved.
27  */
28 
29 
30 #include <sys/types.h>
31 #include <sys/sysmacros.h>
32 #include <sys/buf.h>
33 #include <sys/errno.h>
34 #include <sys/modctl.h>
35 #include <sys/conf.h>
36 #include <sys/stat.h>
37 #include <sys/kmem.h>
38 #include <sys/proc.h>
39 #include <sys/cpuvar.h>
40 #include <sys/ddi_impldefs.h>
41 #include <sys/ddi.h>
42 #include <sys/fm/protocol.h>
43 #include <sys/fm/util.h>
44 #include <sys/fm/io/ddi.h>
45 #include <sys/sysevent/eventdefs.h>
46 #include <sys/sunddi.h>
47 #include <sys/sunndi.h>
48 #include <sys/debug.h>
49 #include <sys/bofi.h>
50 #include <sys/dvma.h>
51 #include <sys/bofi_impl.h>
52 
53 /*
54  * Testing the resilience of a hardened device driver requires a suitably wide
55  * range of different types of "typical" hardware faults to be injected,
56  * preferably in a controlled and repeatable fashion. This is not in general
57  * possible via hardware, so the "fault injection test harness" is provided.
58  * This works by intercepting calls from the driver to various DDI routines,
59  * and then corrupting the result of those DDI routine calls as if the
60  * hardware had caused the corruption.
61  *
62  * Conceptually, the bofi driver consists of two parts:
63  *
64  * A driver interface that supports a number of ioctls which allow error
65  * definitions ("errdefs") to be defined and subsequently managed. The
66  * driver is a clone driver, so each open will create a separate
67  * invocation. Any errdefs created by using ioctls to that invocation
68  * will automatically be deleted when that invocation is closed.
69  *
70  * Intercept routines: When the bofi driver is attached, it edits the
71  * bus_ops structure of the bus nexus specified by the "bofi-nexus"
72  * field in the "bofi.conf" file, thus allowing the
73  * bofi driver to intercept various ddi functions. These intercept
74  * routines primarily carry out fault injections based on the errdefs
75  * created for that device.
76  *
77  * Faults can be injected into:
78  *
79  * DMA (corrupting data for DMA to/from memory areas defined by
80  * ddi_dma_setup(), ddi_dma_bind_handle(), etc)
81  *
82  * Physical IO (corrupting data sent/received via ddi_get8(), ddi_put8(),
83  * etc),
84  *
85  * Interrupts (generating spurious interrupts, losing interrupts,
86  * delaying interrupts).
87  *
88  * By default, ddi routines called from all drivers will be intercepted
89  * and faults potentially injected. However, the "bofi-to-test" field in
90  * the "bofi.conf" file can be set to a space-separated list of drivers to
91  * test (or by preceding each driver name in the list with an "!", a list
92  * of drivers not to test).
93  *
94  * In addition to fault injection, the bofi driver does a number of static
95  * checks which are controlled by properties in the "bofi.conf" file.
96  *
97  * "bofi-ddi-check" - if set will validate that there are no PIO access
98  * other than those using the DDI routines (ddi_get8(), ddi_put8(), etc).
99  *
100  * "bofi-range-check" - if set to values 1 (warning) or 2 (panic), will
101  * validate that calls to ddi_get8(), ddi_put8(), etc are not made
102  * specifying addresses outside the range of the access_handle.
103  *
104  * "bofi-sync-check" - if set will validate that calls to ddi_dma_sync()
105  * are being made correctly.
106  */
107 
108 extern void *bp_mapin_common(struct buf *, int);
109 
110 static int bofi_ddi_check;
111 static int bofi_sync_check;
112 static int bofi_range_check;
113 
114 static struct bofi_link bofi_link_array[BOFI_NLINKS], *bofi_link_freelist;
115 
116 #define	LLSZMASK (sizeof (uint64_t)-1)
117 
118 #define	HDL_HASH_TBL_SIZE 64
119 static struct bofi_shadow hhash_table[HDL_HASH_TBL_SIZE];
120 static struct bofi_shadow dhash_table[HDL_HASH_TBL_SIZE];
121 #define	HDL_DHASH(x) \
122 	(&dhash_table[((uintptr_t)(x) >> 3) & (HDL_HASH_TBL_SIZE-1)])
123 #define	HDL_HHASH(x) \
124 	(&hhash_table[((uintptr_t)(x) >> 5) & (HDL_HASH_TBL_SIZE-1)])
125 
126 static struct bofi_shadow shadow_list;
127 static struct bofi_errent *errent_listp;
128 
129 static char driver_list[NAMESIZE];
130 static int driver_list_size;
131 static int driver_list_neg;
132 static char nexus_name[NAMESIZE];
133 
134 static int initialized = 0;
135 
136 #define	NCLONES 2560
137 static int clone_tab[NCLONES];
138 
139 static dev_info_t *our_dip;
140 
141 static kmutex_t bofi_mutex;
142 static kmutex_t clone_tab_mutex;
143 static kmutex_t bofi_low_mutex;
144 static ddi_iblock_cookie_t bofi_low_cookie;
145 static uint_t	bofi_signal(caddr_t arg);
146 static int	bofi_getinfo(dev_info_t *, ddi_info_cmd_t, void *, void **);
147 static int	bofi_attach(dev_info_t *, ddi_attach_cmd_t);
148 static int	bofi_detach(dev_info_t *, ddi_detach_cmd_t);
149 static int	bofi_open(dev_t *, int, int, cred_t *);
150 static int	bofi_close(dev_t, int, int, cred_t *);
151 static int	bofi_ioctl(dev_t, int, intptr_t, int, cred_t *, int *);
152 static int	bofi_errdef_alloc(struct bofi_errdef *, char *,
153 		    struct bofi_errent *);
154 static int	bofi_errdef_free(struct bofi_errent *);
155 static void	bofi_start(struct bofi_errctl *, char *);
156 static void	bofi_stop(struct bofi_errctl *, char *);
157 static void	bofi_broadcast(struct bofi_errctl *, char *);
158 static void	bofi_clear_acc_chk(struct bofi_errctl *, char *);
159 static void	bofi_clear_errors(struct bofi_errctl *, char *);
160 static void	bofi_clear_errdefs(struct bofi_errctl *, char *);
161 static int	bofi_errdef_check(struct bofi_errstate *,
162 		    struct acc_log_elem **);
163 static int	bofi_errdef_check_w(struct bofi_errstate *,
164 		    struct acc_log_elem **);
165 static int	bofi_map(dev_info_t *, dev_info_t *, ddi_map_req_t *,
166 		    off_t, off_t, caddr_t *);
167 static int	bofi_dma_allochdl(dev_info_t *, dev_info_t *,
168 		    ddi_dma_attr_t *, int (*)(caddr_t), caddr_t,
169 		    ddi_dma_handle_t *);
170 static int	bofi_dma_freehdl(dev_info_t *, dev_info_t *,
171 		    ddi_dma_handle_t);
172 static int	bofi_dma_bindhdl(dev_info_t *, dev_info_t *,
173 		    ddi_dma_handle_t, struct ddi_dma_req *, ddi_dma_cookie_t *,
174 		    uint_t *);
175 static int	bofi_dma_unbindhdl(dev_info_t *, dev_info_t *,
176 		    ddi_dma_handle_t);
177 static int	bofi_dma_flush(dev_info_t *, dev_info_t *, ddi_dma_handle_t,
178 		    off_t, size_t, uint_t);
179 static int	bofi_dma_ctl(dev_info_t *, dev_info_t *, ddi_dma_handle_t,
180 		    enum ddi_dma_ctlops, off_t *, size_t *, caddr_t *, uint_t);
181 static int	bofi_dma_win(dev_info_t *, dev_info_t *, ddi_dma_handle_t,
182 		    uint_t, off_t *, size_t *, ddi_dma_cookie_t *, uint_t *);
183 static int	bofi_intr_ops(dev_info_t *dip, dev_info_t *rdip,
184 		    ddi_intr_op_t intr_op, ddi_intr_handle_impl_t *hdlp,
185 		    void *result);
186 static int	bofi_fm_ereport_callback(sysevent_t *ev, void *cookie);
187 
188 evchan_t *bofi_error_chan;
189 
190 #define	FM_SIMULATED_DMA "simulated.dma"
191 #define	FM_SIMULATED_PIO "simulated.pio"
192 
193 #if defined(__sparc)
194 static void	bofi_dvma_kaddr_load(ddi_dma_handle_t, caddr_t, uint_t,
195 		    uint_t, ddi_dma_cookie_t *);
196 static void	bofi_dvma_unload(ddi_dma_handle_t, uint_t, uint_t);
197 static void	bofi_dvma_sync(ddi_dma_handle_t, uint_t, uint_t);
198 static void	bofi_dvma_reserve(dev_info_t *, ddi_dma_handle_t);
199 #endif
200 static int	driver_under_test(dev_info_t *);
201 static int	bofi_check_acc_hdl(ddi_acc_impl_t *);
202 static int	bofi_check_dma_hdl(ddi_dma_impl_t *);
203 static int	bofi_post_event(dev_info_t *dip, dev_info_t *rdip,
204 		    ddi_eventcookie_t eventhdl, void *impl_data);
205 
206 static struct bus_ops bofi_bus_ops = {
207 	BUSO_REV,
208 	bofi_map,
209 	NULL,
210 	NULL,
211 	NULL,
212 	i_ddi_map_fault,
213 	NULL,
214 	bofi_dma_allochdl,
215 	bofi_dma_freehdl,
216 	bofi_dma_bindhdl,
217 	bofi_dma_unbindhdl,
218 	bofi_dma_flush,
219 	bofi_dma_win,
220 	bofi_dma_ctl,
221 	NULL,
222 	ddi_bus_prop_op,
223 	ndi_busop_get_eventcookie,
224 	ndi_busop_add_eventcall,
225 	ndi_busop_remove_eventcall,
226 	bofi_post_event,
227 	NULL,
228 	0,
229 	0,
230 	0,
231 	0,
232 	0,
233 	0,
234 	0,
235 	bofi_intr_ops
236 };
237 
238 static struct cb_ops bofi_cb_ops = {
239 	bofi_open,		/* open */
240 	bofi_close,		/* close */
241 	nodev,			/* strategy */
242 	nodev,			/* print */
243 	nodev,			/* dump */
244 	nodev,			/* read */
245 	nodev,			/* write */
246 	bofi_ioctl,		/* ioctl */
247 	nodev,			/* devmap */
248 	nodev,			/* mmap */
249 	nodev,			/* segmap */
250 	nochpoll,		/* chpoll */
251 	ddi_prop_op,		/* prop_op */
252 	NULL,			/* for STREAMS drivers */
253 	D_MP,			/* driver compatibility flag */
254 	CB_REV,			/* cb_ops revision */
255 	nodev,			/* aread */
256 	nodev			/* awrite */
257 };
258 
259 static struct dev_ops bofi_ops = {
260 	DEVO_REV,		/* driver build version */
261 	0,			/* device reference count */
262 	bofi_getinfo,
263 	nulldev,
264 	nulldev,		/* probe */
265 	bofi_attach,
266 	bofi_detach,
267 	nulldev,		/* reset */
268 	&bofi_cb_ops,
269 	(struct bus_ops *)NULL,
270 	nulldev,		/* power */
271 	ddi_quiesce_not_needed,		/* quiesce */
272 };
273 
274 /* module configuration stuff */
275 static void    *statep;
276 
277 static struct modldrv modldrv = {
278 	&mod_driverops,
279 	"bofi driver",
280 	&bofi_ops
281 };
282 
283 static struct modlinkage modlinkage = {
284 	MODREV_1,
285 	&modldrv,
286 	0
287 };
288 
289 static struct bus_ops save_bus_ops;
290 
291 #if defined(__sparc)
292 static struct dvma_ops bofi_dvma_ops = {
293 	DVMAO_REV,
294 	bofi_dvma_kaddr_load,
295 	bofi_dvma_unload,
296 	bofi_dvma_sync
297 };
298 #endif
299 
300 /*
301  * support routine - map user page into kernel virtual
302  */
303 static caddr_t
304 dmareq_mapin(offset_t len, caddr_t addr, struct as *as, int flag)
305 {
306 	struct buf buf;
307 	struct proc proc;
308 
309 	/*
310 	 * mock up a buf structure so we can call bp_mapin_common()
311 	 */
312 	buf.b_flags = B_PHYS;
313 	buf.b_un.b_addr = (caddr_t)addr;
314 	buf.b_bcount = (size_t)len;
315 	proc.p_as = as;
316 	buf.b_proc = &proc;
317 	return (bp_mapin_common(&buf, flag));
318 }
319 
320 
321 /*
322  * support routine - map page chain into kernel virtual
323  */
324 static caddr_t
325 dmareq_pp_mapin(offset_t len, uint_t offset, page_t *pp, int flag)
326 {
327 	struct buf buf;
328 
329 	/*
330 	 * mock up a buf structure so we can call bp_mapin_common()
331 	 */
332 	buf.b_flags = B_PAGEIO;
333 	buf.b_un.b_addr = (caddr_t)(uintptr_t)offset;
334 	buf.b_bcount = (size_t)len;
335 	buf.b_pages = pp;
336 	return (bp_mapin_common(&buf, flag));
337 }
338 
339 
340 /*
341  * support routine - map page array into kernel virtual
342  */
343 static caddr_t
344 dmareq_pplist_mapin(uint_t len, caddr_t addr, page_t **pplist, struct as *as,
345     int flag)
346 {
347 	struct buf buf;
348 	struct proc proc;
349 
350 	/*
351 	 * mock up a buf structure so we can call bp_mapin_common()
352 	 */
353 	buf.b_flags = B_PHYS|B_SHADOW;
354 	buf.b_un.b_addr = addr;
355 	buf.b_bcount = len;
356 	buf.b_shadow = pplist;
357 	proc.p_as = as;
358 	buf.b_proc = &proc;
359 	return (bp_mapin_common(&buf, flag));
360 }
361 
362 
363 /*
364  * support routine - map dmareq into kernel virtual if not already
365  * fills in *lenp with length
366  * *mapaddr will be new kernel virtual address - or null if no mapping needed
367  */
368 static caddr_t
369 ddi_dmareq_mapin(struct ddi_dma_req *dmareqp, caddr_t *mapaddrp,
370     offset_t *lenp)
371 {
372 	int sleep = (dmareqp->dmar_fp == DDI_DMA_SLEEP) ? VM_SLEEP: VM_NOSLEEP;
373 
374 	*lenp = dmareqp->dmar_object.dmao_size;
375 	if (dmareqp->dmar_object.dmao_type == DMA_OTYP_PAGES) {
376 		*mapaddrp = dmareq_pp_mapin(dmareqp->dmar_object.dmao_size,
377 		    dmareqp->dmar_object.dmao_obj.pp_obj.pp_offset,
378 		    dmareqp->dmar_object.dmao_obj.pp_obj.pp_pp, sleep);
379 		return (*mapaddrp);
380 	} else if (dmareqp->dmar_object.dmao_obj.virt_obj.v_priv != NULL) {
381 		*mapaddrp = dmareq_pplist_mapin(dmareqp->dmar_object.dmao_size,
382 		    dmareqp->dmar_object.dmao_obj.virt_obj.v_addr,
383 		    dmareqp->dmar_object.dmao_obj.virt_obj.v_priv,
384 		    dmareqp->dmar_object.dmao_obj.virt_obj.v_as, sleep);
385 		return (*mapaddrp);
386 	} else if (dmareqp->dmar_object.dmao_obj.virt_obj.v_as == &kas) {
387 		*mapaddrp = NULL;
388 		return (dmareqp->dmar_object.dmao_obj.virt_obj.v_addr);
389 	} else if (dmareqp->dmar_object.dmao_obj.virt_obj.v_as == NULL) {
390 		*mapaddrp = NULL;
391 		return (dmareqp->dmar_object.dmao_obj.virt_obj.v_addr);
392 	} else {
393 		*mapaddrp = dmareq_mapin(dmareqp->dmar_object.dmao_size,
394 		    dmareqp->dmar_object.dmao_obj.virt_obj.v_addr,
395 		    dmareqp->dmar_object.dmao_obj.virt_obj.v_as, sleep);
396 		return (*mapaddrp);
397 	}
398 }
399 
400 
401 /*
402  * support routine - free off kernel virtual mapping as allocated by
403  * ddi_dmareq_mapin()
404  */
405 static void
406 ddi_dmareq_mapout(caddr_t addr, offset_t len, int map_flags, page_t *pp,
407     page_t **pplist)
408 {
409 	struct buf buf;
410 
411 	if (addr == NULL)
412 		return;
413 	/*
414 	 * mock up a buf structure
415 	 */
416 	buf.b_flags = B_REMAPPED | map_flags;
417 	buf.b_un.b_addr = addr;
418 	buf.b_bcount = (size_t)len;
419 	buf.b_pages = pp;
420 	buf.b_shadow = pplist;
421 	bp_mapout(&buf);
422 }
423 
424 static time_t
425 bofi_gettime()
426 {
427 	timestruc_t ts;
428 
429 	gethrestime(&ts);
430 	return (ts.tv_sec);
431 }
432 
433 /*
434  * reset the bus_ops structure of the specified nexus to point to
435  * the original values in the save_bus_ops structure.
436  *
437  * Note that both this routine and modify_bus_ops() rely on the current
438  * behavior of the framework in that nexus drivers are not unloadable
439  *
440  */
441 
442 static int
443 reset_bus_ops(char *name, struct bus_ops *bop)
444 {
445 	struct modctl *modp;
446 	struct modldrv *mp;
447 	struct bus_ops *bp;
448 	struct dev_ops *ops;
449 
450 	mutex_enter(&mod_lock);
451 	/*
452 	 * find specified module
453 	 */
454 	modp = &modules;
455 	do {
456 		if (strcmp(name, modp->mod_modname) == 0) {
457 			if (!modp->mod_linkage) {
458 				mutex_exit(&mod_lock);
459 				return (0);
460 			}
461 			mp = modp->mod_linkage->ml_linkage[0];
462 			if (!mp || !mp->drv_dev_ops) {
463 				mutex_exit(&mod_lock);
464 				return (0);
465 			}
466 			ops = mp->drv_dev_ops;
467 			bp = ops->devo_bus_ops;
468 			if (!bp) {
469 				mutex_exit(&mod_lock);
470 				return (0);
471 			}
472 			if (ops->devo_refcnt > 0) {
473 				/*
474 				 * As long as devices are active with modified
475 				 * bus ops bofi must not go away. There may be
476 				 * drivers with modified access or dma handles.
477 				 */
478 				mutex_exit(&mod_lock);
479 				return (0);
480 			}
481 			cmn_err(CE_NOTE, "bofi reset bus_ops for %s",
482 			    mp->drv_linkinfo);
483 			bp->bus_intr_op = bop->bus_intr_op;
484 			bp->bus_post_event = bop->bus_post_event;
485 			bp->bus_map = bop->bus_map;
486 			bp->bus_dma_map = bop->bus_dma_map;
487 			bp->bus_dma_allochdl = bop->bus_dma_allochdl;
488 			bp->bus_dma_freehdl = bop->bus_dma_freehdl;
489 			bp->bus_dma_bindhdl = bop->bus_dma_bindhdl;
490 			bp->bus_dma_unbindhdl = bop->bus_dma_unbindhdl;
491 			bp->bus_dma_flush = bop->bus_dma_flush;
492 			bp->bus_dma_win = bop->bus_dma_win;
493 			bp->bus_dma_ctl = bop->bus_dma_ctl;
494 			mutex_exit(&mod_lock);
495 			return (1);
496 		}
497 	} while ((modp = modp->mod_next) != &modules);
498 	mutex_exit(&mod_lock);
499 	return (0);
500 }
501 
502 /*
503  * modify the bus_ops structure of the specified nexus to point to bofi
504  * routines, saving the original values in the save_bus_ops structure
505  */
506 
507 static int
508 modify_bus_ops(char *name, struct bus_ops *bop)
509 {
510 	struct modctl *modp;
511 	struct modldrv *mp;
512 	struct bus_ops *bp;
513 	struct dev_ops *ops;
514 
515 	if (ddi_name_to_major(name) == -1)
516 		return (0);
517 
518 	mutex_enter(&mod_lock);
519 	/*
520 	 * find specified module
521 	 */
522 	modp = &modules;
523 	do {
524 		if (strcmp(name, modp->mod_modname) == 0) {
525 			if (!modp->mod_linkage) {
526 				mutex_exit(&mod_lock);
527 				return (0);
528 			}
529 			mp = modp->mod_linkage->ml_linkage[0];
530 			if (!mp || !mp->drv_dev_ops) {
531 				mutex_exit(&mod_lock);
532 				return (0);
533 			}
534 			ops = mp->drv_dev_ops;
535 			bp = ops->devo_bus_ops;
536 			if (!bp) {
537 				mutex_exit(&mod_lock);
538 				return (0);
539 			}
540 			if (ops->devo_refcnt == 0) {
541 				/*
542 				 * If there is no device active for this
543 				 * module then there is nothing to do for bofi.
544 				 */
545 				mutex_exit(&mod_lock);
546 				return (0);
547 			}
548 			cmn_err(CE_NOTE, "bofi modify bus_ops for %s",
549 			    mp->drv_linkinfo);
550 			save_bus_ops = *bp;
551 			bp->bus_intr_op = bop->bus_intr_op;
552 			bp->bus_post_event = bop->bus_post_event;
553 			bp->bus_map = bop->bus_map;
554 			bp->bus_dma_map = bop->bus_dma_map;
555 			bp->bus_dma_allochdl = bop->bus_dma_allochdl;
556 			bp->bus_dma_freehdl = bop->bus_dma_freehdl;
557 			bp->bus_dma_bindhdl = bop->bus_dma_bindhdl;
558 			bp->bus_dma_unbindhdl = bop->bus_dma_unbindhdl;
559 			bp->bus_dma_flush = bop->bus_dma_flush;
560 			bp->bus_dma_win = bop->bus_dma_win;
561 			bp->bus_dma_ctl = bop->bus_dma_ctl;
562 			mutex_exit(&mod_lock);
563 			return (1);
564 		}
565 	} while ((modp = modp->mod_next) != &modules);
566 	mutex_exit(&mod_lock);
567 	return (0);
568 }
569 
570 
571 int
572 _init(void)
573 {
574 	int    e;
575 
576 	e = ddi_soft_state_init(&statep, sizeof (struct bofi_errent), 1);
577 	if (e != 0)
578 		return (e);
579 	if ((e = mod_install(&modlinkage)) != 0)
580 		ddi_soft_state_fini(&statep);
581 	return (e);
582 }
583 
584 
585 int
586 _fini(void)
587 {
588 	int e;
589 
590 	if ((e = mod_remove(&modlinkage)) != 0)
591 		return (e);
592 	ddi_soft_state_fini(&statep);
593 	return (e);
594 }
595 
596 
597 int
598 _info(struct modinfo *modinfop)
599 {
600 	return (mod_info(&modlinkage, modinfop));
601 }
602 
603 
604 static int
605 bofi_attach(dev_info_t *dip, ddi_attach_cmd_t cmd)
606 {
607 	char *name;
608 	char buf[80];
609 	int i;
610 	int s, ss;
611 	int size = NAMESIZE;
612 	int new_string;
613 	char *ptr;
614 
615 	if (cmd != DDI_ATTACH)
616 		return (DDI_FAILURE);
617 	/*
618 	 * only one instance - but we clone using the open routine
619 	 */
620 	if (ddi_get_instance(dip) > 0)
621 		return (DDI_FAILURE);
622 
623 	if (!initialized) {
624 		if ((name = ddi_get_name(dip)) == NULL)
625 			return (DDI_FAILURE);
626 		(void) snprintf(buf, sizeof (buf), "%s,ctl", name);
627 		if (ddi_create_minor_node(dip, buf, S_IFCHR, 0,
628 		    DDI_PSEUDO, 0) == DDI_FAILURE)
629 			return (DDI_FAILURE);
630 
631 		if (ddi_get_soft_iblock_cookie(dip, DDI_SOFTINT_MED,
632 		    &bofi_low_cookie) != DDI_SUCCESS) {
633 			ddi_remove_minor_node(dip, buf);
634 			return (DDI_FAILURE); /* fail attach */
635 		}
636 		/*
637 		 * get nexus name (from conf file)
638 		 */
639 		if (ddi_prop_op(DDI_DEV_T_ANY, dip, PROP_LEN_AND_VAL_BUF, 0,
640 		    "bofi-nexus", nexus_name, &size) != DDI_PROP_SUCCESS) {
641 			ddi_remove_minor_node(dip, buf);
642 			return (DDI_FAILURE);
643 		}
644 		/*
645 		 * get whether to do dma map kmem private checking
646 		 */
647 		if ((bofi_range_check = ddi_prop_lookup_string(DDI_DEV_T_ANY,
648 		    dip, 0, "bofi-range-check", &ptr)) != DDI_PROP_SUCCESS)
649 			bofi_range_check = 0;
650 		else if (strcmp(ptr, "panic") == 0)
651 			bofi_range_check = 2;
652 		else if (strcmp(ptr, "warn") == 0)
653 			bofi_range_check = 1;
654 		else
655 			bofi_range_check = 0;
656 		ddi_prop_free(ptr);
657 
658 		/*
659 		 * get whether to prevent direct access to register
660 		 */
661 		if ((bofi_ddi_check = ddi_prop_lookup_string(DDI_DEV_T_ANY,
662 		    dip, 0, "bofi-ddi-check", &ptr)) != DDI_PROP_SUCCESS)
663 			bofi_ddi_check = 0;
664 		else if (strcmp(ptr, "on") == 0)
665 			bofi_ddi_check = 1;
666 		else
667 			bofi_ddi_check = 0;
668 		ddi_prop_free(ptr);
669 
670 		/*
671 		 * get whether to do copy on ddi_dma_sync
672 		 */
673 		if ((bofi_sync_check = ddi_prop_lookup_string(DDI_DEV_T_ANY,
674 		    dip, 0, "bofi-sync-check", &ptr)) != DDI_PROP_SUCCESS)
675 			bofi_sync_check = 0;
676 		else if (strcmp(ptr, "on") == 0)
677 			bofi_sync_check = 1;
678 		else
679 			bofi_sync_check = 0;
680 		ddi_prop_free(ptr);
681 
682 		/*
683 		 * get driver-under-test names (from conf file)
684 		 */
685 		size = NAMESIZE;
686 		if (ddi_prop_op(DDI_DEV_T_ANY, dip, PROP_LEN_AND_VAL_BUF, 0,
687 		    "bofi-to-test", driver_list, &size) != DDI_PROP_SUCCESS)
688 			driver_list[0] = 0;
689 		/*
690 		 * and convert into a sequence of strings
691 		 */
692 		driver_list_neg = 1;
693 		new_string = 1;
694 		driver_list_size = strlen(driver_list);
695 		for (i = 0; i < driver_list_size; i++) {
696 			if (driver_list[i] == ' ') {
697 				driver_list[i] = '\0';
698 				new_string = 1;
699 			} else if (new_string) {
700 				if (driver_list[i] != '!')
701 					driver_list_neg = 0;
702 				new_string = 0;
703 			}
704 		}
705 		/*
706 		 * initialize mutex, lists
707 		 */
708 		mutex_init(&clone_tab_mutex, NULL, MUTEX_DRIVER,
709 		    NULL);
710 		/*
711 		 * fake up iblock cookie - need to protect outselves
712 		 * against drivers that use hilevel interrupts
713 		 */
714 		ss = spl8();
715 		s = spl8();
716 		splx(ss);
717 		mutex_init(&bofi_mutex, NULL, MUTEX_SPIN, (void *)(uintptr_t)s);
718 		mutex_init(&bofi_low_mutex, NULL, MUTEX_DRIVER,
719 		    (void *)bofi_low_cookie);
720 		shadow_list.next = &shadow_list;
721 		shadow_list.prev = &shadow_list;
722 		for (i = 0; i < HDL_HASH_TBL_SIZE; i++) {
723 			hhash_table[i].hnext = &hhash_table[i];
724 			hhash_table[i].hprev = &hhash_table[i];
725 			dhash_table[i].dnext = &dhash_table[i];
726 			dhash_table[i].dprev = &dhash_table[i];
727 		}
728 		for (i = 1; i < BOFI_NLINKS; i++)
729 			bofi_link_array[i].link = &bofi_link_array[i-1];
730 		bofi_link_freelist = &bofi_link_array[BOFI_NLINKS - 1];
731 		/*
732 		 * overlay bus_ops structure
733 		 */
734 		if (modify_bus_ops(nexus_name, &bofi_bus_ops) == 0) {
735 			ddi_remove_minor_node(dip, buf);
736 			mutex_destroy(&clone_tab_mutex);
737 			mutex_destroy(&bofi_mutex);
738 			mutex_destroy(&bofi_low_mutex);
739 			return (DDI_FAILURE);
740 		}
741 		if (sysevent_evc_bind(FM_ERROR_CHAN, &bofi_error_chan, 0) == 0)
742 			(void) sysevent_evc_subscribe(bofi_error_chan, "bofi",
743 			    EC_FM, bofi_fm_ereport_callback, NULL, 0);
744 
745 		/*
746 		 * save dip for getinfo
747 		 */
748 		our_dip = dip;
749 		ddi_report_dev(dip);
750 		initialized = 1;
751 	}
752 	return (DDI_SUCCESS);
753 }
754 
755 
756 static int
757 bofi_detach(dev_info_t *dip, ddi_detach_cmd_t cmd)
758 {
759 	char *name;
760 	char buf[80];
761 
762 	if (cmd != DDI_DETACH)
763 		return (DDI_FAILURE);
764 	if (ddi_get_instance(dip) > 0)
765 		return (DDI_FAILURE);
766 	if ((name = ddi_get_name(dip)) == NULL)
767 		return (DDI_FAILURE);
768 	(void) snprintf(buf, sizeof (buf), "%s,ctl", name);
769 	mutex_enter(&bofi_low_mutex);
770 	mutex_enter(&bofi_mutex);
771 	/*
772 	 * make sure test bofi is no longer in use
773 	 */
774 	if (shadow_list.next != &shadow_list || errent_listp != NULL) {
775 		mutex_exit(&bofi_mutex);
776 		mutex_exit(&bofi_low_mutex);
777 		return (DDI_FAILURE);
778 	}
779 	mutex_exit(&bofi_mutex);
780 	mutex_exit(&bofi_low_mutex);
781 
782 	/*
783 	 * restore bus_ops structure
784 	 */
785 	if (reset_bus_ops(nexus_name, &save_bus_ops) == 0)
786 		return (DDI_FAILURE);
787 
788 	(void) sysevent_evc_unbind(bofi_error_chan);
789 
790 	mutex_destroy(&clone_tab_mutex);
791 	mutex_destroy(&bofi_mutex);
792 	mutex_destroy(&bofi_low_mutex);
793 	ddi_remove_minor_node(dip, buf);
794 	our_dip = NULL;
795 	initialized = 0;
796 	return (DDI_SUCCESS);
797 }
798 
799 
800 /* ARGSUSED */
801 static int
802 bofi_getinfo(dev_info_t *dip, ddi_info_cmd_t cmd, void *arg, void **result)
803 {
804 	dev_t	dev = (dev_t)arg;
805 	int	minor = (int)getminor(dev);
806 	int	retval;
807 
808 	switch (cmd) {
809 	case DDI_INFO_DEVT2DEVINFO:
810 		if (minor != 0 || our_dip == NULL) {
811 			*result = (void *)NULL;
812 			retval = DDI_FAILURE;
813 		} else {
814 			*result = (void *)our_dip;
815 			retval = DDI_SUCCESS;
816 		}
817 		break;
818 	case DDI_INFO_DEVT2INSTANCE:
819 		*result = (void *)0;
820 		retval = DDI_SUCCESS;
821 		break;
822 	default:
823 		retval = DDI_FAILURE;
824 	}
825 	return (retval);
826 }
827 
828 
829 /* ARGSUSED */
830 static int
831 bofi_open(dev_t *devp, int flag, int otyp, cred_t *credp)
832 {
833 	int	minor = (int)getminor(*devp);
834 	struct bofi_errent *softc;
835 
836 	/*
837 	 * only allow open on minor=0 - the clone device
838 	 */
839 	if (minor != 0)
840 		return (ENXIO);
841 	/*
842 	 * fail if not attached
843 	 */
844 	if (!initialized)
845 		return (ENXIO);
846 	/*
847 	 * find a free slot and grab it
848 	 */
849 	mutex_enter(&clone_tab_mutex);
850 	for (minor = 1; minor < NCLONES; minor++) {
851 		if (clone_tab[minor] == 0) {
852 			clone_tab[minor] = 1;
853 			break;
854 		}
855 	}
856 	mutex_exit(&clone_tab_mutex);
857 	if (minor == NCLONES)
858 		return (EAGAIN);
859 	/*
860 	 * soft state structure for this clone is used to maintain a list
861 	 * of allocated errdefs so they can be freed on close
862 	 */
863 	if (ddi_soft_state_zalloc(statep, minor) != DDI_SUCCESS) {
864 		mutex_enter(&clone_tab_mutex);
865 		clone_tab[minor] = 0;
866 		mutex_exit(&clone_tab_mutex);
867 		return (EAGAIN);
868 	}
869 	softc = ddi_get_soft_state(statep, minor);
870 	softc->cnext = softc;
871 	softc->cprev = softc;
872 
873 	*devp = makedevice(getmajor(*devp), minor);
874 	return (0);
875 }
876 
877 
878 /* ARGSUSED */
879 static int
880 bofi_close(dev_t dev, int flag, int otyp, cred_t *credp)
881 {
882 	int	minor = (int)getminor(dev);
883 	struct bofi_errent *softc;
884 	struct bofi_errent *ep, *next_ep;
885 
886 	softc = ddi_get_soft_state(statep, minor);
887 	if (softc == NULL)
888 		return (ENXIO);
889 	/*
890 	 * find list of errdefs and free them off
891 	 */
892 	for (ep = softc->cnext; ep != softc; ) {
893 		next_ep = ep->cnext;
894 		(void) bofi_errdef_free(ep);
895 		ep = next_ep;
896 	}
897 	/*
898 	 * free clone tab slot
899 	 */
900 	mutex_enter(&clone_tab_mutex);
901 	clone_tab[minor] = 0;
902 	mutex_exit(&clone_tab_mutex);
903 
904 	ddi_soft_state_free(statep, minor);
905 	return (0);
906 }
907 
908 
909 /* ARGSUSED */
910 static int
911 bofi_ioctl(dev_t dev, int cmd, intptr_t arg, int mode, cred_t *credp,
912     int *rvalp)
913 {
914 	struct bofi_errent *softc;
915 	int	minor = (int)getminor(dev);
916 	struct bofi_errdef errdef;
917 	struct bofi_errctl errctl;
918 	struct bofi_errstate errstate;
919 	void *ed_handle;
920 	struct bofi_get_handles get_handles;
921 	struct bofi_get_hdl_info hdl_info;
922 	struct handle_info *hdlip;
923 	struct handle_info *hib;
924 
925 	char *buffer;
926 	char *bufptr;
927 	char *endbuf;
928 	int req_count, count, err;
929 	char *namep;
930 	struct bofi_shadow *hp;
931 	int retval;
932 	struct bofi_shadow *hhashp;
933 	int i;
934 
935 	switch (cmd) {
936 	case BOFI_ADD_DEF:
937 		/*
938 		 * add a new error definition
939 		 */
940 #ifdef _MULTI_DATAMODEL
941 		switch (ddi_model_convert_from(mode & FMODELS)) {
942 		case DDI_MODEL_ILP32:
943 		{
944 			/*
945 			 * For use when a 32 bit app makes a call into a
946 			 * 64 bit ioctl
947 			 */
948 			struct bofi_errdef32	errdef_32;
949 
950 			if (ddi_copyin((void *)arg, &errdef_32,
951 			    sizeof (struct bofi_errdef32), mode)) {
952 				return (EFAULT);
953 			}
954 			errdef.namesize = errdef_32.namesize;
955 			(void) strncpy(errdef.name, errdef_32.name, NAMESIZE);
956 			errdef.instance = errdef_32.instance;
957 			errdef.rnumber = errdef_32.rnumber;
958 			errdef.offset = errdef_32.offset;
959 			errdef.len = errdef_32.len;
960 			errdef.access_type = errdef_32.access_type;
961 			errdef.access_count = errdef_32.access_count;
962 			errdef.fail_count = errdef_32.fail_count;
963 			errdef.acc_chk = errdef_32.acc_chk;
964 			errdef.optype = errdef_32.optype;
965 			errdef.operand = errdef_32.operand;
966 			errdef.log.logsize = errdef_32.log.logsize;
967 			errdef.log.entries = errdef_32.log.entries;
968 			errdef.log.flags = errdef_32.log.flags;
969 			errdef.log.wrapcnt = errdef_32.log.wrapcnt;
970 			errdef.log.start_time = errdef_32.log.start_time;
971 			errdef.log.stop_time = errdef_32.log.stop_time;
972 			errdef.log.logbase =
973 			    (caddr_t)(uintptr_t)errdef_32.log.logbase;
974 			errdef.errdef_handle = errdef_32.errdef_handle;
975 			break;
976 		}
977 		case DDI_MODEL_NONE:
978 			if (ddi_copyin((void *)arg, &errdef,
979 			    sizeof (struct bofi_errdef), mode))
980 				return (EFAULT);
981 			break;
982 		}
983 #else /* ! _MULTI_DATAMODEL */
984 		if (ddi_copyin((void *)arg, &errdef,
985 		    sizeof (struct bofi_errdef), mode) != 0)
986 			return (EFAULT);
987 #endif /* _MULTI_DATAMODEL */
988 		/*
989 		 * do some validation
990 		 */
991 		if (errdef.fail_count == 0)
992 			errdef.optype = 0;
993 		if (errdef.optype != 0) {
994 			if (errdef.access_type & BOFI_INTR &&
995 			    errdef.optype != BOFI_DELAY_INTR &&
996 			    errdef.optype != BOFI_LOSE_INTR &&
997 			    errdef.optype != BOFI_EXTRA_INTR)
998 				return (EINVAL);
999 			if ((errdef.access_type & (BOFI_DMA_RW|BOFI_PIO_R)) &&
1000 			    errdef.optype == BOFI_NO_TRANSFER)
1001 				return (EINVAL);
1002 			if ((errdef.access_type & (BOFI_PIO_RW)) &&
1003 			    errdef.optype != BOFI_EQUAL &&
1004 			    errdef.optype != BOFI_OR &&
1005 			    errdef.optype != BOFI_XOR &&
1006 			    errdef.optype != BOFI_AND &&
1007 			    errdef.optype != BOFI_NO_TRANSFER)
1008 				return (EINVAL);
1009 		}
1010 		/*
1011 		 * find softstate for this clone, so we can tag
1012 		 * new errdef on to it
1013 		 */
1014 		softc = ddi_get_soft_state(statep, minor);
1015 		if (softc == NULL)
1016 			return (ENXIO);
1017 		/*
1018 		 * read in name
1019 		 */
1020 		if (errdef.namesize > NAMESIZE)
1021 			return (EINVAL);
1022 		namep = kmem_zalloc(errdef.namesize+1, KM_SLEEP);
1023 		(void) strncpy(namep, errdef.name, errdef.namesize);
1024 
1025 		if (bofi_errdef_alloc(&errdef, namep, softc) != DDI_SUCCESS) {
1026 			(void) bofi_errdef_free((struct bofi_errent *)
1027 			    (uintptr_t)errdef.errdef_handle);
1028 			kmem_free(namep, errdef.namesize+1);
1029 			return (EINVAL);
1030 		}
1031 		/*
1032 		 * copy out errdef again, including filled in errdef_handle
1033 		 */
1034 #ifdef _MULTI_DATAMODEL
1035 		switch (ddi_model_convert_from(mode & FMODELS)) {
1036 		case DDI_MODEL_ILP32:
1037 		{
1038 			/*
1039 			 * For use when a 32 bit app makes a call into a
1040 			 * 64 bit ioctl
1041 			 */
1042 			struct bofi_errdef32	errdef_32;
1043 
1044 			errdef_32.namesize = errdef.namesize;
1045 			(void) strncpy(errdef_32.name, errdef.name, NAMESIZE);
1046 			errdef_32.instance = errdef.instance;
1047 			errdef_32.rnumber = errdef.rnumber;
1048 			errdef_32.offset = errdef.offset;
1049 			errdef_32.len = errdef.len;
1050 			errdef_32.access_type = errdef.access_type;
1051 			errdef_32.access_count = errdef.access_count;
1052 			errdef_32.fail_count = errdef.fail_count;
1053 			errdef_32.acc_chk = errdef.acc_chk;
1054 			errdef_32.optype = errdef.optype;
1055 			errdef_32.operand = errdef.operand;
1056 			errdef_32.log.logsize = errdef.log.logsize;
1057 			errdef_32.log.entries = errdef.log.entries;
1058 			errdef_32.log.flags = errdef.log.flags;
1059 			errdef_32.log.wrapcnt = errdef.log.wrapcnt;
1060 			errdef_32.log.start_time = errdef.log.start_time;
1061 			errdef_32.log.stop_time = errdef.log.stop_time;
1062 			errdef_32.log.logbase =
1063 			    (caddr32_t)(uintptr_t)errdef.log.logbase;
1064 			errdef_32.errdef_handle = errdef.errdef_handle;
1065 			if (ddi_copyout(&errdef_32, (void *)arg,
1066 			    sizeof (struct bofi_errdef32), mode) != 0) {
1067 				(void) bofi_errdef_free((struct bofi_errent *)
1068 				    errdef.errdef_handle);
1069 				kmem_free(namep, errdef.namesize+1);
1070 				return (EFAULT);
1071 			}
1072 			break;
1073 		}
1074 		case DDI_MODEL_NONE:
1075 			if (ddi_copyout(&errdef, (void *)arg,
1076 			    sizeof (struct bofi_errdef), mode) != 0) {
1077 				(void) bofi_errdef_free((struct bofi_errent *)
1078 				    errdef.errdef_handle);
1079 				kmem_free(namep, errdef.namesize+1);
1080 				return (EFAULT);
1081 			}
1082 			break;
1083 		}
1084 #else /* ! _MULTI_DATAMODEL */
1085 		if (ddi_copyout(&errdef, (void *)arg,
1086 		    sizeof (struct bofi_errdef), mode) != 0) {
1087 			(void) bofi_errdef_free((struct bofi_errent *)
1088 			    (uintptr_t)errdef.errdef_handle);
1089 			kmem_free(namep, errdef.namesize+1);
1090 			return (EFAULT);
1091 		}
1092 #endif /* _MULTI_DATAMODEL */
1093 		return (0);
1094 	case BOFI_DEL_DEF:
1095 		/*
1096 		 * delete existing errdef
1097 		 */
1098 		if (ddi_copyin((void *)arg, &ed_handle,
1099 		    sizeof (void *), mode) != 0)
1100 			return (EFAULT);
1101 		return (bofi_errdef_free((struct bofi_errent *)ed_handle));
1102 	case BOFI_START:
1103 		/*
1104 		 * start all errdefs corresponding to
1105 		 * this name and instance
1106 		 */
1107 		if (ddi_copyin((void *)arg, &errctl,
1108 		    sizeof (struct bofi_errctl), mode) != 0)
1109 			return (EFAULT);
1110 		/*
1111 		 * copy in name
1112 		 */
1113 		if (errctl.namesize > NAMESIZE)
1114 			return (EINVAL);
1115 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1116 		(void) strncpy(namep, errctl.name, errctl.namesize);
1117 		bofi_start(&errctl, namep);
1118 		kmem_free(namep, errctl.namesize+1);
1119 		return (0);
1120 	case BOFI_STOP:
1121 		/*
1122 		 * stop all errdefs corresponding to
1123 		 * this name and instance
1124 		 */
1125 		if (ddi_copyin((void *)arg, &errctl,
1126 		    sizeof (struct bofi_errctl), mode) != 0)
1127 			return (EFAULT);
1128 		/*
1129 		 * copy in name
1130 		 */
1131 		if (errctl.namesize > NAMESIZE)
1132 			return (EINVAL);
1133 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1134 		(void) strncpy(namep, errctl.name, errctl.namesize);
1135 		bofi_stop(&errctl, namep);
1136 		kmem_free(namep, errctl.namesize+1);
1137 		return (0);
1138 	case BOFI_BROADCAST:
1139 		/*
1140 		 * wakeup all errdefs corresponding to
1141 		 * this name and instance
1142 		 */
1143 		if (ddi_copyin((void *)arg, &errctl,
1144 		    sizeof (struct bofi_errctl), mode) != 0)
1145 			return (EFAULT);
1146 		/*
1147 		 * copy in name
1148 		 */
1149 		if (errctl.namesize > NAMESIZE)
1150 			return (EINVAL);
1151 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1152 		(void) strncpy(namep, errctl.name, errctl.namesize);
1153 		bofi_broadcast(&errctl, namep);
1154 		kmem_free(namep, errctl.namesize+1);
1155 		return (0);
1156 	case BOFI_CLEAR_ACC_CHK:
1157 		/*
1158 		 * clear "acc_chk" for all errdefs corresponding to
1159 		 * this name and instance
1160 		 */
1161 		if (ddi_copyin((void *)arg, &errctl,
1162 		    sizeof (struct bofi_errctl), mode) != 0)
1163 			return (EFAULT);
1164 		/*
1165 		 * copy in name
1166 		 */
1167 		if (errctl.namesize > NAMESIZE)
1168 			return (EINVAL);
1169 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1170 		(void) strncpy(namep, errctl.name, errctl.namesize);
1171 		bofi_clear_acc_chk(&errctl, namep);
1172 		kmem_free(namep, errctl.namesize+1);
1173 		return (0);
1174 	case BOFI_CLEAR_ERRORS:
1175 		/*
1176 		 * set "fail_count" to 0 for all errdefs corresponding to
1177 		 * this name and instance whose "access_count"
1178 		 * has expired.
1179 		 */
1180 		if (ddi_copyin((void *)arg, &errctl,
1181 		    sizeof (struct bofi_errctl), mode) != 0)
1182 			return (EFAULT);
1183 		/*
1184 		 * copy in name
1185 		 */
1186 		if (errctl.namesize > NAMESIZE)
1187 			return (EINVAL);
1188 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1189 		(void) strncpy(namep, errctl.name, errctl.namesize);
1190 		bofi_clear_errors(&errctl, namep);
1191 		kmem_free(namep, errctl.namesize+1);
1192 		return (0);
1193 	case BOFI_CLEAR_ERRDEFS:
1194 		/*
1195 		 * set "access_count" and "fail_count" to 0 for all errdefs
1196 		 * corresponding to this name and instance
1197 		 */
1198 		if (ddi_copyin((void *)arg, &errctl,
1199 		    sizeof (struct bofi_errctl), mode) != 0)
1200 			return (EFAULT);
1201 		/*
1202 		 * copy in name
1203 		 */
1204 		if (errctl.namesize > NAMESIZE)
1205 			return (EINVAL);
1206 		namep = kmem_zalloc(errctl.namesize+1, KM_SLEEP);
1207 		(void) strncpy(namep, errctl.name, errctl.namesize);
1208 		bofi_clear_errdefs(&errctl, namep);
1209 		kmem_free(namep, errctl.namesize+1);
1210 		return (0);
1211 	case BOFI_CHK_STATE:
1212 	{
1213 		struct acc_log_elem *klg;
1214 		size_t uls;
1215 		/*
1216 		 * get state for this errdef - read in dummy errstate
1217 		 * with just the errdef_handle filled in
1218 		 */
1219 #ifdef _MULTI_DATAMODEL
1220 		switch (ddi_model_convert_from(mode & FMODELS)) {
1221 		case DDI_MODEL_ILP32:
1222 		{
1223 			/*
1224 			 * For use when a 32 bit app makes a call into a
1225 			 * 64 bit ioctl
1226 			 */
1227 			struct bofi_errstate32	errstate_32;
1228 
1229 			if (ddi_copyin((void *)arg, &errstate_32,
1230 			    sizeof (struct bofi_errstate32), mode) != 0) {
1231 				return (EFAULT);
1232 			}
1233 			errstate.fail_time = errstate_32.fail_time;
1234 			errstate.msg_time = errstate_32.msg_time;
1235 			errstate.access_count = errstate_32.access_count;
1236 			errstate.fail_count = errstate_32.fail_count;
1237 			errstate.acc_chk = errstate_32.acc_chk;
1238 			errstate.errmsg_count = errstate_32.errmsg_count;
1239 			(void) strncpy(errstate.buffer, errstate_32.buffer,
1240 			    ERRMSGSIZE);
1241 			errstate.severity = errstate_32.severity;
1242 			errstate.log.logsize = errstate_32.log.logsize;
1243 			errstate.log.entries = errstate_32.log.entries;
1244 			errstate.log.flags = errstate_32.log.flags;
1245 			errstate.log.wrapcnt = errstate_32.log.wrapcnt;
1246 			errstate.log.start_time = errstate_32.log.start_time;
1247 			errstate.log.stop_time = errstate_32.log.stop_time;
1248 			errstate.log.logbase =
1249 			    (caddr_t)(uintptr_t)errstate_32.log.logbase;
1250 			errstate.errdef_handle = errstate_32.errdef_handle;
1251 			break;
1252 		}
1253 		case DDI_MODEL_NONE:
1254 			if (ddi_copyin((void *)arg, &errstate,
1255 			    sizeof (struct bofi_errstate), mode) != 0)
1256 				return (EFAULT);
1257 			break;
1258 		}
1259 #else /* ! _MULTI_DATAMODEL */
1260 		if (ddi_copyin((void *)arg, &errstate,
1261 		    sizeof (struct bofi_errstate), mode) != 0)
1262 			return (EFAULT);
1263 #endif /* _MULTI_DATAMODEL */
1264 		if ((retval = bofi_errdef_check(&errstate, &klg)) == EINVAL)
1265 			return (EINVAL);
1266 		/*
1267 		 * copy out real errstate structure
1268 		 */
1269 		uls = errstate.log.logsize;
1270 		if (errstate.log.entries > uls && uls)
1271 			/* insufficient user memory */
1272 			errstate.log.entries = uls;
1273 		/* always pass back a time */
1274 		if (errstate.log.stop_time == 0ul)
1275 			(void) drv_getparm(TIME, &(errstate.log.stop_time));
1276 
1277 #ifdef _MULTI_DATAMODEL
1278 		switch (ddi_model_convert_from(mode & FMODELS)) {
1279 		case DDI_MODEL_ILP32:
1280 		{
1281 			/*
1282 			 * For use when a 32 bit app makes a call into a
1283 			 * 64 bit ioctl
1284 			 */
1285 			struct bofi_errstate32	errstate_32;
1286 
1287 			errstate_32.fail_time = errstate.fail_time;
1288 			errstate_32.msg_time = errstate.msg_time;
1289 			errstate_32.access_count = errstate.access_count;
1290 			errstate_32.fail_count = errstate.fail_count;
1291 			errstate_32.acc_chk = errstate.acc_chk;
1292 			errstate_32.errmsg_count = errstate.errmsg_count;
1293 			(void) strncpy(errstate_32.buffer, errstate.buffer,
1294 			    ERRMSGSIZE);
1295 			errstate_32.severity = errstate.severity;
1296 			errstate_32.log.logsize = errstate.log.logsize;
1297 			errstate_32.log.entries = errstate.log.entries;
1298 			errstate_32.log.flags = errstate.log.flags;
1299 			errstate_32.log.wrapcnt = errstate.log.wrapcnt;
1300 			errstate_32.log.start_time = errstate.log.start_time;
1301 			errstate_32.log.stop_time = errstate.log.stop_time;
1302 			errstate_32.log.logbase =
1303 			    (caddr32_t)(uintptr_t)errstate.log.logbase;
1304 			errstate_32.errdef_handle = errstate.errdef_handle;
1305 			if (ddi_copyout(&errstate_32, (void *)arg,
1306 			    sizeof (struct bofi_errstate32), mode) != 0)
1307 				return (EFAULT);
1308 			break;
1309 		}
1310 		case DDI_MODEL_NONE:
1311 			if (ddi_copyout(&errstate, (void *)arg,
1312 			    sizeof (struct bofi_errstate), mode) != 0)
1313 				return (EFAULT);
1314 			break;
1315 		}
1316 #else /* ! _MULTI_DATAMODEL */
1317 		if (ddi_copyout(&errstate, (void *)arg,
1318 		    sizeof (struct bofi_errstate), mode) != 0)
1319 			return (EFAULT);
1320 #endif /* _MULTI_DATAMODEL */
1321 		if (uls && errstate.log.entries &&
1322 		    ddi_copyout(klg, errstate.log.logbase,
1323 		    errstate.log.entries * sizeof (struct acc_log_elem),
1324 		    mode) != 0) {
1325 			return (EFAULT);
1326 		}
1327 		return (retval);
1328 	}
1329 	case BOFI_CHK_STATE_W:
1330 	{
1331 		struct acc_log_elem *klg;
1332 		size_t uls;
1333 		/*
1334 		 * get state for this errdef - read in dummy errstate
1335 		 * with just the errdef_handle filled in. Then wait for
1336 		 * a ddi_report_fault message to come back
1337 		 */
1338 #ifdef _MULTI_DATAMODEL
1339 		switch (ddi_model_convert_from(mode & FMODELS)) {
1340 		case DDI_MODEL_ILP32:
1341 		{
1342 			/*
1343 			 * For use when a 32 bit app makes a call into a
1344 			 * 64 bit ioctl
1345 			 */
1346 			struct bofi_errstate32	errstate_32;
1347 
1348 			if (ddi_copyin((void *)arg, &errstate_32,
1349 			    sizeof (struct bofi_errstate32), mode) != 0) {
1350 				return (EFAULT);
1351 			}
1352 			errstate.fail_time = errstate_32.fail_time;
1353 			errstate.msg_time = errstate_32.msg_time;
1354 			errstate.access_count = errstate_32.access_count;
1355 			errstate.fail_count = errstate_32.fail_count;
1356 			errstate.acc_chk = errstate_32.acc_chk;
1357 			errstate.errmsg_count = errstate_32.errmsg_count;
1358 			(void) strncpy(errstate.buffer, errstate_32.buffer,
1359 			    ERRMSGSIZE);
1360 			errstate.severity = errstate_32.severity;
1361 			errstate.log.logsize = errstate_32.log.logsize;
1362 			errstate.log.entries = errstate_32.log.entries;
1363 			errstate.log.flags = errstate_32.log.flags;
1364 			errstate.log.wrapcnt = errstate_32.log.wrapcnt;
1365 			errstate.log.start_time = errstate_32.log.start_time;
1366 			errstate.log.stop_time = errstate_32.log.stop_time;
1367 			errstate.log.logbase =
1368 			    (caddr_t)(uintptr_t)errstate_32.log.logbase;
1369 			errstate.errdef_handle = errstate_32.errdef_handle;
1370 			break;
1371 		}
1372 		case DDI_MODEL_NONE:
1373 			if (ddi_copyin((void *)arg, &errstate,
1374 			    sizeof (struct bofi_errstate), mode) != 0)
1375 				return (EFAULT);
1376 			break;
1377 		}
1378 #else /* ! _MULTI_DATAMODEL */
1379 		if (ddi_copyin((void *)arg, &errstate,
1380 		    sizeof (struct bofi_errstate), mode) != 0)
1381 			return (EFAULT);
1382 #endif /* _MULTI_DATAMODEL */
1383 		if ((retval = bofi_errdef_check_w(&errstate, &klg)) == EINVAL)
1384 			return (EINVAL);
1385 		/*
1386 		 * copy out real errstate structure
1387 		 */
1388 		uls = errstate.log.logsize;
1389 		uls = errstate.log.logsize;
1390 		if (errstate.log.entries > uls && uls)
1391 			/* insufficient user memory */
1392 			errstate.log.entries = uls;
1393 		/* always pass back a time */
1394 		if (errstate.log.stop_time == 0ul)
1395 			(void) drv_getparm(TIME, &(errstate.log.stop_time));
1396 
1397 #ifdef _MULTI_DATAMODEL
1398 		switch (ddi_model_convert_from(mode & FMODELS)) {
1399 		case DDI_MODEL_ILP32:
1400 		{
1401 			/*
1402 			 * For use when a 32 bit app makes a call into a
1403 			 * 64 bit ioctl
1404 			 */
1405 			struct bofi_errstate32	errstate_32;
1406 
1407 			errstate_32.fail_time = errstate.fail_time;
1408 			errstate_32.msg_time = errstate.msg_time;
1409 			errstate_32.access_count = errstate.access_count;
1410 			errstate_32.fail_count = errstate.fail_count;
1411 			errstate_32.acc_chk = errstate.acc_chk;
1412 			errstate_32.errmsg_count = errstate.errmsg_count;
1413 			(void) strncpy(errstate_32.buffer, errstate.buffer,
1414 			    ERRMSGSIZE);
1415 			errstate_32.severity = errstate.severity;
1416 			errstate_32.log.logsize = errstate.log.logsize;
1417 			errstate_32.log.entries = errstate.log.entries;
1418 			errstate_32.log.flags = errstate.log.flags;
1419 			errstate_32.log.wrapcnt = errstate.log.wrapcnt;
1420 			errstate_32.log.start_time = errstate.log.start_time;
1421 			errstate_32.log.stop_time = errstate.log.stop_time;
1422 			errstate_32.log.logbase =
1423 			    (caddr32_t)(uintptr_t)errstate.log.logbase;
1424 			errstate_32.errdef_handle = errstate.errdef_handle;
1425 			if (ddi_copyout(&errstate_32, (void *)arg,
1426 			    sizeof (struct bofi_errstate32), mode) != 0)
1427 				return (EFAULT);
1428 			break;
1429 		}
1430 		case DDI_MODEL_NONE:
1431 			if (ddi_copyout(&errstate, (void *)arg,
1432 			    sizeof (struct bofi_errstate), mode) != 0)
1433 				return (EFAULT);
1434 			break;
1435 		}
1436 #else /* ! _MULTI_DATAMODEL */
1437 		if (ddi_copyout(&errstate, (void *)arg,
1438 		    sizeof (struct bofi_errstate), mode) != 0)
1439 			return (EFAULT);
1440 #endif /* _MULTI_DATAMODEL */
1441 
1442 		if (uls && errstate.log.entries &&
1443 		    ddi_copyout(klg, errstate.log.logbase,
1444 		    errstate.log.entries * sizeof (struct acc_log_elem),
1445 		    mode) != 0) {
1446 			return (EFAULT);
1447 		}
1448 		return (retval);
1449 	}
1450 	case BOFI_GET_HANDLES:
1451 		/*
1452 		 * display existing handles
1453 		 */
1454 #ifdef _MULTI_DATAMODEL
1455 		switch (ddi_model_convert_from(mode & FMODELS)) {
1456 		case DDI_MODEL_ILP32:
1457 		{
1458 			/*
1459 			 * For use when a 32 bit app makes a call into a
1460 			 * 64 bit ioctl
1461 			 */
1462 			struct bofi_get_handles32	get_handles_32;
1463 
1464 			if (ddi_copyin((void *)arg, &get_handles_32,
1465 			    sizeof (get_handles_32), mode) != 0) {
1466 				return (EFAULT);
1467 			}
1468 			get_handles.namesize = get_handles_32.namesize;
1469 			(void) strncpy(get_handles.name, get_handles_32.name,
1470 			    NAMESIZE);
1471 			get_handles.instance = get_handles_32.instance;
1472 			get_handles.count = get_handles_32.count;
1473 			get_handles.buffer =
1474 			    (caddr_t)(uintptr_t)get_handles_32.buffer;
1475 			break;
1476 		}
1477 		case DDI_MODEL_NONE:
1478 			if (ddi_copyin((void *)arg, &get_handles,
1479 			    sizeof (get_handles), mode) != 0)
1480 				return (EFAULT);
1481 			break;
1482 		}
1483 #else /* ! _MULTI_DATAMODEL */
1484 		if (ddi_copyin((void *)arg, &get_handles,
1485 		    sizeof (get_handles), mode) != 0)
1486 			return (EFAULT);
1487 #endif /* _MULTI_DATAMODEL */
1488 		/*
1489 		 * read in name
1490 		 */
1491 		if (get_handles.namesize > NAMESIZE)
1492 			return (EINVAL);
1493 		namep = kmem_zalloc(get_handles.namesize+1, KM_SLEEP);
1494 		(void) strncpy(namep, get_handles.name, get_handles.namesize);
1495 		req_count = get_handles.count;
1496 		bufptr = buffer = kmem_zalloc(req_count, KM_SLEEP);
1497 		endbuf = bufptr + req_count;
1498 		/*
1499 		 * display existing handles
1500 		 */
1501 		mutex_enter(&bofi_low_mutex);
1502 		mutex_enter(&bofi_mutex);
1503 		for (i = 0; i < HDL_HASH_TBL_SIZE; i++) {
1504 			hhashp = &hhash_table[i];
1505 			for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext) {
1506 				if (!driver_under_test(hp->dip))
1507 					continue;
1508 				if (ddi_name_to_major(ddi_get_name(hp->dip)) !=
1509 				    ddi_name_to_major(namep))
1510 					continue;
1511 				if (hp->instance != get_handles.instance)
1512 					continue;
1513 				/*
1514 				 * print information per handle - note that
1515 				 * DMA* means an unbound DMA handle
1516 				 */
1517 				(void) snprintf(bufptr, (size_t)(endbuf-bufptr),
1518 				    "  %s %d %s ", hp->name, hp->instance,
1519 				    (hp->type == BOFI_INT_HDL) ? "INTR" :
1520 				    (hp->type == BOFI_ACC_HDL) ? "PIO" :
1521 				    (hp->type == BOFI_DMA_HDL) ? "DMA" :
1522 				    (hp->hparrayp != NULL) ? "DVMA" : "DMA*");
1523 				bufptr += strlen(bufptr);
1524 				if (hp->type == BOFI_ACC_HDL) {
1525 					if (hp->len == INT_MAX - hp->offset)
1526 						(void) snprintf(bufptr,
1527 						    (size_t)(endbuf-bufptr),
1528 						    "reg set %d off 0x%llx\n",
1529 						    hp->rnumber, hp->offset);
1530 					else
1531 						(void) snprintf(bufptr,
1532 						    (size_t)(endbuf-bufptr),
1533 						    "reg set %d off 0x%llx"
1534 						    " len 0x%llx\n",
1535 						    hp->rnumber, hp->offset,
1536 						    hp->len);
1537 				} else if (hp->type == BOFI_DMA_HDL)
1538 					(void) snprintf(bufptr,
1539 					    (size_t)(endbuf-bufptr),
1540 					    "handle no %d len 0x%llx"
1541 					    " addr 0x%p\n", hp->rnumber,
1542 					    hp->len, (void *)hp->addr);
1543 				else if (hp->type == BOFI_NULL &&
1544 				    hp->hparrayp == NULL)
1545 					(void) snprintf(bufptr,
1546 					    (size_t)(endbuf-bufptr),
1547 					    "handle no %d\n", hp->rnumber);
1548 				else
1549 					(void) snprintf(bufptr,
1550 					    (size_t)(endbuf-bufptr), "\n");
1551 				bufptr += strlen(bufptr);
1552 			}
1553 		}
1554 		mutex_exit(&bofi_mutex);
1555 		mutex_exit(&bofi_low_mutex);
1556 		err = ddi_copyout(buffer, get_handles.buffer, req_count, mode);
1557 		kmem_free(namep, get_handles.namesize+1);
1558 		kmem_free(buffer, req_count);
1559 		if (err != 0)
1560 			return (EFAULT);
1561 		else
1562 			return (0);
1563 	case BOFI_GET_HANDLE_INFO:
1564 		/*
1565 		 * display existing handles
1566 		 */
1567 #ifdef _MULTI_DATAMODEL
1568 		switch (ddi_model_convert_from(mode & FMODELS)) {
1569 		case DDI_MODEL_ILP32:
1570 		{
1571 			/*
1572 			 * For use when a 32 bit app makes a call into a
1573 			 * 64 bit ioctl
1574 			 */
1575 			struct bofi_get_hdl_info32	hdl_info_32;
1576 
1577 			if (ddi_copyin((void *)arg, &hdl_info_32,
1578 			    sizeof (hdl_info_32), mode)) {
1579 				return (EFAULT);
1580 			}
1581 			hdl_info.namesize = hdl_info_32.namesize;
1582 			(void) strncpy(hdl_info.name, hdl_info_32.name,
1583 			    NAMESIZE);
1584 			hdl_info.count = hdl_info_32.count;
1585 			hdl_info.hdli = (caddr_t)(uintptr_t)hdl_info_32.hdli;
1586 			break;
1587 		}
1588 		case DDI_MODEL_NONE:
1589 			if (ddi_copyin((void *)arg, &hdl_info,
1590 			    sizeof (hdl_info), mode))
1591 				return (EFAULT);
1592 			break;
1593 		}
1594 #else /* ! _MULTI_DATAMODEL */
1595 		if (ddi_copyin((void *)arg, &hdl_info,
1596 		    sizeof (hdl_info), mode))
1597 			return (EFAULT);
1598 #endif /* _MULTI_DATAMODEL */
1599 		if (hdl_info.namesize > NAMESIZE)
1600 			return (EINVAL);
1601 		namep = kmem_zalloc(hdl_info.namesize + 1, KM_SLEEP);
1602 		(void) strncpy(namep, hdl_info.name, hdl_info.namesize);
1603 		req_count = hdl_info.count;
1604 		count = hdl_info.count = 0; /* the actual no of handles */
1605 		if (req_count > 0) {
1606 			hib = hdlip =
1607 			    kmem_zalloc(req_count * sizeof (struct handle_info),
1608 			    KM_SLEEP);
1609 		} else {
1610 			hib = hdlip = 0;
1611 			req_count = hdl_info.count = 0;
1612 		}
1613 
1614 		/*
1615 		 * display existing handles
1616 		 */
1617 		mutex_enter(&bofi_low_mutex);
1618 		mutex_enter(&bofi_mutex);
1619 		for (i = 0; i < HDL_HASH_TBL_SIZE; i++) {
1620 			hhashp = &hhash_table[i];
1621 			for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext) {
1622 				if (!driver_under_test(hp->dip) ||
1623 				    ddi_name_to_major(ddi_get_name(hp->dip)) !=
1624 				    ddi_name_to_major(namep) ||
1625 				    ++(hdl_info.count) > req_count ||
1626 				    count == req_count)
1627 					continue;
1628 
1629 				hdlip->instance = hp->instance;
1630 				hdlip->rnumber = hp->rnumber;
1631 				switch (hp->type) {
1632 				case BOFI_ACC_HDL:
1633 					hdlip->access_type = BOFI_PIO_RW;
1634 					hdlip->offset = hp->offset;
1635 					hdlip->len = hp->len;
1636 					break;
1637 				case BOFI_DMA_HDL:
1638 					hdlip->access_type = 0;
1639 					if (hp->flags & DDI_DMA_WRITE)
1640 						hdlip->access_type |=
1641 						    BOFI_DMA_W;
1642 					if (hp->flags & DDI_DMA_READ)
1643 						hdlip->access_type |=
1644 						    BOFI_DMA_R;
1645 					hdlip->len = hp->len;
1646 					hdlip->addr_cookie =
1647 					    (uint64_t)(uintptr_t)hp->addr;
1648 					break;
1649 				case BOFI_INT_HDL:
1650 					hdlip->access_type = BOFI_INTR;
1651 					break;
1652 				default:
1653 					hdlip->access_type = 0;
1654 					break;
1655 				}
1656 				hdlip++;
1657 				count++;
1658 			}
1659 		}
1660 		mutex_exit(&bofi_mutex);
1661 		mutex_exit(&bofi_low_mutex);
1662 		err = 0;
1663 #ifdef _MULTI_DATAMODEL
1664 		switch (ddi_model_convert_from(mode & FMODELS)) {
1665 		case DDI_MODEL_ILP32:
1666 		{
1667 			/*
1668 			 * For use when a 32 bit app makes a call into a
1669 			 * 64 bit ioctl
1670 			 */
1671 			struct bofi_get_hdl_info32	hdl_info_32;
1672 
1673 			hdl_info_32.namesize = hdl_info.namesize;
1674 			(void) strncpy(hdl_info_32.name, hdl_info.name,
1675 			    NAMESIZE);
1676 			hdl_info_32.count = hdl_info.count;
1677 			hdl_info_32.hdli = (caddr32_t)(uintptr_t)hdl_info.hdli;
1678 			if (ddi_copyout(&hdl_info_32, (void *)arg,
1679 			    sizeof (hdl_info_32), mode) != 0) {
1680 				kmem_free(namep, hdl_info.namesize+1);
1681 				if (req_count > 0)
1682 					kmem_free(hib,
1683 					    req_count * sizeof (*hib));
1684 				return (EFAULT);
1685 			}
1686 			break;
1687 		}
1688 		case DDI_MODEL_NONE:
1689 			if (ddi_copyout(&hdl_info, (void *)arg,
1690 			    sizeof (hdl_info), mode) != 0) {
1691 				kmem_free(namep, hdl_info.namesize+1);
1692 				if (req_count > 0)
1693 					kmem_free(hib,
1694 					    req_count * sizeof (*hib));
1695 				return (EFAULT);
1696 			}
1697 			break;
1698 		}
1699 #else /* ! _MULTI_DATAMODEL */
1700 		if (ddi_copyout(&hdl_info, (void *)arg,
1701 		    sizeof (hdl_info), mode) != 0) {
1702 			kmem_free(namep, hdl_info.namesize+1);
1703 			if (req_count > 0)
1704 				kmem_free(hib, req_count * sizeof (*hib));
1705 			return (EFAULT);
1706 		}
1707 #endif /* ! _MULTI_DATAMODEL */
1708 		if (count > 0) {
1709 			if (ddi_copyout(hib, hdl_info.hdli,
1710 			    count * sizeof (*hib), mode) != 0) {
1711 				kmem_free(namep, hdl_info.namesize+1);
1712 				if (req_count > 0)
1713 					kmem_free(hib,
1714 					    req_count * sizeof (*hib));
1715 				return (EFAULT);
1716 			}
1717 		}
1718 		kmem_free(namep, hdl_info.namesize+1);
1719 		if (req_count > 0)
1720 			kmem_free(hib, req_count * sizeof (*hib));
1721 		return (err);
1722 	default:
1723 		return (ENOTTY);
1724 	}
1725 }
1726 
1727 
1728 /*
1729  * add a new error definition
1730  */
1731 static int
1732 bofi_errdef_alloc(struct bofi_errdef *errdefp, char *namep,
1733     struct bofi_errent *softc)
1734 {
1735 	struct bofi_errent *ep;
1736 	struct bofi_shadow *hp;
1737 	struct bofi_link   *lp;
1738 
1739 	/*
1740 	 * allocate errdef structure and put on in-use list
1741 	 */
1742 	ep = kmem_zalloc(sizeof (struct bofi_errent), KM_SLEEP);
1743 	ep->errdef = *errdefp;
1744 	ep->name = namep;
1745 	ep->errdef.errdef_handle = (uint64_t)(uintptr_t)ep;
1746 	ep->errstate.severity = DDI_SERVICE_RESTORED;
1747 	ep->errstate.errdef_handle = (uint64_t)(uintptr_t)ep;
1748 	cv_init(&ep->cv, NULL, CV_DRIVER, NULL);
1749 	/*
1750 	 * allocate space for logging
1751 	 */
1752 	ep->errdef.log.entries = 0;
1753 	ep->errdef.log.wrapcnt = 0;
1754 	if (ep->errdef.access_type & BOFI_LOG)
1755 		ep->logbase = kmem_alloc(sizeof (struct acc_log_elem) *
1756 		    ep->errdef.log.logsize, KM_SLEEP);
1757 	else
1758 		ep->logbase = NULL;
1759 	/*
1760 	 * put on in-use list
1761 	 */
1762 	mutex_enter(&bofi_low_mutex);
1763 	mutex_enter(&bofi_mutex);
1764 	ep->next = errent_listp;
1765 	errent_listp = ep;
1766 	/*
1767 	 * and add it to the per-clone list
1768 	 */
1769 	ep->cnext = softc->cnext;
1770 	softc->cnext->cprev = ep;
1771 	ep->cprev = softc;
1772 	softc->cnext = ep;
1773 
1774 	/*
1775 	 * look for corresponding shadow handle structures and if we find any
1776 	 * tag this errdef structure on to their link lists.
1777 	 */
1778 	for (hp = shadow_list.next; hp != &shadow_list; hp = hp->next) {
1779 		if (ddi_name_to_major(hp->name) == ddi_name_to_major(namep) &&
1780 		    hp->instance == errdefp->instance &&
1781 		    (((errdefp->access_type & BOFI_DMA_RW) &&
1782 		    (ep->errdef.rnumber == -1 ||
1783 		    hp->rnumber == ep->errdef.rnumber) &&
1784 		    hp->type == BOFI_DMA_HDL &&
1785 		    (((uintptr_t)(hp->addr + ep->errdef.offset +
1786 		    ep->errdef.len) & ~LLSZMASK) >
1787 		    ((uintptr_t)((hp->addr + ep->errdef.offset) +
1788 		    LLSZMASK) & ~LLSZMASK))) ||
1789 		    ((errdefp->access_type & BOFI_INTR) &&
1790 		    hp->type == BOFI_INT_HDL) ||
1791 		    ((errdefp->access_type & BOFI_PIO_RW) &&
1792 		    hp->type == BOFI_ACC_HDL &&
1793 		    (errdefp->rnumber == -1 ||
1794 		    hp->rnumber == errdefp->rnumber) &&
1795 		    (errdefp->len == 0 ||
1796 		    hp->offset < errdefp->offset + errdefp->len) &&
1797 		    hp->offset + hp->len > errdefp->offset))) {
1798 			lp = bofi_link_freelist;
1799 			if (lp != NULL) {
1800 				bofi_link_freelist = lp->link;
1801 				lp->errentp = ep;
1802 				lp->link = hp->link;
1803 				hp->link = lp;
1804 			}
1805 		}
1806 	}
1807 	errdefp->errdef_handle = (uint64_t)(uintptr_t)ep;
1808 	mutex_exit(&bofi_mutex);
1809 	mutex_exit(&bofi_low_mutex);
1810 	ep->softintr_id = NULL;
1811 	return (ddi_add_softintr(our_dip, DDI_SOFTINT_MED, &ep->softintr_id,
1812 	    NULL, NULL, bofi_signal, (caddr_t)&ep->errdef));
1813 }
1814 
1815 
1816 /*
1817  * delete existing errdef
1818  */
1819 static int
1820 bofi_errdef_free(struct bofi_errent *ep)
1821 {
1822 	struct bofi_errent *hep, *prev_hep;
1823 	struct bofi_link *lp, *prev_lp, *next_lp;
1824 	struct bofi_shadow *hp;
1825 
1826 	mutex_enter(&bofi_low_mutex);
1827 	mutex_enter(&bofi_mutex);
1828 	/*
1829 	 * don't just assume its a valid ep - check that its on the
1830 	 * in-use list
1831 	 */
1832 	prev_hep = NULL;
1833 	for (hep = errent_listp; hep != NULL; ) {
1834 		if (hep == ep)
1835 			break;
1836 		prev_hep = hep;
1837 		hep = hep->next;
1838 	}
1839 	if (hep == NULL) {
1840 		mutex_exit(&bofi_mutex);
1841 		mutex_exit(&bofi_low_mutex);
1842 		return (EINVAL);
1843 	}
1844 	/*
1845 	 * found it - delete from in-use list
1846 	 */
1847 
1848 	if (prev_hep)
1849 		prev_hep->next = hep->next;
1850 	else
1851 		errent_listp = hep->next;
1852 	/*
1853 	 * and take it off the per-clone list
1854 	 */
1855 	hep->cnext->cprev = hep->cprev;
1856 	hep->cprev->cnext = hep->cnext;
1857 	/*
1858 	 * see if we are on any shadow handle link lists - and if we
1859 	 * are then take us off
1860 	 */
1861 	for (hp = shadow_list.next; hp != &shadow_list; hp = hp->next) {
1862 		prev_lp = NULL;
1863 		for (lp = hp->link; lp != NULL; ) {
1864 			if (lp->errentp == ep) {
1865 				if (prev_lp)
1866 					prev_lp->link = lp->link;
1867 				else
1868 					hp->link = lp->link;
1869 				next_lp = lp->link;
1870 				lp->link = bofi_link_freelist;
1871 				bofi_link_freelist = lp;
1872 				lp = next_lp;
1873 			} else {
1874 				prev_lp = lp;
1875 				lp = lp->link;
1876 			}
1877 		}
1878 	}
1879 	mutex_exit(&bofi_mutex);
1880 	mutex_exit(&bofi_low_mutex);
1881 
1882 	cv_destroy(&ep->cv);
1883 	kmem_free(ep->name, ep->errdef.namesize+1);
1884 	if ((ep->errdef.access_type & BOFI_LOG) &&
1885 	    ep->errdef.log.logsize && ep->logbase) /* double check */
1886 		kmem_free(ep->logbase,
1887 		    sizeof (struct acc_log_elem) * ep->errdef.log.logsize);
1888 
1889 	if (ep->softintr_id)
1890 		ddi_remove_softintr(ep->softintr_id);
1891 	kmem_free(ep, sizeof (struct bofi_errent));
1892 	return (0);
1893 }
1894 
1895 
1896 /*
1897  * start all errdefs corresponding to this name and instance
1898  */
1899 static void
1900 bofi_start(struct bofi_errctl *errctlp, char *namep)
1901 {
1902 	struct bofi_errent *ep;
1903 
1904 	/*
1905 	 * look for any errdefs with matching name and instance
1906 	 */
1907 	mutex_enter(&bofi_low_mutex);
1908 	for (ep = errent_listp; ep != NULL; ep = ep->next)
1909 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
1910 		    errctlp->instance == ep->errdef.instance) {
1911 			ep->state |= BOFI_DEV_ACTIVE;
1912 			(void) drv_getparm(TIME, &(ep->errdef.log.start_time));
1913 			ep->errdef.log.stop_time = 0ul;
1914 		}
1915 	mutex_exit(&bofi_low_mutex);
1916 }
1917 
1918 
1919 /*
1920  * stop all errdefs corresponding to this name and instance
1921  */
1922 static void
1923 bofi_stop(struct bofi_errctl *errctlp, char *namep)
1924 {
1925 	struct bofi_errent *ep;
1926 
1927 	/*
1928 	 * look for any errdefs with matching name and instance
1929 	 */
1930 	mutex_enter(&bofi_low_mutex);
1931 	for (ep = errent_listp; ep != NULL; ep = ep->next)
1932 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
1933 		    errctlp->instance == ep->errdef.instance) {
1934 			ep->state &= ~BOFI_DEV_ACTIVE;
1935 			if (ep->errdef.log.stop_time == 0ul)
1936 				(void) drv_getparm(TIME,
1937 				    &(ep->errdef.log.stop_time));
1938 		}
1939 	mutex_exit(&bofi_low_mutex);
1940 }
1941 
1942 
1943 /*
1944  * wake up any thread waiting on this errdefs
1945  */
1946 static uint_t
1947 bofi_signal(caddr_t arg)
1948 {
1949 	struct bofi_errdef *edp = (struct bofi_errdef *)arg;
1950 	struct bofi_errent *hep;
1951 	struct bofi_errent *ep =
1952 	    (struct bofi_errent *)(uintptr_t)edp->errdef_handle;
1953 
1954 	mutex_enter(&bofi_low_mutex);
1955 	for (hep = errent_listp; hep != NULL; ) {
1956 		if (hep == ep)
1957 			break;
1958 		hep = hep->next;
1959 	}
1960 	if (hep == NULL) {
1961 		mutex_exit(&bofi_low_mutex);
1962 		return (DDI_INTR_UNCLAIMED);
1963 	}
1964 	if ((ep->errdef.access_type & BOFI_LOG) &&
1965 	    (edp->log.flags & BOFI_LOG_FULL)) {
1966 		edp->log.stop_time = bofi_gettime();
1967 		ep->state |= BOFI_NEW_MESSAGE;
1968 		if (ep->state & BOFI_MESSAGE_WAIT)
1969 			cv_broadcast(&ep->cv);
1970 		ep->state &= ~BOFI_MESSAGE_WAIT;
1971 	}
1972 	if (ep->errstate.msg_time != 0) {
1973 		ep->state |= BOFI_NEW_MESSAGE;
1974 		if (ep->state & BOFI_MESSAGE_WAIT)
1975 			cv_broadcast(&ep->cv);
1976 		ep->state &= ~BOFI_MESSAGE_WAIT;
1977 	}
1978 	mutex_exit(&bofi_low_mutex);
1979 	return (DDI_INTR_CLAIMED);
1980 }
1981 
1982 
1983 /*
1984  * wake up all errdefs corresponding to this name and instance
1985  */
1986 static void
1987 bofi_broadcast(struct bofi_errctl *errctlp, char *namep)
1988 {
1989 	struct bofi_errent *ep;
1990 
1991 	/*
1992 	 * look for any errdefs with matching name and instance
1993 	 */
1994 	mutex_enter(&bofi_low_mutex);
1995 	for (ep = errent_listp; ep != NULL; ep = ep->next)
1996 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
1997 		    errctlp->instance == ep->errdef.instance) {
1998 			/*
1999 			 * wake up sleepers
2000 			 */
2001 			ep->state |= BOFI_NEW_MESSAGE;
2002 			if (ep->state & BOFI_MESSAGE_WAIT)
2003 				cv_broadcast(&ep->cv);
2004 			ep->state &= ~BOFI_MESSAGE_WAIT;
2005 		}
2006 	mutex_exit(&bofi_low_mutex);
2007 }
2008 
2009 
2010 /*
2011  * clear "acc_chk" for all errdefs corresponding to this name and instance
2012  * and wake them up.
2013  */
2014 static void
2015 bofi_clear_acc_chk(struct bofi_errctl *errctlp, char *namep)
2016 {
2017 	struct bofi_errent *ep;
2018 
2019 	/*
2020 	 * look for any errdefs with matching name and instance
2021 	 */
2022 	mutex_enter(&bofi_low_mutex);
2023 	for (ep = errent_listp; ep != NULL; ep = ep->next)
2024 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
2025 		    errctlp->instance == ep->errdef.instance) {
2026 			mutex_enter(&bofi_mutex);
2027 			if (ep->errdef.access_count == 0 &&
2028 			    ep->errdef.fail_count == 0)
2029 				ep->errdef.acc_chk = 0;
2030 			mutex_exit(&bofi_mutex);
2031 			/*
2032 			 * wake up sleepers
2033 			 */
2034 			ep->state |= BOFI_NEW_MESSAGE;
2035 			if (ep->state & BOFI_MESSAGE_WAIT)
2036 				cv_broadcast(&ep->cv);
2037 			ep->state &= ~BOFI_MESSAGE_WAIT;
2038 		}
2039 	mutex_exit(&bofi_low_mutex);
2040 }
2041 
2042 
2043 /*
2044  * set "fail_count" to 0 for all errdefs corresponding to this name and instance
2045  * whose "access_count" has expired, set "acc_chk" to 0 and wake them up.
2046  */
2047 static void
2048 bofi_clear_errors(struct bofi_errctl *errctlp, char *namep)
2049 {
2050 	struct bofi_errent *ep;
2051 
2052 	/*
2053 	 * look for any errdefs with matching name and instance
2054 	 */
2055 	mutex_enter(&bofi_low_mutex);
2056 	for (ep = errent_listp; ep != NULL; ep = ep->next)
2057 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
2058 		    errctlp->instance == ep->errdef.instance) {
2059 			mutex_enter(&bofi_mutex);
2060 			if (ep->errdef.access_count == 0) {
2061 				ep->errdef.acc_chk = 0;
2062 				ep->errdef.fail_count = 0;
2063 				mutex_exit(&bofi_mutex);
2064 				if (ep->errdef.log.stop_time == 0ul)
2065 					(void) drv_getparm(TIME,
2066 					    &(ep->errdef.log.stop_time));
2067 			} else
2068 				mutex_exit(&bofi_mutex);
2069 			/*
2070 			 * wake up sleepers
2071 			 */
2072 			ep->state |= BOFI_NEW_MESSAGE;
2073 			if (ep->state & BOFI_MESSAGE_WAIT)
2074 				cv_broadcast(&ep->cv);
2075 			ep->state &= ~BOFI_MESSAGE_WAIT;
2076 		}
2077 	mutex_exit(&bofi_low_mutex);
2078 }
2079 
2080 
2081 /*
2082  * set "access_count" and "fail_count" to 0 for all errdefs corresponding to
2083  * this name and instance, set "acc_chk" to 0, and wake them up.
2084  */
2085 static void
2086 bofi_clear_errdefs(struct bofi_errctl *errctlp, char *namep)
2087 {
2088 	struct bofi_errent *ep;
2089 
2090 	/*
2091 	 * look for any errdefs with matching name and instance
2092 	 */
2093 	mutex_enter(&bofi_low_mutex);
2094 	for (ep = errent_listp; ep != NULL; ep = ep->next)
2095 		if (strncmp(namep, ep->name, NAMESIZE) == 0 &&
2096 		    errctlp->instance == ep->errdef.instance) {
2097 			mutex_enter(&bofi_mutex);
2098 			ep->errdef.acc_chk = 0;
2099 			ep->errdef.access_count = 0;
2100 			ep->errdef.fail_count = 0;
2101 			mutex_exit(&bofi_mutex);
2102 			if (ep->errdef.log.stop_time == 0ul)
2103 				(void) drv_getparm(TIME,
2104 				    &(ep->errdef.log.stop_time));
2105 			/*
2106 			 * wake up sleepers
2107 			 */
2108 			ep->state |= BOFI_NEW_MESSAGE;
2109 			if (ep->state & BOFI_MESSAGE_WAIT)
2110 				cv_broadcast(&ep->cv);
2111 			ep->state &= ~BOFI_MESSAGE_WAIT;
2112 		}
2113 	mutex_exit(&bofi_low_mutex);
2114 }
2115 
2116 
2117 /*
2118  * get state for this errdef
2119  */
2120 static int
2121 bofi_errdef_check(struct bofi_errstate *errstatep, struct acc_log_elem **logpp)
2122 {
2123 	struct bofi_errent *hep;
2124 	struct bofi_errent *ep;
2125 
2126 	ep = (struct bofi_errent *)(uintptr_t)errstatep->errdef_handle;
2127 	mutex_enter(&bofi_low_mutex);
2128 	/*
2129 	 * don't just assume its a valid ep - check that its on the
2130 	 * in-use list
2131 	 */
2132 	for (hep = errent_listp; hep != NULL; hep = hep->next)
2133 		if (hep == ep)
2134 			break;
2135 	if (hep == NULL) {
2136 		mutex_exit(&bofi_low_mutex);
2137 		return (EINVAL);
2138 	}
2139 	mutex_enter(&bofi_mutex);
2140 	ep->errstate.access_count = ep->errdef.access_count;
2141 	ep->errstate.fail_count = ep->errdef.fail_count;
2142 	ep->errstate.acc_chk = ep->errdef.acc_chk;
2143 	ep->errstate.log = ep->errdef.log;
2144 	*logpp = ep->logbase;
2145 	*errstatep = ep->errstate;
2146 	mutex_exit(&bofi_mutex);
2147 	mutex_exit(&bofi_low_mutex);
2148 	return (0);
2149 }
2150 
2151 
2152 /*
2153  * Wait for a ddi_report_fault message to come back for this errdef
2154  * Then return state for this errdef.
2155  * fault report is intercepted by bofi_post_event, which triggers
2156  * bofi_signal via a softint, which will wake up this routine if
2157  * we are waiting
2158  */
2159 static int
2160 bofi_errdef_check_w(struct bofi_errstate *errstatep,
2161     struct acc_log_elem **logpp)
2162 {
2163 	struct bofi_errent *hep;
2164 	struct bofi_errent *ep;
2165 	int rval = 0;
2166 
2167 	ep = (struct bofi_errent *)(uintptr_t)errstatep->errdef_handle;
2168 	mutex_enter(&bofi_low_mutex);
2169 retry:
2170 	/*
2171 	 * don't just assume its a valid ep - check that its on the
2172 	 * in-use list
2173 	 */
2174 	for (hep = errent_listp; hep != NULL; hep = hep->next)
2175 		if (hep == ep)
2176 			break;
2177 	if (hep == NULL) {
2178 		mutex_exit(&bofi_low_mutex);
2179 		return (EINVAL);
2180 	}
2181 	/*
2182 	 * wait for ddi_report_fault for the devinfo corresponding
2183 	 * to this errdef
2184 	 */
2185 	if (rval == 0 && !(ep->state & BOFI_NEW_MESSAGE)) {
2186 		ep->state |= BOFI_MESSAGE_WAIT;
2187 		if (cv_wait_sig(&ep->cv, &bofi_low_mutex) == 0) {
2188 			if (!(ep->state & BOFI_NEW_MESSAGE))
2189 				rval = EINTR;
2190 		}
2191 		goto retry;
2192 	}
2193 	ep->state &= ~BOFI_NEW_MESSAGE;
2194 	/*
2195 	 * we either didn't need to sleep, we've been woken up or we've been
2196 	 * signaled - either way return state now
2197 	 */
2198 	mutex_enter(&bofi_mutex);
2199 	ep->errstate.access_count = ep->errdef.access_count;
2200 	ep->errstate.fail_count = ep->errdef.fail_count;
2201 	ep->errstate.acc_chk = ep->errdef.acc_chk;
2202 	ep->errstate.log = ep->errdef.log;
2203 	*logpp = ep->logbase;
2204 	*errstatep = ep->errstate;
2205 	mutex_exit(&bofi_mutex);
2206 	mutex_exit(&bofi_low_mutex);
2207 	return (rval);
2208 }
2209 
2210 
2211 /*
2212  * support routine - check if requested driver is defined as under test in the
2213  * conf file.
2214  */
2215 static int
2216 driver_under_test(dev_info_t *rdip)
2217 {
2218 	int i;
2219 	char	*rname;
2220 	major_t rmaj;
2221 
2222 	rname = ddi_get_name(rdip);
2223 	rmaj = ddi_name_to_major(rname);
2224 
2225 	/*
2226 	 * Enforce the user to specifically request the following drivers.
2227 	 */
2228 	for (i = 0; i < driver_list_size; i += (1 + strlen(&driver_list[i]))) {
2229 		if (driver_list_neg == 0) {
2230 			if (rmaj == ddi_name_to_major(&driver_list[i]))
2231 				return (1);
2232 		} else {
2233 			if (rmaj == ddi_name_to_major(&driver_list[i+1]))
2234 				return (0);
2235 		}
2236 	}
2237 	if (driver_list_neg == 0)
2238 		return (0);
2239 	else
2240 		return (1);
2241 
2242 }
2243 
2244 
2245 static void
2246 log_acc_event(struct bofi_errent *ep, uint_t at, offset_t offset, off_t len,
2247     size_t repcount, uint64_t *valuep)
2248 {
2249 	struct bofi_errdef *edp = &(ep->errdef);
2250 	struct acc_log *log = &edp->log;
2251 
2252 	ASSERT(log != NULL);
2253 	ASSERT(MUTEX_HELD(&bofi_mutex));
2254 
2255 	if (log->flags & BOFI_LOG_REPIO)
2256 		repcount = 1;
2257 	else if (repcount == 0 && edp->access_count > 0 &&
2258 	    (log->flags & BOFI_LOG_FULL) == 0)
2259 		edp->access_count += 1;
2260 
2261 	if (repcount && log->entries < log->logsize) {
2262 		struct acc_log_elem *elem = ep->logbase + log->entries;
2263 
2264 		if (log->flags & BOFI_LOG_TIMESTAMP)
2265 			elem->access_time = bofi_gettime();
2266 		elem->access_type = at;
2267 		elem->offset = offset;
2268 		elem->value = valuep ? *valuep : 0ll;
2269 		elem->size = len;
2270 		elem->repcount = repcount;
2271 		++log->entries;
2272 		if (log->entries == log->logsize) {
2273 			log->flags |= BOFI_LOG_FULL;
2274 			ddi_trigger_softintr(((struct bofi_errent *)
2275 			    (uintptr_t)edp->errdef_handle)->softintr_id);
2276 		}
2277 	}
2278 	if ((log->flags & BOFI_LOG_WRAP) && edp->access_count <= 1) {
2279 		log->wrapcnt++;
2280 		edp->access_count = log->logsize;
2281 		log->entries = 0;	/* wrap back to the start */
2282 	}
2283 }
2284 
2285 
2286 /*
2287  * got a condition match on dma read/write - check counts and corrupt
2288  * data if necessary
2289  *
2290  * bofi_mutex always held when this is called.
2291  */
2292 static void
2293 do_dma_corrupt(struct bofi_shadow *hp, struct bofi_errent *ep,
2294     uint_t synctype, off_t off, off_t length)
2295 {
2296 	uint64_t operand;
2297 	int i;
2298 	off_t len;
2299 	caddr_t logaddr;
2300 	uint64_t *addr;
2301 	uint64_t *endaddr;
2302 	ddi_dma_impl_t *hdlp;
2303 	ndi_err_t *errp;
2304 
2305 	ASSERT(MUTEX_HELD(&bofi_mutex));
2306 	if ((ep->errdef.access_count ||
2307 	    ep->errdef.fail_count) &&
2308 	    (ep->errdef.access_type & BOFI_LOG)) {
2309 		uint_t atype;
2310 
2311 		if (synctype == DDI_DMA_SYNC_FORDEV)
2312 			atype = BOFI_DMA_W;
2313 		else if (synctype == DDI_DMA_SYNC_FORCPU ||
2314 		    synctype == DDI_DMA_SYNC_FORKERNEL)
2315 			atype = BOFI_DMA_R;
2316 		else
2317 			atype = 0;
2318 		if ((off <= ep->errdef.offset &&
2319 		    off + length > ep->errdef.offset) ||
2320 		    (off > ep->errdef.offset &&
2321 		    off < ep->errdef.offset + ep->errdef.len)) {
2322 			logaddr = (caddr_t)((uintptr_t)(hp->addr +
2323 			    off + LLSZMASK) & ~LLSZMASK);
2324 
2325 			log_acc_event(ep, atype, logaddr - hp->addr,
2326 			    length, 1, 0);
2327 		}
2328 	}
2329 	if (ep->errdef.access_count > 1) {
2330 		ep->errdef.access_count--;
2331 	} else if (ep->errdef.fail_count > 0) {
2332 		ep->errdef.fail_count--;
2333 		ep->errdef.access_count = 0;
2334 		/*
2335 		 * OK do the corruption
2336 		 */
2337 		if (ep->errstate.fail_time == 0)
2338 			ep->errstate.fail_time = bofi_gettime();
2339 		/*
2340 		 * work out how much to corrupt
2341 		 *
2342 		 * Make sure endaddr isn't greater than hp->addr + hp->len.
2343 		 * If endaddr becomes less than addr len becomes negative
2344 		 * and the following loop isn't entered.
2345 		 */
2346 		addr = (uint64_t *)((uintptr_t)((hp->addr +
2347 		    ep->errdef.offset) + LLSZMASK) & ~LLSZMASK);
2348 		endaddr = (uint64_t *)((uintptr_t)(hp->addr + min(hp->len,
2349 		    ep->errdef.offset + ep->errdef.len)) & ~LLSZMASK);
2350 		len = endaddr - addr;
2351 		operand = ep->errdef.operand;
2352 		hdlp = (ddi_dma_impl_t *)(hp->hdl.dma_handle);
2353 		errp = &hdlp->dmai_error;
2354 		if (ep->errdef.acc_chk & 2) {
2355 			uint64_t ena;
2356 			char buf[FM_MAX_CLASS];
2357 
2358 			errp->err_status = DDI_FM_NONFATAL;
2359 			(void) snprintf(buf, FM_MAX_CLASS, FM_SIMULATED_DMA);
2360 			ena = fm_ena_generate(0, FM_ENA_FMT1);
2361 			ddi_fm_ereport_post(hp->dip, buf, ena,
2362 			    DDI_NOSLEEP, FM_VERSION, DATA_TYPE_UINT8,
2363 			    FM_EREPORT_VERS0, NULL);
2364 		}
2365 		switch (ep->errdef.optype) {
2366 		case BOFI_EQUAL :
2367 			for (i = 0; i < len; i++)
2368 				*(addr + i) = operand;
2369 			break;
2370 		case BOFI_AND :
2371 			for (i = 0; i < len; i++)
2372 				*(addr + i) &= operand;
2373 			break;
2374 		case BOFI_OR :
2375 			for (i = 0; i < len; i++)
2376 				*(addr + i) |= operand;
2377 			break;
2378 		case BOFI_XOR :
2379 			for (i = 0; i < len; i++)
2380 				*(addr + i) ^= operand;
2381 			break;
2382 		default:
2383 			/* do nothing */
2384 			break;
2385 		}
2386 	}
2387 }
2388 
2389 
2390 static uint64_t do_bofi_rd8(struct bofi_shadow *, caddr_t);
2391 static uint64_t do_bofi_rd16(struct bofi_shadow *, caddr_t);
2392 static uint64_t do_bofi_rd32(struct bofi_shadow *, caddr_t);
2393 static uint64_t do_bofi_rd64(struct bofi_shadow *, caddr_t);
2394 
2395 
2396 /*
2397  * check all errdefs linked to this shadow handle. If we've got a condition
2398  * match check counts and corrupt data if necessary
2399  *
2400  * bofi_mutex always held when this is called.
2401  *
2402  * because of possibility of BOFI_NO_TRANSFER, we couldn't get data
2403  * from io-space before calling this, so we pass in the func to do the
2404  * transfer as a parameter.
2405  */
2406 static uint64_t
2407 do_pior_corrupt(struct bofi_shadow *hp, caddr_t addr,
2408     uint64_t (*func)(), size_t repcount, size_t accsize)
2409 {
2410 	struct bofi_errent *ep;
2411 	struct bofi_link   *lp;
2412 	uint64_t operand;
2413 	uintptr_t minlen;
2414 	intptr_t base;
2415 	int done_get = 0;
2416 	uint64_t get_val, gv;
2417 	ddi_acc_impl_t *hdlp;
2418 	ndi_err_t *errp;
2419 
2420 	ASSERT(MUTEX_HELD(&bofi_mutex));
2421 	/*
2422 	 * check through all errdefs associated with this shadow handle
2423 	 */
2424 	for (lp = hp->link; lp != NULL; lp = lp->link) {
2425 		ep = lp->errentp;
2426 		if (ep->errdef.len == 0)
2427 			minlen = hp->len;
2428 		else
2429 			minlen = min(hp->len, ep->errdef.len);
2430 		base = addr - hp->addr - ep->errdef.offset + hp->offset;
2431 		if ((ep->errdef.access_type & BOFI_PIO_R) &&
2432 		    (ep->state & BOFI_DEV_ACTIVE) &&
2433 		    base >= 0 && base < minlen) {
2434 			/*
2435 			 * condition match for pio read
2436 			 */
2437 			if (ep->errdef.access_count > 1) {
2438 				ep->errdef.access_count--;
2439 				if (done_get == 0) {
2440 					done_get = 1;
2441 					gv = get_val = func(hp, addr);
2442 				}
2443 				if (ep->errdef.access_type & BOFI_LOG) {
2444 					log_acc_event(ep, BOFI_PIO_R,
2445 					    addr - hp->addr,
2446 					    accsize, repcount, &gv);
2447 				}
2448 			} else if (ep->errdef.fail_count > 0) {
2449 				ep->errdef.fail_count--;
2450 				ep->errdef.access_count = 0;
2451 				/*
2452 				 * OK do corruption
2453 				 */
2454 				if (ep->errstate.fail_time == 0)
2455 					ep->errstate.fail_time = bofi_gettime();
2456 				operand = ep->errdef.operand;
2457 				if (done_get == 0) {
2458 					if (ep->errdef.optype ==
2459 					    BOFI_NO_TRANSFER)
2460 						/*
2461 						 * no transfer - bomb out
2462 						 */
2463 						return (operand);
2464 					done_get = 1;
2465 					gv = get_val = func(hp, addr);
2466 
2467 				}
2468 				if (ep->errdef.access_type & BOFI_LOG) {
2469 					log_acc_event(ep, BOFI_PIO_R,
2470 					    addr - hp->addr,
2471 					    accsize, repcount, &gv);
2472 				}
2473 				hdlp = (ddi_acc_impl_t *)(hp->hdl.acc_handle);
2474 				errp = hdlp->ahi_err;
2475 				if (ep->errdef.acc_chk & 1) {
2476 					uint64_t ena;
2477 					char buf[FM_MAX_CLASS];
2478 
2479 					errp->err_status = DDI_FM_NONFATAL;
2480 					(void) snprintf(buf, FM_MAX_CLASS,
2481 					    FM_SIMULATED_PIO);
2482 					ena = fm_ena_generate(0, FM_ENA_FMT1);
2483 					ddi_fm_ereport_post(hp->dip, buf, ena,
2484 					    DDI_NOSLEEP, FM_VERSION,
2485 					    DATA_TYPE_UINT8, FM_EREPORT_VERS0,
2486 					    NULL);
2487 				}
2488 				switch (ep->errdef.optype) {
2489 				case BOFI_EQUAL :
2490 					get_val = operand;
2491 					break;
2492 				case BOFI_AND :
2493 					get_val &= operand;
2494 					break;
2495 				case BOFI_OR :
2496 					get_val |= operand;
2497 					break;
2498 				case BOFI_XOR :
2499 					get_val ^= operand;
2500 					break;
2501 				default:
2502 					/* do nothing */
2503 					break;
2504 				}
2505 			}
2506 		}
2507 	}
2508 	if (done_get == 0)
2509 		return (func(hp, addr));
2510 	else
2511 		return (get_val);
2512 }
2513 
2514 
2515 /*
2516  * check all errdefs linked to this shadow handle. If we've got a condition
2517  * match check counts and corrupt data if necessary
2518  *
2519  * bofi_mutex always held when this is called.
2520  *
2521  * because of possibility of BOFI_NO_TRANSFER, we return 0 if no data
2522  * is to be written out to io-space, 1 otherwise
2523  */
2524 static int
2525 do_piow_corrupt(struct bofi_shadow *hp, caddr_t addr, uint64_t *valuep,
2526     size_t size, size_t repcount)
2527 {
2528 	struct bofi_errent *ep;
2529 	struct bofi_link   *lp;
2530 	uintptr_t minlen;
2531 	intptr_t base;
2532 	uint64_t v = *valuep;
2533 	ddi_acc_impl_t *hdlp;
2534 	ndi_err_t *errp;
2535 
2536 	ASSERT(MUTEX_HELD(&bofi_mutex));
2537 	/*
2538 	 * check through all errdefs associated with this shadow handle
2539 	 */
2540 	for (lp = hp->link; lp != NULL; lp = lp->link) {
2541 		ep = lp->errentp;
2542 		if (ep->errdef.len == 0)
2543 			minlen = hp->len;
2544 		else
2545 			minlen = min(hp->len, ep->errdef.len);
2546 		base = (caddr_t)addr - hp->addr - ep->errdef.offset +hp->offset;
2547 		if ((ep->errdef.access_type & BOFI_PIO_W) &&
2548 		    (ep->state & BOFI_DEV_ACTIVE) &&
2549 		    base >= 0 && base < minlen) {
2550 			/*
2551 			 * condition match for pio write
2552 			 */
2553 
2554 			if (ep->errdef.access_count > 1) {
2555 				ep->errdef.access_count--;
2556 				if (ep->errdef.access_type & BOFI_LOG)
2557 					log_acc_event(ep, BOFI_PIO_W,
2558 					    addr - hp->addr, size,
2559 					    repcount, &v);
2560 			} else if (ep->errdef.fail_count > 0) {
2561 				ep->errdef.fail_count--;
2562 				ep->errdef.access_count = 0;
2563 				if (ep->errdef.access_type & BOFI_LOG)
2564 					log_acc_event(ep, BOFI_PIO_W,
2565 					    addr - hp->addr, size,
2566 					    repcount, &v);
2567 				/*
2568 				 * OK do corruption
2569 				 */
2570 				if (ep->errstate.fail_time == 0)
2571 					ep->errstate.fail_time = bofi_gettime();
2572 				hdlp = (ddi_acc_impl_t *)(hp->hdl.acc_handle);
2573 				errp = hdlp->ahi_err;
2574 				if (ep->errdef.acc_chk & 1) {
2575 					uint64_t ena;
2576 					char buf[FM_MAX_CLASS];
2577 
2578 					errp->err_status = DDI_FM_NONFATAL;
2579 					(void) snprintf(buf, FM_MAX_CLASS,
2580 					    FM_SIMULATED_PIO);
2581 					ena = fm_ena_generate(0, FM_ENA_FMT1);
2582 					ddi_fm_ereport_post(hp->dip, buf, ena,
2583 					    DDI_NOSLEEP, FM_VERSION,
2584 					    DATA_TYPE_UINT8, FM_EREPORT_VERS0,
2585 					    NULL);
2586 				}
2587 				switch (ep->errdef.optype) {
2588 				case BOFI_EQUAL :
2589 					*valuep = ep->errdef.operand;
2590 					break;
2591 				case BOFI_AND :
2592 					*valuep &= ep->errdef.operand;
2593 					break;
2594 				case BOFI_OR :
2595 					*valuep |= ep->errdef.operand;
2596 					break;
2597 				case BOFI_XOR :
2598 					*valuep ^= ep->errdef.operand;
2599 					break;
2600 				case BOFI_NO_TRANSFER :
2601 					/*
2602 					 * no transfer - bomb out
2603 					 */
2604 					return (0);
2605 				default:
2606 					/* do nothing */
2607 					break;
2608 				}
2609 			}
2610 		}
2611 	}
2612 	return (1);
2613 }
2614 
2615 
2616 static uint64_t
2617 do_bofi_rd8(struct bofi_shadow *hp, caddr_t addr)
2618 {
2619 	return (hp->save.acc.ahi_get8(&hp->save.acc, (uint8_t *)addr));
2620 }
2621 
2622 #define	BOFI_READ_CHECKS(type) \
2623 	if (bofi_ddi_check) \
2624 		addr = (type *)((uintptr_t)addr - 64 + hp->addr); \
2625 	if (bofi_range_check && ((caddr_t)addr < hp->addr || \
2626 	    (caddr_t)addr - hp->addr >= hp->len)) { \
2627 		cmn_err((bofi_range_check == 2) ? CE_PANIC : CE_WARN, \
2628 		    "ddi_get() out of range addr %p not in %p/%llx", \
2629 		    (void *)addr, (void *)hp->addr, hp->len); \
2630 		return (0); \
2631 	}
2632 
2633 /*
2634  * our getb() routine - use tryenter
2635  */
2636 static uint8_t
2637 bofi_rd8(ddi_acc_impl_t *handle, uint8_t *addr)
2638 {
2639 	struct bofi_shadow *hp;
2640 	uint8_t retval;
2641 
2642 	hp = handle->ahi_common.ah_bus_private;
2643 	BOFI_READ_CHECKS(uint8_t)
2644 	if (!hp->link || !mutex_tryenter(&bofi_mutex))
2645 		return (hp->save.acc.ahi_get8(&hp->save.acc, addr));
2646 	retval = (uint8_t)do_pior_corrupt(hp, (caddr_t)addr, do_bofi_rd8, 1,
2647 	    1);
2648 	mutex_exit(&bofi_mutex);
2649 	return (retval);
2650 }
2651 
2652 
2653 static uint64_t
2654 do_bofi_rd16(struct bofi_shadow *hp, caddr_t addr)
2655 {
2656 	return (hp->save.acc.ahi_get16(&hp->save.acc, (uint16_t *)addr));
2657 }
2658 
2659 
2660 /*
2661  * our getw() routine - use tryenter
2662  */
2663 static uint16_t
2664 bofi_rd16(ddi_acc_impl_t *handle, uint16_t *addr)
2665 {
2666 	struct bofi_shadow *hp;
2667 	uint16_t retval;
2668 
2669 	hp = handle->ahi_common.ah_bus_private;
2670 	BOFI_READ_CHECKS(uint16_t)
2671 	if (!hp->link || !mutex_tryenter(&bofi_mutex))
2672 		return (hp->save.acc.ahi_get16(&hp->save.acc, addr));
2673 	retval = (uint16_t)do_pior_corrupt(hp, (caddr_t)addr, do_bofi_rd16, 1,
2674 	    2);
2675 	mutex_exit(&bofi_mutex);
2676 	return (retval);
2677 }
2678 
2679 
2680 static uint64_t
2681 do_bofi_rd32(struct bofi_shadow *hp, caddr_t addr)
2682 {
2683 	return (hp->save.acc.ahi_get32(&hp->save.acc, (uint32_t *)addr));
2684 }
2685 
2686 
2687 /*
2688  * our getl() routine - use tryenter
2689  */
2690 static uint32_t
2691 bofi_rd32(ddi_acc_impl_t *handle, uint32_t *addr)
2692 {
2693 	struct bofi_shadow *hp;
2694 	uint32_t retval;
2695 
2696 	hp = handle->ahi_common.ah_bus_private;
2697 	BOFI_READ_CHECKS(uint32_t)
2698 	if (!hp->link || !mutex_tryenter(&bofi_mutex))
2699 		return (hp->save.acc.ahi_get32(&hp->save.acc, addr));
2700 	retval = (uint32_t)do_pior_corrupt(hp, (caddr_t)addr, do_bofi_rd32, 1,
2701 	    4);
2702 	mutex_exit(&bofi_mutex);
2703 	return (retval);
2704 }
2705 
2706 
2707 static uint64_t
2708 do_bofi_rd64(struct bofi_shadow *hp, caddr_t addr)
2709 {
2710 	return (hp->save.acc.ahi_get64(&hp->save.acc, (uint64_t *)addr));
2711 }
2712 
2713 
2714 /*
2715  * our getll() routine - use tryenter
2716  */
2717 static uint64_t
2718 bofi_rd64(ddi_acc_impl_t *handle, uint64_t *addr)
2719 {
2720 	struct bofi_shadow *hp;
2721 	uint64_t retval;
2722 
2723 	hp = handle->ahi_common.ah_bus_private;
2724 	BOFI_READ_CHECKS(uint64_t)
2725 	if (!hp->link || !mutex_tryenter(&bofi_mutex))
2726 		return (hp->save.acc.ahi_get64(&hp->save.acc, addr));
2727 	retval = (uint64_t)do_pior_corrupt(hp, (caddr_t)addr, do_bofi_rd64, 1,
2728 	    8);
2729 	mutex_exit(&bofi_mutex);
2730 	return (retval);
2731 }
2732 
2733 #define	BOFI_WRITE_TESTS(type) \
2734 	if (bofi_ddi_check) \
2735 		addr = (type *)((uintptr_t)addr - 64 + hp->addr); \
2736 	if (bofi_range_check && ((caddr_t)addr < hp->addr || \
2737 	    (caddr_t)addr - hp->addr >= hp->len)) { \
2738 		cmn_err((bofi_range_check == 2) ? CE_PANIC : CE_WARN, \
2739 		    "ddi_put() out of range addr %p not in %p/%llx\n", \
2740 		    (void *)addr, (void *)hp->addr, hp->len); \
2741 		return; \
2742 	}
2743 
2744 /*
2745  * our putb() routine - use tryenter
2746  */
2747 static void
2748 bofi_wr8(ddi_acc_impl_t *handle, uint8_t *addr, uint8_t value)
2749 {
2750 	struct bofi_shadow *hp;
2751 	uint64_t llvalue = value;
2752 
2753 	hp = handle->ahi_common.ah_bus_private;
2754 	BOFI_WRITE_TESTS(uint8_t)
2755 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2756 		hp->save.acc.ahi_put8(&hp->save.acc, addr, (uint8_t)llvalue);
2757 		return;
2758 	}
2759 	if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 1, 1))
2760 		hp->save.acc.ahi_put8(&hp->save.acc, addr, (uint8_t)llvalue);
2761 	mutex_exit(&bofi_mutex);
2762 }
2763 
2764 
2765 /*
2766  * our putw() routine - use tryenter
2767  */
2768 static void
2769 bofi_wr16(ddi_acc_impl_t *handle, uint16_t *addr, uint16_t value)
2770 {
2771 	struct bofi_shadow *hp;
2772 	uint64_t llvalue = value;
2773 
2774 	hp = handle->ahi_common.ah_bus_private;
2775 	BOFI_WRITE_TESTS(uint16_t)
2776 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2777 		hp->save.acc.ahi_put16(&hp->save.acc, addr, (uint16_t)llvalue);
2778 		return;
2779 	}
2780 	if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 2, 1))
2781 		hp->save.acc.ahi_put16(&hp->save.acc, addr, (uint16_t)llvalue);
2782 	mutex_exit(&bofi_mutex);
2783 }
2784 
2785 
2786 /*
2787  * our putl() routine - use tryenter
2788  */
2789 static void
2790 bofi_wr32(ddi_acc_impl_t *handle, uint32_t *addr, uint32_t value)
2791 {
2792 	struct bofi_shadow *hp;
2793 	uint64_t llvalue = value;
2794 
2795 	hp = handle->ahi_common.ah_bus_private;
2796 	BOFI_WRITE_TESTS(uint32_t)
2797 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2798 		hp->save.acc.ahi_put32(&hp->save.acc, addr, (uint32_t)llvalue);
2799 		return;
2800 	}
2801 	if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 4, 1))
2802 		hp->save.acc.ahi_put32(&hp->save.acc, addr, (uint32_t)llvalue);
2803 	mutex_exit(&bofi_mutex);
2804 }
2805 
2806 
2807 /*
2808  * our putll() routine - use tryenter
2809  */
2810 static void
2811 bofi_wr64(ddi_acc_impl_t *handle, uint64_t *addr, uint64_t value)
2812 {
2813 	struct bofi_shadow *hp;
2814 	uint64_t llvalue = value;
2815 
2816 	hp = handle->ahi_common.ah_bus_private;
2817 	BOFI_WRITE_TESTS(uint64_t)
2818 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2819 		hp->save.acc.ahi_put64(&hp->save.acc, addr, (uint64_t)llvalue);
2820 		return;
2821 	}
2822 	if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 8, 1))
2823 		hp->save.acc.ahi_put64(&hp->save.acc, addr, (uint64_t)llvalue);
2824 	mutex_exit(&bofi_mutex);
2825 }
2826 
2827 #define	BOFI_REP_READ_TESTS(type) \
2828 	if (bofi_ddi_check) \
2829 		dev_addr = (type *)((uintptr_t)dev_addr - 64 + hp->addr); \
2830 	if (bofi_range_check && ((caddr_t)dev_addr < hp->addr || \
2831 	    (caddr_t)(dev_addr + repcount) - hp->addr > hp->len)) { \
2832 		cmn_err((bofi_range_check == 2) ? CE_PANIC : CE_WARN, \
2833 		    "ddi_rep_get() out of range addr %p not in %p/%llx\n", \
2834 		    (void *)dev_addr, (void *)hp->addr, hp->len); \
2835 		if ((caddr_t)dev_addr < hp->addr || \
2836 		    (caddr_t)dev_addr - hp->addr >= hp->len) \
2837 			return; \
2838 		repcount = (type *)(hp->addr + hp->len) - dev_addr; \
2839 	}
2840 
2841 /*
2842  * our rep_getb() routine - use tryenter
2843  */
2844 static void
2845 bofi_rep_rd8(ddi_acc_impl_t *handle, uint8_t *host_addr, uint8_t *dev_addr,
2846     size_t repcount, uint_t flags)
2847 {
2848 	struct bofi_shadow *hp;
2849 	int i;
2850 	uint8_t *addr;
2851 
2852 	hp = handle->ahi_common.ah_bus_private;
2853 	BOFI_REP_READ_TESTS(uint8_t)
2854 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2855 		hp->save.acc.ahi_rep_get8(&hp->save.acc, host_addr, dev_addr,
2856 		    repcount, flags);
2857 		return;
2858 	}
2859 	for (i = 0; i < repcount; i++) {
2860 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
2861 		*(host_addr + i) = (uint8_t)do_pior_corrupt(hp, (caddr_t)addr,
2862 		    do_bofi_rd8, i ? 0 : repcount, 1);
2863 	}
2864 	mutex_exit(&bofi_mutex);
2865 }
2866 
2867 
2868 /*
2869  * our rep_getw() routine - use tryenter
2870  */
2871 static void
2872 bofi_rep_rd16(ddi_acc_impl_t *handle, uint16_t *host_addr,
2873     uint16_t *dev_addr, size_t repcount, uint_t flags)
2874 {
2875 	struct bofi_shadow *hp;
2876 	int i;
2877 	uint16_t *addr;
2878 
2879 	hp = handle->ahi_common.ah_bus_private;
2880 	BOFI_REP_READ_TESTS(uint16_t)
2881 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2882 		hp->save.acc.ahi_rep_get16(&hp->save.acc, host_addr, dev_addr,
2883 		    repcount, flags);
2884 		return;
2885 	}
2886 	for (i = 0; i < repcount; i++) {
2887 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
2888 		*(host_addr + i) = (uint16_t)do_pior_corrupt(hp, (caddr_t)addr,
2889 		    do_bofi_rd16, i ? 0 : repcount, 2);
2890 	}
2891 	mutex_exit(&bofi_mutex);
2892 }
2893 
2894 
2895 /*
2896  * our rep_getl() routine - use tryenter
2897  */
2898 static void
2899 bofi_rep_rd32(ddi_acc_impl_t *handle, uint32_t *host_addr,
2900     uint32_t *dev_addr, size_t repcount, uint_t flags)
2901 {
2902 	struct bofi_shadow *hp;
2903 	int i;
2904 	uint32_t *addr;
2905 
2906 	hp = handle->ahi_common.ah_bus_private;
2907 	BOFI_REP_READ_TESTS(uint32_t)
2908 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2909 		hp->save.acc.ahi_rep_get32(&hp->save.acc, host_addr, dev_addr,
2910 		    repcount, flags);
2911 		return;
2912 	}
2913 	for (i = 0; i < repcount; i++) {
2914 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
2915 		*(host_addr + i) = (uint32_t)do_pior_corrupt(hp, (caddr_t)addr,
2916 		    do_bofi_rd32, i ? 0 : repcount, 4);
2917 	}
2918 	mutex_exit(&bofi_mutex);
2919 }
2920 
2921 
2922 /*
2923  * our rep_getll() routine - use tryenter
2924  */
2925 static void
2926 bofi_rep_rd64(ddi_acc_impl_t *handle, uint64_t *host_addr,
2927     uint64_t *dev_addr, size_t repcount, uint_t flags)
2928 {
2929 	struct bofi_shadow *hp;
2930 	int i;
2931 	uint64_t *addr;
2932 
2933 	hp = handle->ahi_common.ah_bus_private;
2934 	BOFI_REP_READ_TESTS(uint64_t)
2935 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2936 		hp->save.acc.ahi_rep_get64(&hp->save.acc, host_addr, dev_addr,
2937 		    repcount, flags);
2938 		return;
2939 	}
2940 	for (i = 0; i < repcount; i++) {
2941 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
2942 		*(host_addr + i) = (uint64_t)do_pior_corrupt(hp, (caddr_t)addr,
2943 		    do_bofi_rd64, i ? 0 : repcount, 8);
2944 	}
2945 	mutex_exit(&bofi_mutex);
2946 }
2947 
2948 #define	BOFI_REP_WRITE_TESTS(type) \
2949 	if (bofi_ddi_check) \
2950 		dev_addr = (type *)((uintptr_t)dev_addr - 64 + hp->addr); \
2951 	if (bofi_range_check && ((caddr_t)dev_addr < hp->addr || \
2952 	    (caddr_t)(dev_addr + repcount) - hp->addr > hp->len)) { \
2953 		cmn_err((bofi_range_check == 2) ? CE_PANIC : CE_WARN, \
2954 		    "ddi_rep_put() out of range addr %p not in %p/%llx\n", \
2955 		    (void *)dev_addr, (void *)hp->addr, hp->len); \
2956 		if ((caddr_t)dev_addr < hp->addr || \
2957 		    (caddr_t)dev_addr - hp->addr >= hp->len) \
2958 			return; \
2959 		repcount = (type *)(hp->addr + hp->len) - dev_addr; \
2960 	}
2961 
2962 /*
2963  * our rep_putb() routine - use tryenter
2964  */
2965 static void
2966 bofi_rep_wr8(ddi_acc_impl_t *handle, uint8_t *host_addr, uint8_t *dev_addr,
2967     size_t repcount, uint_t flags)
2968 {
2969 	struct bofi_shadow *hp;
2970 	int i;
2971 	uint64_t llvalue;
2972 	uint8_t *addr;
2973 
2974 	hp = handle->ahi_common.ah_bus_private;
2975 	BOFI_REP_WRITE_TESTS(uint8_t)
2976 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
2977 		hp->save.acc.ahi_rep_put8(&hp->save.acc, host_addr, dev_addr,
2978 		    repcount, flags);
2979 		return;
2980 	}
2981 	for (i = 0; i < repcount; i++) {
2982 		llvalue = *(host_addr + i);
2983 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
2984 		if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 1, i ? 0 :
2985 		    repcount))
2986 			hp->save.acc.ahi_put8(&hp->save.acc, addr,
2987 			    (uint8_t)llvalue);
2988 	}
2989 	mutex_exit(&bofi_mutex);
2990 }
2991 
2992 
2993 /*
2994  * our rep_putw() routine - use tryenter
2995  */
2996 static void
2997 bofi_rep_wr16(ddi_acc_impl_t *handle, uint16_t *host_addr,
2998     uint16_t *dev_addr, size_t repcount, uint_t flags)
2999 {
3000 	struct bofi_shadow *hp;
3001 	int i;
3002 	uint64_t llvalue;
3003 	uint16_t *addr;
3004 
3005 	hp = handle->ahi_common.ah_bus_private;
3006 	BOFI_REP_WRITE_TESTS(uint16_t)
3007 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
3008 		hp->save.acc.ahi_rep_put16(&hp->save.acc, host_addr, dev_addr,
3009 		    repcount, flags);
3010 		return;
3011 	}
3012 	for (i = 0; i < repcount; i++) {
3013 		llvalue = *(host_addr + i);
3014 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
3015 		if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 2, i ? 0 :
3016 		    repcount))
3017 			hp->save.acc.ahi_put16(&hp->save.acc, addr,
3018 			    (uint16_t)llvalue);
3019 	}
3020 	mutex_exit(&bofi_mutex);
3021 }
3022 
3023 
3024 /*
3025  * our rep_putl() routine - use tryenter
3026  */
3027 static void
3028 bofi_rep_wr32(ddi_acc_impl_t *handle, uint32_t *host_addr,
3029     uint32_t *dev_addr, size_t repcount, uint_t flags)
3030 {
3031 	struct bofi_shadow *hp;
3032 	int i;
3033 	uint64_t llvalue;
3034 	uint32_t *addr;
3035 
3036 	hp = handle->ahi_common.ah_bus_private;
3037 	BOFI_REP_WRITE_TESTS(uint32_t)
3038 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
3039 		hp->save.acc.ahi_rep_put32(&hp->save.acc, host_addr, dev_addr,
3040 		    repcount, flags);
3041 		return;
3042 	}
3043 	for (i = 0; i < repcount; i++) {
3044 		llvalue = *(host_addr + i);
3045 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
3046 		if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 4, i ? 0 :
3047 		    repcount))
3048 			hp->save.acc.ahi_put32(&hp->save.acc, addr,
3049 			    (uint32_t)llvalue);
3050 	}
3051 	mutex_exit(&bofi_mutex);
3052 }
3053 
3054 
3055 /*
3056  * our rep_putll() routine - use tryenter
3057  */
3058 static void
3059 bofi_rep_wr64(ddi_acc_impl_t *handle, uint64_t *host_addr,
3060     uint64_t *dev_addr, size_t repcount, uint_t flags)
3061 {
3062 	struct bofi_shadow *hp;
3063 	int i;
3064 	uint64_t llvalue;
3065 	uint64_t *addr;
3066 
3067 	hp = handle->ahi_common.ah_bus_private;
3068 	BOFI_REP_WRITE_TESTS(uint64_t)
3069 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
3070 		hp->save.acc.ahi_rep_put64(&hp->save.acc, host_addr, dev_addr,
3071 		    repcount, flags);
3072 		return;
3073 	}
3074 	for (i = 0; i < repcount; i++) {
3075 		llvalue = *(host_addr + i);
3076 		addr = dev_addr + ((flags == DDI_DEV_AUTOINCR) ? i : 0);
3077 		if (do_piow_corrupt(hp, (caddr_t)addr, &llvalue, 8, i ? 0 :
3078 		    repcount))
3079 			hp->save.acc.ahi_put64(&hp->save.acc, addr,
3080 			    (uint64_t)llvalue);
3081 	}
3082 	mutex_exit(&bofi_mutex);
3083 }
3084 
3085 
3086 /*
3087  * our ddi_map routine
3088  */
3089 static int
3090 bofi_map(dev_info_t *dip, dev_info_t *rdip,
3091     ddi_map_req_t *reqp, off_t offset, off_t len, caddr_t *vaddrp)
3092 {
3093 	ddi_acc_impl_t *ap;
3094 	struct bofi_shadow *hp;
3095 	struct bofi_errent *ep;
3096 	struct bofi_link   *lp, *next_lp;
3097 	int retval;
3098 	struct bofi_shadow *dhashp;
3099 	struct bofi_shadow *hhashp;
3100 
3101 	switch (reqp->map_op) {
3102 	case DDI_MO_MAP_LOCKED:
3103 		/*
3104 		 * for this case get nexus to do real work first
3105 		 */
3106 		retval = save_bus_ops.bus_map(dip, rdip, reqp, offset, len,
3107 		    vaddrp);
3108 		if (retval != DDI_SUCCESS)
3109 			return (retval);
3110 
3111 		ap = (ddi_acc_impl_t *)reqp->map_handlep;
3112 		if (ap == NULL)
3113 			return (DDI_SUCCESS);
3114 		/*
3115 		 * if driver_list is set, only intercept those drivers
3116 		 */
3117 		if (!driver_under_test(ap->ahi_common.ah_dip))
3118 			return (DDI_SUCCESS);
3119 
3120 		/*
3121 		 * support for ddi_regs_map_setup()
3122 		 * - allocate shadow handle structure and fill it in
3123 		 */
3124 		hp = kmem_zalloc(sizeof (struct bofi_shadow), KM_SLEEP);
3125 		(void) strncpy(hp->name, ddi_get_name(ap->ahi_common.ah_dip),
3126 		    NAMESIZE);
3127 		hp->instance = ddi_get_instance(ap->ahi_common.ah_dip);
3128 		hp->dip = ap->ahi_common.ah_dip;
3129 		hp->addr = *vaddrp;
3130 		/*
3131 		 * return spurious value to catch direct access to registers
3132 		 */
3133 		if (bofi_ddi_check)
3134 			*vaddrp = (caddr_t)64;
3135 		hp->rnumber = ((ddi_acc_hdl_t *)ap)->ah_rnumber;
3136 		hp->offset = offset;
3137 		if (len == 0)
3138 			hp->len = INT_MAX - offset;
3139 		else
3140 			hp->len = min(len, INT_MAX - offset);
3141 		hp->hdl.acc_handle = (ddi_acc_handle_t)ap;
3142 		hp->link = NULL;
3143 		hp->type = BOFI_ACC_HDL;
3144 		/*
3145 		 * save existing function pointers and plug in our own
3146 		 */
3147 		hp->save.acc = *ap;
3148 		ap->ahi_get8 = bofi_rd8;
3149 		ap->ahi_get16 = bofi_rd16;
3150 		ap->ahi_get32 = bofi_rd32;
3151 		ap->ahi_get64 = bofi_rd64;
3152 		ap->ahi_put8 = bofi_wr8;
3153 		ap->ahi_put16 = bofi_wr16;
3154 		ap->ahi_put32 = bofi_wr32;
3155 		ap->ahi_put64 = bofi_wr64;
3156 		ap->ahi_rep_get8 = bofi_rep_rd8;
3157 		ap->ahi_rep_get16 = bofi_rep_rd16;
3158 		ap->ahi_rep_get32 = bofi_rep_rd32;
3159 		ap->ahi_rep_get64 = bofi_rep_rd64;
3160 		ap->ahi_rep_put8 = bofi_rep_wr8;
3161 		ap->ahi_rep_put16 = bofi_rep_wr16;
3162 		ap->ahi_rep_put32 = bofi_rep_wr32;
3163 		ap->ahi_rep_put64 = bofi_rep_wr64;
3164 		ap->ahi_fault_check = bofi_check_acc_hdl;
3165 #if defined(__sparc)
3166 #else
3167 		ap->ahi_acc_attr &= ~DDI_ACCATTR_DIRECT;
3168 #endif
3169 		/*
3170 		 * stick in a pointer to our shadow handle
3171 		 */
3172 		ap->ahi_common.ah_bus_private = hp;
3173 		/*
3174 		 * add to dhash, hhash and inuse lists
3175 		 */
3176 		mutex_enter(&bofi_low_mutex);
3177 		mutex_enter(&bofi_mutex);
3178 		hp->next = shadow_list.next;
3179 		shadow_list.next->prev = hp;
3180 		hp->prev = &shadow_list;
3181 		shadow_list.next = hp;
3182 		hhashp = HDL_HHASH(ap);
3183 		hp->hnext = hhashp->hnext;
3184 		hhashp->hnext->hprev = hp;
3185 		hp->hprev = hhashp;
3186 		hhashp->hnext = hp;
3187 		dhashp = HDL_DHASH(hp->dip);
3188 		hp->dnext = dhashp->dnext;
3189 		dhashp->dnext->dprev = hp;
3190 		hp->dprev = dhashp;
3191 		dhashp->dnext = hp;
3192 		/*
3193 		 * chain on any pre-existing errdefs that apply to this
3194 		 * acc_handle
3195 		 */
3196 		for (ep = errent_listp; ep != NULL; ep = ep->next) {
3197 			if (ddi_name_to_major(hp->name) ==
3198 			    ddi_name_to_major(ep->name) &&
3199 			    hp->instance == ep->errdef.instance &&
3200 			    (ep->errdef.access_type & BOFI_PIO_RW) &&
3201 			    (ep->errdef.rnumber == -1 ||
3202 			    hp->rnumber == ep->errdef.rnumber) &&
3203 			    (ep->errdef.len == 0 ||
3204 			    offset < ep->errdef.offset + ep->errdef.len) &&
3205 			    offset + hp->len > ep->errdef.offset) {
3206 				lp = bofi_link_freelist;
3207 				if (lp != NULL) {
3208 					bofi_link_freelist = lp->link;
3209 					lp->errentp = ep;
3210 					lp->link = hp->link;
3211 					hp->link = lp;
3212 				}
3213 			}
3214 		}
3215 		mutex_exit(&bofi_mutex);
3216 		mutex_exit(&bofi_low_mutex);
3217 		return (DDI_SUCCESS);
3218 	case DDI_MO_UNMAP:
3219 
3220 		ap = (ddi_acc_impl_t *)reqp->map_handlep;
3221 		if (ap == NULL)
3222 			break;
3223 		/*
3224 		 * support for ddi_regs_map_free()
3225 		 * - check we really have a shadow handle for this one
3226 		 */
3227 		mutex_enter(&bofi_low_mutex);
3228 		mutex_enter(&bofi_mutex);
3229 		hhashp = HDL_HHASH(ap);
3230 		for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3231 			if (hp->hdl.acc_handle == (ddi_acc_handle_t)ap)
3232 				break;
3233 		if (hp == hhashp) {
3234 			mutex_exit(&bofi_mutex);
3235 			mutex_exit(&bofi_low_mutex);
3236 			break;
3237 		}
3238 		/*
3239 		 * got a shadow handle - restore original pointers
3240 		 */
3241 		*ap = hp->save.acc;
3242 		*vaddrp = hp->addr;
3243 		/*
3244 		 * remove from dhash, hhash and inuse lists
3245 		 */
3246 		hp->hnext->hprev = hp->hprev;
3247 		hp->hprev->hnext = hp->hnext;
3248 		hp->dnext->dprev = hp->dprev;
3249 		hp->dprev->dnext = hp->dnext;
3250 		hp->next->prev = hp->prev;
3251 		hp->prev->next = hp->next;
3252 		/*
3253 		 * free any errdef link structures tagged onto the shadow handle
3254 		 */
3255 		for (lp = hp->link; lp != NULL; ) {
3256 			next_lp = lp->link;
3257 			lp->link = bofi_link_freelist;
3258 			bofi_link_freelist = lp;
3259 			lp = next_lp;
3260 		}
3261 		hp->link = NULL;
3262 		mutex_exit(&bofi_mutex);
3263 		mutex_exit(&bofi_low_mutex);
3264 		/*
3265 		 * finally delete shadow handle
3266 		 */
3267 		kmem_free(hp, sizeof (struct bofi_shadow));
3268 		break;
3269 	default:
3270 		break;
3271 	}
3272 	return (save_bus_ops.bus_map(dip, rdip, reqp, offset, len, vaddrp));
3273 }
3274 
3275 
3276 /*
3277  * chain any pre-existing errdefs on to newly created dma handle
3278  * if required call do_dma_corrupt() to corrupt data
3279  */
3280 static void
3281 chain_on_errdefs(struct bofi_shadow *hp)
3282 {
3283 	struct bofi_errent *ep;
3284 	struct bofi_link   *lp;
3285 
3286 	ASSERT(MUTEX_HELD(&bofi_mutex));
3287 	/*
3288 	 * chain on any pre-existing errdefs that apply to this dma_handle
3289 	 */
3290 	for (ep = errent_listp; ep != NULL; ep = ep->next) {
3291 		if (ddi_name_to_major(hp->name) ==
3292 		    ddi_name_to_major(ep->name) &&
3293 		    hp->instance == ep->errdef.instance &&
3294 		    (ep->errdef.rnumber == -1 ||
3295 		    hp->rnumber == ep->errdef.rnumber) &&
3296 		    ((ep->errdef.access_type & BOFI_DMA_RW) &&
3297 		    (((uintptr_t)(hp->addr + ep->errdef.offset +
3298 		    ep->errdef.len) & ~LLSZMASK) >
3299 		    ((uintptr_t)((hp->addr + ep->errdef.offset) +
3300 		    LLSZMASK) & ~LLSZMASK)))) {
3301 			/*
3302 			 * got a match - link it on
3303 			 */
3304 			lp = bofi_link_freelist;
3305 			if (lp != NULL) {
3306 				bofi_link_freelist = lp->link;
3307 				lp->errentp = ep;
3308 				lp->link = hp->link;
3309 				hp->link = lp;
3310 				if ((ep->errdef.access_type & BOFI_DMA_W) &&
3311 				    (hp->flags & DDI_DMA_WRITE) &&
3312 				    (ep->state & BOFI_DEV_ACTIVE)) {
3313 					do_dma_corrupt(hp, ep,
3314 					    DDI_DMA_SYNC_FORDEV,
3315 					    0, hp->len);
3316 				}
3317 			}
3318 		}
3319 	}
3320 }
3321 
3322 
3323 /*
3324  * need to do copy byte-by-byte in case one of pages is little-endian
3325  */
3326 static void
3327 xbcopy(void *from, void *to, u_longlong_t len)
3328 {
3329 	uchar_t *f = from;
3330 	uchar_t *t = to;
3331 
3332 	while (len--)
3333 		*t++ = *f++;
3334 }
3335 
3336 
3337 /*
3338  * our ddi_dma_allochdl routine
3339  */
3340 static int
3341 bofi_dma_allochdl(dev_info_t *dip, dev_info_t *rdip, ddi_dma_attr_t *attrp,
3342     int (*waitfp)(caddr_t), caddr_t arg, ddi_dma_handle_t *handlep)
3343 {
3344 	int retval = DDI_DMA_NORESOURCES;
3345 	struct bofi_shadow *hp, *xhp;
3346 	int maxrnumber = 0;
3347 	struct bofi_shadow *dhashp;
3348 	struct bofi_shadow *hhashp;
3349 	ddi_dma_impl_t *mp;
3350 
3351 	/*
3352 	 * if driver_list is set, only intercept those drivers
3353 	 */
3354 	if (!driver_under_test(rdip))
3355 		return (save_bus_ops.bus_dma_allochdl(dip, rdip, attrp,
3356 		    waitfp, arg, handlep));
3357 
3358 	/*
3359 	 * allocate shadow handle structure and fill it in
3360 	 */
3361 	hp = kmem_zalloc(sizeof (struct bofi_shadow),
3362 	    ((waitfp == DDI_DMA_SLEEP) ? KM_SLEEP : KM_NOSLEEP));
3363 	if (hp == NULL) {
3364 		/*
3365 		 * what to do here? Wait a bit and try again
3366 		 */
3367 		if (waitfp != DDI_DMA_DONTWAIT)
3368 			(void) timeout((void (*)())(uintptr_t)waitfp, arg, 10);
3369 		return (retval);
3370 	}
3371 	(void) strncpy(hp->name, ddi_get_name(rdip), NAMESIZE);
3372 	hp->instance = ddi_get_instance(rdip);
3373 	hp->dip = rdip;
3374 	hp->link = NULL;
3375 	hp->type = BOFI_NULL;
3376 	/*
3377 	 * call nexus to do the real work
3378 	 */
3379 	retval = save_bus_ops.bus_dma_allochdl(dip, rdip, attrp, waitfp, arg,
3380 	    handlep);
3381 	if (retval != DDI_SUCCESS) {
3382 		kmem_free(hp, sizeof (struct bofi_shadow));
3383 		return (retval);
3384 	}
3385 	/*
3386 	 * now point set dma_handle to point to real handle
3387 	 */
3388 	hp->hdl.dma_handle = *handlep;
3389 	mp = (ddi_dma_impl_t *)*handlep;
3390 	mp->dmai_fault_check = bofi_check_dma_hdl;
3391 	/*
3392 	 * bind and unbind are cached in devinfo - must overwrite them
3393 	 * - note that our bind and unbind are quite happy dealing with
3394 	 * any handles for this devinfo that were previously allocated
3395 	 */
3396 	if (save_bus_ops.bus_dma_bindhdl == DEVI(rdip)->devi_bus_dma_bindfunc)
3397 		DEVI(rdip)->devi_bus_dma_bindfunc = bofi_dma_bindhdl;
3398 	if (save_bus_ops.bus_dma_unbindhdl ==
3399 	    DEVI(rdip)->devi_bus_dma_unbindfunc)
3400 		DEVI(rdip)->devi_bus_dma_unbindfunc = bofi_dma_unbindhdl;
3401 	mutex_enter(&bofi_low_mutex);
3402 	mutex_enter(&bofi_mutex);
3403 	/*
3404 	 * get an "rnumber" for this handle - really just seeking to
3405 	 * get a unique number - generally only care for early allocated
3406 	 * handles - so we get as far as INT_MAX, just stay there
3407 	 */
3408 	dhashp = HDL_DHASH(hp->dip);
3409 	for (xhp = dhashp->dnext; xhp != dhashp; xhp = xhp->dnext)
3410 		if (ddi_name_to_major(xhp->name) ==
3411 		    ddi_name_to_major(hp->name) &&
3412 		    xhp->instance == hp->instance &&
3413 		    (xhp->type == BOFI_DMA_HDL ||
3414 		    xhp->type == BOFI_NULL))
3415 			if (xhp->rnumber >= maxrnumber) {
3416 				if (xhp->rnumber == INT_MAX)
3417 					maxrnumber = INT_MAX;
3418 				else
3419 					maxrnumber = xhp->rnumber + 1;
3420 			}
3421 	hp->rnumber = maxrnumber;
3422 	/*
3423 	 * add to dhash, hhash and inuse lists
3424 	 */
3425 	hp->next = shadow_list.next;
3426 	shadow_list.next->prev = hp;
3427 	hp->prev = &shadow_list;
3428 	shadow_list.next = hp;
3429 	hhashp = HDL_HHASH(*handlep);
3430 	hp->hnext = hhashp->hnext;
3431 	hhashp->hnext->hprev = hp;
3432 	hp->hprev = hhashp;
3433 	hhashp->hnext = hp;
3434 	dhashp = HDL_DHASH(hp->dip);
3435 	hp->dnext = dhashp->dnext;
3436 	dhashp->dnext->dprev = hp;
3437 	hp->dprev = dhashp;
3438 	dhashp->dnext = hp;
3439 	mutex_exit(&bofi_mutex);
3440 	mutex_exit(&bofi_low_mutex);
3441 	return (retval);
3442 }
3443 
3444 
3445 /*
3446  * our ddi_dma_freehdl routine
3447  */
3448 static int
3449 bofi_dma_freehdl(dev_info_t *dip, dev_info_t *rdip, ddi_dma_handle_t handle)
3450 {
3451 	int retval;
3452 	struct bofi_shadow *hp;
3453 	struct bofi_shadow *hhashp;
3454 
3455 	/*
3456 	 * find shadow for this handle
3457 	 */
3458 	mutex_enter(&bofi_low_mutex);
3459 	mutex_enter(&bofi_mutex);
3460 	hhashp = HDL_HHASH(handle);
3461 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3462 		if (hp->hdl.dma_handle == handle)
3463 			break;
3464 	mutex_exit(&bofi_mutex);
3465 	mutex_exit(&bofi_low_mutex);
3466 	/*
3467 	 * call nexus to do the real work
3468 	 */
3469 	retval = save_bus_ops.bus_dma_freehdl(dip, rdip, handle);
3470 	if (retval != DDI_SUCCESS) {
3471 		return (retval);
3472 	}
3473 	/*
3474 	 * did we really have a shadow for this handle
3475 	 */
3476 	if (hp == hhashp)
3477 		return (retval);
3478 	/*
3479 	 * yes we have - see if it's still bound
3480 	 */
3481 	mutex_enter(&bofi_low_mutex);
3482 	mutex_enter(&bofi_mutex);
3483 	if (hp->type != BOFI_NULL)
3484 		panic("driver freeing bound dma_handle");
3485 	/*
3486 	 * remove from dhash, hhash and inuse lists
3487 	 */
3488 	hp->hnext->hprev = hp->hprev;
3489 	hp->hprev->hnext = hp->hnext;
3490 	hp->dnext->dprev = hp->dprev;
3491 	hp->dprev->dnext = hp->dnext;
3492 	hp->next->prev = hp->prev;
3493 	hp->prev->next = hp->next;
3494 	mutex_exit(&bofi_mutex);
3495 	mutex_exit(&bofi_low_mutex);
3496 
3497 	kmem_free(hp, sizeof (struct bofi_shadow));
3498 	return (retval);
3499 }
3500 
3501 
3502 /*
3503  * our ddi_dma_bindhdl routine
3504  */
3505 static int
3506 bofi_dma_bindhdl(dev_info_t *dip, dev_info_t *rdip,
3507     ddi_dma_handle_t handle, struct ddi_dma_req *dmareqp,
3508     ddi_dma_cookie_t *cookiep, uint_t *ccountp)
3509 {
3510 	int retval = DDI_DMA_NORESOURCES;
3511 	auto struct ddi_dma_req dmareq;
3512 	struct bofi_shadow *hp;
3513 	struct bofi_shadow *hhashp;
3514 	ddi_dma_impl_t *mp;
3515 	unsigned long pagemask = ddi_ptob(rdip, 1) - 1;
3516 
3517 	/*
3518 	 * check we really have a shadow for this handle
3519 	 */
3520 	mutex_enter(&bofi_low_mutex);
3521 	mutex_enter(&bofi_mutex);
3522 	hhashp = HDL_HHASH(handle);
3523 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3524 		if (hp->hdl.dma_handle == handle)
3525 			break;
3526 	mutex_exit(&bofi_mutex);
3527 	mutex_exit(&bofi_low_mutex);
3528 	if (hp == hhashp) {
3529 		/*
3530 		 * no we don't - just call nexus to do the real work
3531 		 */
3532 		return save_bus_ops.bus_dma_bindhdl(dip, rdip, handle, dmareqp,
3533 		    cookiep, ccountp);
3534 	}
3535 	/*
3536 	 * yes we have - see if it's already bound
3537 	 */
3538 	if (hp->type != BOFI_NULL)
3539 		return (DDI_DMA_INUSE);
3540 
3541 	hp->flags = dmareqp->dmar_flags;
3542 	if (dmareqp->dmar_object.dmao_type == DMA_OTYP_PAGES) {
3543 		hp->map_flags = B_PAGEIO;
3544 		hp->map_pp = dmareqp->dmar_object.dmao_obj.pp_obj.pp_pp;
3545 	} else if (dmareqp->dmar_object.dmao_obj.virt_obj.v_priv != NULL) {
3546 		hp->map_flags = B_SHADOW;
3547 		hp->map_pplist = dmareqp->dmar_object.dmao_obj.virt_obj.v_priv;
3548 	} else {
3549 		hp->map_flags = 0;
3550 	}
3551 	/*
3552 	 * get a kernel virtual mapping
3553 	 */
3554 	hp->addr = ddi_dmareq_mapin(dmareqp, &hp->mapaddr, &hp->len);
3555 	if (hp->addr == NULL)
3556 		goto error;
3557 	if (bofi_sync_check) {
3558 		/*
3559 		 * Take a copy and pass pointers to this up to nexus instead.
3560 		 * Data will be copied from the original on explicit
3561 		 * and implicit ddi_dma_sync()
3562 		 *
3563 		 * - maintain page alignment because some devices assume it.
3564 		 */
3565 		hp->origaddr = hp->addr;
3566 		hp->allocaddr = ddi_umem_alloc(
3567 		    ((uintptr_t)hp->addr & pagemask) + hp->len,
3568 		    (dmareqp->dmar_fp == DDI_DMA_SLEEP) ? KM_SLEEP : KM_NOSLEEP,
3569 		    &hp->umem_cookie);
3570 		if (hp->allocaddr == NULL)
3571 			goto error;
3572 		hp->addr = hp->allocaddr + ((uintptr_t)hp->addr & pagemask);
3573 		if (dmareqp->dmar_flags & DDI_DMA_WRITE)
3574 			xbcopy(hp->origaddr, hp->addr, hp->len);
3575 		dmareq = *dmareqp;
3576 		dmareq.dmar_object.dmao_size = hp->len;
3577 		dmareq.dmar_object.dmao_type = DMA_OTYP_VADDR;
3578 		dmareq.dmar_object.dmao_obj.virt_obj.v_as = &kas;
3579 		dmareq.dmar_object.dmao_obj.virt_obj.v_addr = hp->addr;
3580 		dmareq.dmar_object.dmao_obj.virt_obj.v_priv = NULL;
3581 		dmareqp = &dmareq;
3582 	}
3583 	/*
3584 	 * call nexus to do the real work
3585 	 */
3586 	retval = save_bus_ops.bus_dma_bindhdl(dip, rdip, handle, dmareqp,
3587 	    cookiep, ccountp);
3588 	if (retval != DDI_SUCCESS)
3589 		goto error2;
3590 	/*
3591 	 * unset DMP_NOSYNC
3592 	 */
3593 	mp = (ddi_dma_impl_t *)handle;
3594 	mp->dmai_rflags &= ~DMP_NOSYNC;
3595 	/*
3596 	 * chain on any pre-existing errdefs that apply to this
3597 	 * acc_handle and corrupt if required (as there is an implicit
3598 	 * ddi_dma_sync() in this call)
3599 	 */
3600 	mutex_enter(&bofi_low_mutex);
3601 	mutex_enter(&bofi_mutex);
3602 	hp->type = BOFI_DMA_HDL;
3603 	chain_on_errdefs(hp);
3604 	mutex_exit(&bofi_mutex);
3605 	mutex_exit(&bofi_low_mutex);
3606 	return (retval);
3607 
3608 error:
3609 	if (dmareqp->dmar_fp != DDI_DMA_DONTWAIT) {
3610 		/*
3611 		 * what to do here? Wait a bit and try again
3612 		 */
3613 		(void) timeout((void (*)())(uintptr_t)dmareqp->dmar_fp,
3614 		    dmareqp->dmar_arg, 10);
3615 	}
3616 error2:
3617 	if (hp) {
3618 		ddi_dmareq_mapout(hp->mapaddr, hp->len, hp->map_flags,
3619 		    hp->map_pp, hp->map_pplist);
3620 		if (bofi_sync_check && hp->allocaddr)
3621 			ddi_umem_free(hp->umem_cookie);
3622 		hp->mapaddr = NULL;
3623 		hp->allocaddr = NULL;
3624 		hp->origaddr = NULL;
3625 	}
3626 	return (retval);
3627 }
3628 
3629 
3630 /*
3631  * our ddi_dma_unbindhdl routine
3632  */
3633 static int
3634 bofi_dma_unbindhdl(dev_info_t *dip, dev_info_t *rdip, ddi_dma_handle_t handle)
3635 {
3636 	struct bofi_link *lp, *next_lp;
3637 	struct bofi_errent *ep;
3638 	int retval;
3639 	struct bofi_shadow *hp;
3640 	struct bofi_shadow *hhashp;
3641 
3642 	/*
3643 	 * call nexus to do the real work
3644 	 */
3645 	retval = save_bus_ops.bus_dma_unbindhdl(dip, rdip, handle);
3646 	if (retval != DDI_SUCCESS)
3647 		return (retval);
3648 	/*
3649 	 * check we really have a shadow for this handle
3650 	 */
3651 	mutex_enter(&bofi_low_mutex);
3652 	mutex_enter(&bofi_mutex);
3653 	hhashp = HDL_HHASH(handle);
3654 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3655 		if (hp->hdl.dma_handle == handle)
3656 			break;
3657 	if (hp == hhashp) {
3658 		mutex_exit(&bofi_mutex);
3659 		mutex_exit(&bofi_low_mutex);
3660 		return (retval);
3661 	}
3662 	/*
3663 	 * yes we have - see if it's already unbound
3664 	 */
3665 	if (hp->type == BOFI_NULL)
3666 		panic("driver unbinding unbound dma_handle");
3667 	/*
3668 	 * free any errdef link structures tagged on to this
3669 	 * shadow handle
3670 	 */
3671 	for (lp = hp->link; lp != NULL; ) {
3672 		next_lp = lp->link;
3673 		/*
3674 		 * there is an implicit sync_for_cpu on free -
3675 		 * may need to corrupt
3676 		 */
3677 		ep = lp->errentp;
3678 		if ((ep->errdef.access_type & BOFI_DMA_R) &&
3679 		    (hp->flags & DDI_DMA_READ) &&
3680 		    (ep->state & BOFI_DEV_ACTIVE)) {
3681 			do_dma_corrupt(hp, ep, DDI_DMA_SYNC_FORCPU, 0, hp->len);
3682 		}
3683 		lp->link = bofi_link_freelist;
3684 		bofi_link_freelist = lp;
3685 		lp = next_lp;
3686 	}
3687 	hp->link = NULL;
3688 	hp->type = BOFI_NULL;
3689 	mutex_exit(&bofi_mutex);
3690 	mutex_exit(&bofi_low_mutex);
3691 
3692 	if (bofi_sync_check && (hp->flags & DDI_DMA_READ))
3693 		/*
3694 		 * implicit sync_for_cpu - copy data back
3695 		 */
3696 		if (hp->allocaddr)
3697 			xbcopy(hp->addr, hp->origaddr, hp->len);
3698 	ddi_dmareq_mapout(hp->mapaddr, hp->len, hp->map_flags,
3699 	    hp->map_pp, hp->map_pplist);
3700 	if (bofi_sync_check && hp->allocaddr)
3701 		ddi_umem_free(hp->umem_cookie);
3702 	hp->mapaddr = NULL;
3703 	hp->allocaddr = NULL;
3704 	hp->origaddr = NULL;
3705 	return (retval);
3706 }
3707 
3708 
3709 /*
3710  * our ddi_dma_sync routine
3711  */
3712 static int
3713 bofi_dma_flush(dev_info_t *dip, dev_info_t *rdip,
3714     ddi_dma_handle_t handle, off_t off, size_t len, uint_t flags)
3715 {
3716 	struct bofi_link *lp;
3717 	struct bofi_errent *ep;
3718 	struct bofi_shadow *hp;
3719 	struct bofi_shadow *hhashp;
3720 	int retval;
3721 
3722 	if (flags == DDI_DMA_SYNC_FORCPU || flags == DDI_DMA_SYNC_FORKERNEL) {
3723 		/*
3724 		 * in this case get nexus driver to do sync first
3725 		 */
3726 		retval = save_bus_ops.bus_dma_flush(dip, rdip, handle, off,
3727 		    len, flags);
3728 		if (retval != DDI_SUCCESS)
3729 			return (retval);
3730 	}
3731 	/*
3732 	 * check we really have a shadow for this handle
3733 	 */
3734 	mutex_enter(&bofi_low_mutex);
3735 	mutex_enter(&bofi_mutex);
3736 	hhashp = HDL_HHASH(handle);
3737 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3738 		if (hp->hdl.dma_handle == handle &&
3739 		    hp->type == BOFI_DMA_HDL)
3740 			break;
3741 	mutex_exit(&bofi_mutex);
3742 	mutex_exit(&bofi_low_mutex);
3743 	if (hp != hhashp) {
3744 		/*
3745 		 * yes - do we need to copy data from original
3746 		 */
3747 		if (bofi_sync_check && flags == DDI_DMA_SYNC_FORDEV)
3748 			if (hp->allocaddr)
3749 				xbcopy(hp->origaddr+off, hp->addr+off,
3750 				    len ? len : (hp->len - off));
3751 		/*
3752 		 * yes - check if we need to corrupt the data
3753 		 */
3754 		mutex_enter(&bofi_low_mutex);
3755 		mutex_enter(&bofi_mutex);
3756 		for (lp = hp->link; lp != NULL; lp = lp->link) {
3757 			ep = lp->errentp;
3758 			if ((((ep->errdef.access_type & BOFI_DMA_R) &&
3759 			    (flags == DDI_DMA_SYNC_FORCPU ||
3760 			    flags == DDI_DMA_SYNC_FORKERNEL)) ||
3761 			    ((ep->errdef.access_type & BOFI_DMA_W) &&
3762 			    (flags == DDI_DMA_SYNC_FORDEV))) &&
3763 			    (ep->state & BOFI_DEV_ACTIVE)) {
3764 				do_dma_corrupt(hp, ep, flags, off,
3765 				    len ? len : (hp->len - off));
3766 			}
3767 		}
3768 		mutex_exit(&bofi_mutex);
3769 		mutex_exit(&bofi_low_mutex);
3770 		/*
3771 		 *  do we need to copy data to original
3772 		 */
3773 		if (bofi_sync_check && (flags == DDI_DMA_SYNC_FORCPU ||
3774 		    flags == DDI_DMA_SYNC_FORKERNEL))
3775 			if (hp->allocaddr)
3776 				xbcopy(hp->addr+off, hp->origaddr+off,
3777 				    len ? len : (hp->len - off));
3778 	}
3779 	if (flags == DDI_DMA_SYNC_FORDEV)
3780 		/*
3781 		 * in this case get nexus driver to do sync last
3782 		 */
3783 		retval = save_bus_ops.bus_dma_flush(dip, rdip, handle, off,
3784 		    len, flags);
3785 	return (retval);
3786 }
3787 
3788 
3789 /*
3790  * our dma_win routine
3791  */
3792 static int
3793 bofi_dma_win(dev_info_t *dip, dev_info_t *rdip,
3794     ddi_dma_handle_t handle, uint_t win, off_t *offp,
3795     size_t *lenp, ddi_dma_cookie_t *cookiep, uint_t *ccountp)
3796 {
3797 	struct bofi_shadow *hp;
3798 	struct bofi_shadow *hhashp;
3799 	int retval;
3800 	ddi_dma_impl_t *mp;
3801 
3802 	/*
3803 	 * call nexus to do the real work
3804 	 */
3805 	retval = save_bus_ops.bus_dma_win(dip, rdip, handle, win, offp, lenp,
3806 	    cookiep, ccountp);
3807 	if (retval != DDI_SUCCESS)
3808 		return (retval);
3809 	/*
3810 	 * check we really have a shadow for this handle
3811 	 */
3812 	mutex_enter(&bofi_low_mutex);
3813 	mutex_enter(&bofi_mutex);
3814 	hhashp = HDL_HHASH(handle);
3815 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3816 		if (hp->hdl.dma_handle == handle)
3817 			break;
3818 	if (hp != hhashp) {
3819 		/*
3820 		 * yes - make sure DMP_NOSYNC is unset
3821 		 */
3822 		mp = (ddi_dma_impl_t *)handle;
3823 		mp->dmai_rflags &= ~DMP_NOSYNC;
3824 	}
3825 	mutex_exit(&bofi_mutex);
3826 	mutex_exit(&bofi_low_mutex);
3827 	return (retval);
3828 }
3829 
3830 
3831 /*
3832  * our dma_ctl routine
3833  */
3834 static int
3835 bofi_dma_ctl(dev_info_t *dip, dev_info_t *rdip,
3836     ddi_dma_handle_t handle, enum ddi_dma_ctlops request,
3837     off_t *offp, size_t *lenp, caddr_t *objp, uint_t flags)
3838 {
3839 	struct bofi_shadow *hp;
3840 	struct bofi_shadow *hhashp;
3841 	int retval;
3842 	int i;
3843 	struct bofi_shadow *dummyhp;
3844 
3845 	/*
3846 	 * get nexus to do real work
3847 	 */
3848 	retval = save_bus_ops.bus_dma_ctl(dip, rdip, handle, request, offp,
3849 	    lenp, objp, flags);
3850 	if (retval != DDI_SUCCESS)
3851 		return (retval);
3852 	/*
3853 	 * if driver_list is set, only intercept those drivers
3854 	 */
3855 	if (!driver_under_test(rdip))
3856 		return (DDI_SUCCESS);
3857 
3858 #if defined(__sparc)
3859 	/*
3860 	 * check if this is a dvma_reserve - that one's like a
3861 	 * dma_allochdl and needs to be handled separately
3862 	 */
3863 	if (request == DDI_DMA_RESERVE) {
3864 		bofi_dvma_reserve(rdip, *(ddi_dma_handle_t *)objp);
3865 		return (DDI_SUCCESS);
3866 	}
3867 #endif
3868 	/*
3869 	 * check we really have a shadow for this handle
3870 	 */
3871 	mutex_enter(&bofi_low_mutex);
3872 	mutex_enter(&bofi_mutex);
3873 	hhashp = HDL_HHASH(handle);
3874 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
3875 		if (hp->hdl.dma_handle == handle)
3876 			break;
3877 	if (hp == hhashp) {
3878 		mutex_exit(&bofi_mutex);
3879 		mutex_exit(&bofi_low_mutex);
3880 		return (retval);
3881 	}
3882 	/*
3883 	 * yes we have - see what kind of command this is
3884 	 */
3885 	switch (request) {
3886 	case DDI_DMA_RELEASE:
3887 		/*
3888 		 * dvma release - release dummy handle and all the index handles
3889 		 */
3890 		dummyhp = hp;
3891 		dummyhp->hnext->hprev = dummyhp->hprev;
3892 		dummyhp->hprev->hnext = dummyhp->hnext;
3893 		mutex_exit(&bofi_mutex);
3894 		mutex_exit(&bofi_low_mutex);
3895 		for (i = 0; i < dummyhp->len; i++) {
3896 			hp = dummyhp->hparrayp[i];
3897 			/*
3898 			 * chek none of the index handles were still loaded
3899 			 */
3900 			if (hp->type != BOFI_NULL)
3901 				panic("driver releasing loaded dvma");
3902 			/*
3903 			 * remove from dhash and inuse lists
3904 			 */
3905 			mutex_enter(&bofi_low_mutex);
3906 			mutex_enter(&bofi_mutex);
3907 			hp->dnext->dprev = hp->dprev;
3908 			hp->dprev->dnext = hp->dnext;
3909 			hp->next->prev = hp->prev;
3910 			hp->prev->next = hp->next;
3911 			mutex_exit(&bofi_mutex);
3912 			mutex_exit(&bofi_low_mutex);
3913 
3914 			if (bofi_sync_check && hp->allocaddr)
3915 				ddi_umem_free(hp->umem_cookie);
3916 			kmem_free(hp, sizeof (struct bofi_shadow));
3917 		}
3918 		kmem_free(dummyhp->hparrayp, dummyhp->len *
3919 		    sizeof (struct bofi_shadow *));
3920 		kmem_free(dummyhp, sizeof (struct bofi_shadow));
3921 		return (retval);
3922 	default:
3923 		break;
3924 	}
3925 	mutex_exit(&bofi_mutex);
3926 	mutex_exit(&bofi_low_mutex);
3927 	return (retval);
3928 }
3929 
3930 #if defined(__sparc)
3931 /*
3932  * dvma reserve case from bofi_dma_ctl()
3933  */
3934 static void
3935 bofi_dvma_reserve(dev_info_t *rdip, ddi_dma_handle_t handle)
3936 {
3937 	struct bofi_shadow *hp;
3938 	struct bofi_shadow *dummyhp;
3939 	struct bofi_shadow *dhashp;
3940 	struct bofi_shadow *hhashp;
3941 	ddi_dma_impl_t *mp;
3942 	struct fast_dvma *nexus_private;
3943 	int i, count;
3944 
3945 	mp = (ddi_dma_impl_t *)handle;
3946 	count = mp->dmai_ndvmapages;
3947 	/*
3948 	 * allocate dummy shadow handle structure
3949 	 */
3950 	dummyhp = kmem_zalloc(sizeof (*dummyhp), KM_SLEEP);
3951 	if (mp->dmai_rflags & DMP_BYPASSNEXUS) {
3952 		/*
3953 		 * overlay our routines over the nexus's dvma routines
3954 		 */
3955 		nexus_private = (struct fast_dvma *)mp->dmai_nexus_private;
3956 		dummyhp->save.dvma_ops = *(nexus_private->ops);
3957 		nexus_private->ops = &bofi_dvma_ops;
3958 	}
3959 	/*
3960 	 * now fill in the dummy handle. This just gets put on hhash queue
3961 	 * so our dvma routines can find and index off to the handle they
3962 	 * really want.
3963 	 */
3964 	(void) strncpy(dummyhp->name, ddi_get_name(rdip), NAMESIZE);
3965 	dummyhp->instance = ddi_get_instance(rdip);
3966 	dummyhp->rnumber = -1;
3967 	dummyhp->dip = rdip;
3968 	dummyhp->len = count;
3969 	dummyhp->hdl.dma_handle = handle;
3970 	dummyhp->link = NULL;
3971 	dummyhp->type = BOFI_NULL;
3972 	/*
3973 	 * allocate space for real handles
3974 	 */
3975 	dummyhp->hparrayp = kmem_alloc(count *
3976 	    sizeof (struct bofi_shadow *), KM_SLEEP);
3977 	for (i = 0; i < count; i++) {
3978 		/*
3979 		 * allocate shadow handle structures and fill them in
3980 		 */
3981 		hp = kmem_zalloc(sizeof (*hp), KM_SLEEP);
3982 		(void) strncpy(hp->name, ddi_get_name(rdip), NAMESIZE);
3983 		hp->instance = ddi_get_instance(rdip);
3984 		hp->rnumber = -1;
3985 		hp->dip = rdip;
3986 		hp->hdl.dma_handle = 0;
3987 		hp->link = NULL;
3988 		hp->type = BOFI_NULL;
3989 		if (bofi_sync_check) {
3990 			unsigned long pagemask = ddi_ptob(rdip, 1) - 1;
3991 			/*
3992 			 * Take a copy and set this to be hp->addr
3993 			 * Data will be copied to and from the original on
3994 			 * explicit and implicit ddi_dma_sync()
3995 			 *
3996 			 * - maintain page alignment because some devices
3997 			 * assume it.
3998 			 */
3999 			hp->allocaddr = ddi_umem_alloc(
4000 			    ((int)(uintptr_t)hp->addr & pagemask)
4001 			    + pagemask + 1,
4002 			    KM_SLEEP, &hp->umem_cookie);
4003 			hp->addr = hp->allocaddr +
4004 			    ((int)(uintptr_t)hp->addr & pagemask);
4005 		}
4006 		/*
4007 		 * add to dhash and inuse lists.
4008 		 * these don't go on hhash queue.
4009 		 */
4010 		mutex_enter(&bofi_low_mutex);
4011 		mutex_enter(&bofi_mutex);
4012 		hp->next = shadow_list.next;
4013 		shadow_list.next->prev = hp;
4014 		hp->prev = &shadow_list;
4015 		shadow_list.next = hp;
4016 		dhashp = HDL_DHASH(hp->dip);
4017 		hp->dnext = dhashp->dnext;
4018 		dhashp->dnext->dprev = hp;
4019 		hp->dprev = dhashp;
4020 		dhashp->dnext = hp;
4021 		dummyhp->hparrayp[i] = hp;
4022 		mutex_exit(&bofi_mutex);
4023 		mutex_exit(&bofi_low_mutex);
4024 	}
4025 	/*
4026 	 * add dummy handle to hhash list only
4027 	 */
4028 	mutex_enter(&bofi_low_mutex);
4029 	mutex_enter(&bofi_mutex);
4030 	hhashp = HDL_HHASH(handle);
4031 	dummyhp->hnext = hhashp->hnext;
4032 	hhashp->hnext->hprev = dummyhp;
4033 	dummyhp->hprev = hhashp;
4034 	hhashp->hnext = dummyhp;
4035 	mutex_exit(&bofi_mutex);
4036 	mutex_exit(&bofi_low_mutex);
4037 }
4038 
4039 /*
4040  * our dvma_kaddr_load()
4041  */
4042 static void
4043 bofi_dvma_kaddr_load(ddi_dma_handle_t h, caddr_t a, uint_t len, uint_t index,
4044     ddi_dma_cookie_t *cp)
4045 {
4046 	struct bofi_shadow *dummyhp;
4047 	struct bofi_shadow *hp;
4048 	struct bofi_shadow *hhashp;
4049 	struct bofi_errent *ep;
4050 	struct bofi_link   *lp;
4051 
4052 	/*
4053 	 * check we really have a dummy shadow for this handle
4054 	 */
4055 	mutex_enter(&bofi_low_mutex);
4056 	mutex_enter(&bofi_mutex);
4057 	hhashp = HDL_HHASH(h);
4058 	for (dummyhp = hhashp->hnext; dummyhp != hhashp;
4059 	    dummyhp = dummyhp->hnext)
4060 		if (dummyhp->hdl.dma_handle == h)
4061 			break;
4062 	mutex_exit(&bofi_mutex);
4063 	mutex_exit(&bofi_low_mutex);
4064 	if (dummyhp == hhashp) {
4065 		/*
4066 		 * no dummy shadow - panic
4067 		 */
4068 		panic("driver dvma_kaddr_load with no reserve");
4069 	}
4070 
4071 	/*
4072 	 * find real hp
4073 	 */
4074 	hp = dummyhp->hparrayp[index];
4075 	/*
4076 	 * check its not already loaded
4077 	 */
4078 	if (hp->type != BOFI_NULL)
4079 		panic("driver loading loaded dvma");
4080 	/*
4081 	 * if were doing copying, just need to change origaddr and get
4082 	 * nexus to map hp->addr again
4083 	 * if not, set hp->addr to new address.
4084 	 * - note these are always kernel virtual addresses - no need to map
4085 	 */
4086 	if (bofi_sync_check && hp->allocaddr) {
4087 		hp->origaddr = a;
4088 		a = hp->addr;
4089 	} else
4090 		hp->addr = a;
4091 	hp->len = len;
4092 	/*
4093 	 * get nexus to do the real work
4094 	 */
4095 	dummyhp->save.dvma_ops.dvma_kaddr_load(h, a, len, index, cp);
4096 	/*
4097 	 * chain on any pre-existing errdefs that apply to this dma_handle
4098 	 * no need to corrupt - there's no implicit dma_sync on this one
4099 	 */
4100 	mutex_enter(&bofi_low_mutex);
4101 	mutex_enter(&bofi_mutex);
4102 	hp->type = BOFI_DMA_HDL;
4103 	for (ep = errent_listp; ep != NULL; ep = ep->next) {
4104 		if (ddi_name_to_major(hp->name) ==
4105 		    ddi_name_to_major(ep->name) &&
4106 		    hp->instance == ep->errdef.instance &&
4107 		    (ep->errdef.rnumber == -1 ||
4108 		    hp->rnumber == ep->errdef.rnumber) &&
4109 		    ((ep->errdef.access_type & BOFI_DMA_RW) &&
4110 		    (((uintptr_t)(hp->addr + ep->errdef.offset +
4111 		    ep->errdef.len) & ~LLSZMASK) >
4112 		    ((uintptr_t)((hp->addr + ep->errdef.offset) +
4113 		    LLSZMASK) & ~LLSZMASK)))) {
4114 			lp = bofi_link_freelist;
4115 			if (lp != NULL) {
4116 				bofi_link_freelist = lp->link;
4117 				lp->errentp = ep;
4118 				lp->link = hp->link;
4119 				hp->link = lp;
4120 			}
4121 		}
4122 	}
4123 	mutex_exit(&bofi_mutex);
4124 	mutex_exit(&bofi_low_mutex);
4125 }
4126 
4127 /*
4128  * our dvma_unload()
4129  */
4130 static void
4131 bofi_dvma_unload(ddi_dma_handle_t h, uint_t index, uint_t view)
4132 {
4133 	struct bofi_link *lp, *next_lp;
4134 	struct bofi_errent *ep;
4135 	struct bofi_shadow *dummyhp;
4136 	struct bofi_shadow *hp;
4137 	struct bofi_shadow *hhashp;
4138 
4139 	/*
4140 	 * check we really have a dummy shadow for this handle
4141 	 */
4142 	mutex_enter(&bofi_low_mutex);
4143 	mutex_enter(&bofi_mutex);
4144 	hhashp = HDL_HHASH(h);
4145 	for (dummyhp = hhashp->hnext; dummyhp != hhashp;
4146 	    dummyhp = dummyhp->hnext)
4147 		if (dummyhp->hdl.dma_handle == h)
4148 			break;
4149 	mutex_exit(&bofi_mutex);
4150 	mutex_exit(&bofi_low_mutex);
4151 	if (dummyhp == hhashp) {
4152 		/*
4153 		 * no dummy shadow - panic
4154 		 */
4155 		panic("driver dvma_unload with no reserve");
4156 	}
4157 	dummyhp->save.dvma_ops.dvma_unload(h, index, view);
4158 	/*
4159 	 * find real hp
4160 	 */
4161 	hp = dummyhp->hparrayp[index];
4162 	/*
4163 	 * check its not already unloaded
4164 	 */
4165 	if (hp->type == BOFI_NULL)
4166 		panic("driver unloading unloaded dvma");
4167 	/*
4168 	 * free any errdef link structures tagged on to this
4169 	 * shadow handle - do corruption if necessary
4170 	 */
4171 	mutex_enter(&bofi_low_mutex);
4172 	mutex_enter(&bofi_mutex);
4173 	for (lp = hp->link; lp != NULL; ) {
4174 		next_lp = lp->link;
4175 		ep = lp->errentp;
4176 		if ((ep->errdef.access_type & BOFI_DMA_R) &&
4177 		    (view == DDI_DMA_SYNC_FORCPU ||
4178 		    view == DDI_DMA_SYNC_FORKERNEL) &&
4179 		    (ep->state & BOFI_DEV_ACTIVE)) {
4180 			do_dma_corrupt(hp, ep, view, 0, hp->len);
4181 		}
4182 		lp->link = bofi_link_freelist;
4183 		bofi_link_freelist = lp;
4184 		lp = next_lp;
4185 	}
4186 	hp->link = NULL;
4187 	hp->type = BOFI_NULL;
4188 	mutex_exit(&bofi_mutex);
4189 	mutex_exit(&bofi_low_mutex);
4190 	/*
4191 	 * if there is an explicit sync_for_cpu, then do copy to original
4192 	 */
4193 	if (bofi_sync_check &&
4194 	    (view == DDI_DMA_SYNC_FORCPU || view == DDI_DMA_SYNC_FORKERNEL))
4195 		if (hp->allocaddr)
4196 			xbcopy(hp->addr, hp->origaddr, hp->len);
4197 }
4198 
4199 /*
4200  * our dvma_unload()
4201  */
4202 static void
4203 bofi_dvma_sync(ddi_dma_handle_t h, uint_t index, uint_t view)
4204 {
4205 	struct bofi_link *lp;
4206 	struct bofi_errent *ep;
4207 	struct bofi_shadow *hp;
4208 	struct bofi_shadow *dummyhp;
4209 	struct bofi_shadow *hhashp;
4210 
4211 	/*
4212 	 * check we really have a dummy shadow for this handle
4213 	 */
4214 	mutex_enter(&bofi_low_mutex);
4215 	mutex_enter(&bofi_mutex);
4216 	hhashp = HDL_HHASH(h);
4217 	for (dummyhp = hhashp->hnext; dummyhp != hhashp;
4218 	    dummyhp = dummyhp->hnext)
4219 		if (dummyhp->hdl.dma_handle == h)
4220 			break;
4221 	mutex_exit(&bofi_mutex);
4222 	mutex_exit(&bofi_low_mutex);
4223 	if (dummyhp == hhashp) {
4224 		/*
4225 		 * no dummy shadow - panic
4226 		 */
4227 		panic("driver dvma_sync with no reserve");
4228 	}
4229 	/*
4230 	 * find real hp
4231 	 */
4232 	hp = dummyhp->hparrayp[index];
4233 	/*
4234 	 * check its already loaded
4235 	 */
4236 	if (hp->type == BOFI_NULL)
4237 		panic("driver syncing unloaded dvma");
4238 	if (view == DDI_DMA_SYNC_FORCPU || view == DDI_DMA_SYNC_FORKERNEL)
4239 		/*
4240 		 * in this case do sync first
4241 		 */
4242 		dummyhp->save.dvma_ops.dvma_sync(h, index, view);
4243 	/*
4244 	 * if there is an explicit sync_for_dev, then do copy from original
4245 	 */
4246 	if (bofi_sync_check && view == DDI_DMA_SYNC_FORDEV) {
4247 		if (hp->allocaddr)
4248 			xbcopy(hp->origaddr, hp->addr, hp->len);
4249 	}
4250 	/*
4251 	 * do corruption if necessary
4252 	 */
4253 	mutex_enter(&bofi_low_mutex);
4254 	mutex_enter(&bofi_mutex);
4255 	for (lp = hp->link; lp != NULL; lp = lp->link) {
4256 		ep = lp->errentp;
4257 		if ((((ep->errdef.access_type & BOFI_DMA_R) &&
4258 		    (view == DDI_DMA_SYNC_FORCPU ||
4259 		    view == DDI_DMA_SYNC_FORKERNEL)) ||
4260 		    ((ep->errdef.access_type & BOFI_DMA_W) &&
4261 		    (view == DDI_DMA_SYNC_FORDEV))) &&
4262 		    (ep->state & BOFI_DEV_ACTIVE)) {
4263 			do_dma_corrupt(hp, ep, view, 0, hp->len);
4264 		}
4265 	}
4266 	mutex_exit(&bofi_mutex);
4267 	mutex_exit(&bofi_low_mutex);
4268 	/*
4269 	 * if there is an explicit sync_for_cpu, then do copy to original
4270 	 */
4271 	if (bofi_sync_check &&
4272 	    (view == DDI_DMA_SYNC_FORCPU || view == DDI_DMA_SYNC_FORKERNEL)) {
4273 		if (hp->allocaddr)
4274 			xbcopy(hp->addr, hp->origaddr, hp->len);
4275 	}
4276 	if (view == DDI_DMA_SYNC_FORDEV)
4277 		/*
4278 		 * in this case do sync last
4279 		 */
4280 		dummyhp->save.dvma_ops.dvma_sync(h, index, view);
4281 }
4282 #endif
4283 
4284 /*
4285  * bofi intercept routine - gets called instead of users interrupt routine
4286  */
4287 static uint_t
4288 bofi_intercept_intr(caddr_t xp, caddr_t arg2)
4289 {
4290 	struct bofi_errent *ep;
4291 	struct bofi_link   *lp;
4292 	struct bofi_shadow *hp;
4293 	int intr_count = 1;
4294 	int i;
4295 	uint_t retval = DDI_INTR_UNCLAIMED;
4296 	uint_t result;
4297 	int unclaimed_counter = 0;
4298 	int jabber_detected = 0;
4299 
4300 	hp = (struct bofi_shadow *)xp;
4301 	/*
4302 	 * check if nothing to do
4303 	 */
4304 	if (hp->link == NULL)
4305 		return (hp->save.intr.int_handler
4306 		    (hp->save.intr.int_handler_arg1, arg2));
4307 	mutex_enter(&bofi_mutex);
4308 	/*
4309 	 * look for any errdefs
4310 	 */
4311 	for (lp = hp->link; lp != NULL; lp = lp->link) {
4312 		ep = lp->errentp;
4313 		if (ep->state & BOFI_DEV_ACTIVE) {
4314 			/*
4315 			 * got one
4316 			 */
4317 			if ((ep->errdef.access_count ||
4318 			    ep->errdef.fail_count) &&
4319 			    (ep->errdef.access_type & BOFI_LOG))
4320 				log_acc_event(ep, BOFI_INTR, 0, 0, 1, 0);
4321 			if (ep->errdef.access_count > 1) {
4322 				ep->errdef.access_count--;
4323 			} else if (ep->errdef.fail_count > 0) {
4324 				ep->errdef.fail_count--;
4325 				ep->errdef.access_count = 0;
4326 				/*
4327 				 * OK do "corruption"
4328 				 */
4329 				if (ep->errstate.fail_time == 0)
4330 					ep->errstate.fail_time = bofi_gettime();
4331 				switch (ep->errdef.optype) {
4332 				case BOFI_DELAY_INTR:
4333 					if (!hp->hilevel) {
4334 						drv_usecwait
4335 						    (ep->errdef.operand);
4336 					}
4337 					break;
4338 				case BOFI_LOSE_INTR:
4339 					intr_count = 0;
4340 					break;
4341 				case BOFI_EXTRA_INTR:
4342 					intr_count += ep->errdef.operand;
4343 					break;
4344 				default:
4345 					break;
4346 				}
4347 			}
4348 		}
4349 	}
4350 	mutex_exit(&bofi_mutex);
4351 	/*
4352 	 * send extra or fewer interrupts as requested
4353 	 */
4354 	for (i = 0; i < intr_count; i++) {
4355 		result = hp->save.intr.int_handler
4356 		    (hp->save.intr.int_handler_arg1, arg2);
4357 		if (result == DDI_INTR_CLAIMED)
4358 			unclaimed_counter >>= 1;
4359 		else if (++unclaimed_counter >= 20)
4360 			jabber_detected = 1;
4361 		if (i == 0)
4362 			retval = result;
4363 	}
4364 	/*
4365 	 * if more than 1000 spurious interrupts requested and
4366 	 * jabber not detected - give warning
4367 	 */
4368 	if (intr_count > 1000 && !jabber_detected)
4369 		panic("undetected interrupt jabber: %s%d",
4370 		    hp->name, hp->instance);
4371 	/*
4372 	 * return first response - or "unclaimed" if none
4373 	 */
4374 	return (retval);
4375 }
4376 
4377 
4378 /*
4379  * our ddi_check_acc_hdl
4380  */
4381 /* ARGSUSED */
4382 static int
4383 bofi_check_acc_hdl(ddi_acc_impl_t *handle)
4384 {
4385 	struct bofi_shadow *hp;
4386 	struct bofi_link   *lp;
4387 	uint_t result = 0;
4388 
4389 	hp = handle->ahi_common.ah_bus_private;
4390 	if (!hp->link || !mutex_tryenter(&bofi_mutex)) {
4391 		return (0);
4392 	}
4393 	for (lp = hp->link; lp != NULL; lp = lp->link) {
4394 		/*
4395 		 * OR in error state from all associated
4396 		 * errdef structures
4397 		 */
4398 		if (lp->errentp->errdef.access_count == 0 &&
4399 		    (lp->errentp->state & BOFI_DEV_ACTIVE)) {
4400 			result = (lp->errentp->errdef.acc_chk & 1);
4401 		}
4402 	}
4403 	mutex_exit(&bofi_mutex);
4404 	return (result);
4405 }
4406 
4407 /*
4408  * our ddi_check_dma_hdl
4409  */
4410 /* ARGSUSED */
4411 static int
4412 bofi_check_dma_hdl(ddi_dma_impl_t *handle)
4413 {
4414 	struct bofi_shadow *hp;
4415 	struct bofi_link   *lp;
4416 	struct bofi_shadow *hhashp;
4417 	uint_t result = 0;
4418 
4419 	if (!mutex_tryenter(&bofi_mutex)) {
4420 		return (0);
4421 	}
4422 	hhashp = HDL_HHASH(handle);
4423 	for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext)
4424 		if (hp->hdl.dma_handle == (ddi_dma_handle_t)handle)
4425 			break;
4426 	if (hp == hhashp) {
4427 		mutex_exit(&bofi_mutex);
4428 		return (0);
4429 	}
4430 	if (!hp->link) {
4431 		mutex_exit(&bofi_mutex);
4432 		return (0);
4433 	}
4434 	for (lp = hp->link; lp != NULL; lp = lp->link) {
4435 		/*
4436 		 * OR in error state from all associated
4437 		 * errdef structures
4438 		 */
4439 		if (lp->errentp->errdef.access_count == 0 &&
4440 		    (lp->errentp->state & BOFI_DEV_ACTIVE)) {
4441 			result = ((lp->errentp->errdef.acc_chk & 2) ? 1 : 0);
4442 		}
4443 	}
4444 	mutex_exit(&bofi_mutex);
4445 	return (result);
4446 }
4447 
4448 
4449 /* ARGSUSED */
4450 static int
4451 bofi_post_event(dev_info_t *dip, dev_info_t *rdip,
4452     ddi_eventcookie_t eventhdl, void *impl_data)
4453 {
4454 	ddi_eventcookie_t ec;
4455 	struct ddi_fault_event_data *arg;
4456 	struct bofi_errent *ep;
4457 	struct bofi_shadow *hp;
4458 	struct bofi_shadow *dhashp;
4459 	struct bofi_link   *lp;
4460 
4461 	ASSERT(eventhdl);
4462 	if (ddi_get_eventcookie(dip, DDI_DEVI_FAULT_EVENT, &ec) != DDI_SUCCESS)
4463 		return (DDI_FAILURE);
4464 
4465 	if (ec != eventhdl)
4466 		return (save_bus_ops.bus_post_event(dip, rdip, eventhdl,
4467 		    impl_data));
4468 
4469 	arg = (struct ddi_fault_event_data *)impl_data;
4470 	mutex_enter(&bofi_mutex);
4471 	/*
4472 	 * find shadow handles with appropriate dev_infos
4473 	 * and set error reported on all associated errdef structures
4474 	 */
4475 	dhashp = HDL_DHASH(arg->f_dip);
4476 	for (hp = dhashp->dnext; hp != dhashp; hp = hp->dnext) {
4477 		if (hp->dip == arg->f_dip) {
4478 			for (lp = hp->link; lp != NULL; lp = lp->link) {
4479 				ep = lp->errentp;
4480 				ep->errstate.errmsg_count++;
4481 				if ((ep->errstate.msg_time == 0 ||
4482 				    ep->errstate.severity > arg->f_impact) &&
4483 				    (ep->state & BOFI_DEV_ACTIVE)) {
4484 					ep->errstate.msg_time = bofi_gettime();
4485 					ep->errstate.severity = arg->f_impact;
4486 					(void) strncpy(ep->errstate.buffer,
4487 					    arg->f_message, ERRMSGSIZE);
4488 					ddi_trigger_softintr(ep->softintr_id);
4489 				}
4490 			}
4491 		}
4492 	}
4493 	mutex_exit(&bofi_mutex);
4494 	return (save_bus_ops.bus_post_event(dip, rdip, eventhdl, impl_data));
4495 }
4496 
4497 /*ARGSUSED*/
4498 static int
4499 bofi_fm_ereport_callback(sysevent_t *ev, void *cookie)
4500 {
4501 	char *class = "";
4502 	char *path = "";
4503 	char *ptr;
4504 	nvlist_t *nvlist;
4505 	nvlist_t *detector;
4506 	ddi_fault_impact_t impact;
4507 	struct bofi_errent *ep;
4508 	struct bofi_shadow *hp;
4509 	struct bofi_link   *lp;
4510 	char service_class[FM_MAX_CLASS];
4511 	char hppath[MAXPATHLEN];
4512 	int service_ereport = 0;
4513 
4514 	(void) sysevent_get_attr_list(ev, &nvlist);
4515 	(void) nvlist_lookup_string(nvlist, FM_CLASS, &class);
4516 	if (nvlist_lookup_nvlist(nvlist, FM_EREPORT_DETECTOR, &detector) == 0)
4517 		(void) nvlist_lookup_string(detector, FM_FMRI_DEV_PATH, &path);
4518 
4519 	(void) snprintf(service_class, FM_MAX_CLASS, "%s.%s.%s.",
4520 	    FM_EREPORT_CLASS, DDI_IO_CLASS, DDI_FM_SERVICE_IMPACT);
4521 	if (strncmp(class, service_class, strlen(service_class) - 1) == 0)
4522 		service_ereport = 1;
4523 
4524 	mutex_enter(&bofi_mutex);
4525 	/*
4526 	 * find shadow handles with appropriate dev_infos
4527 	 * and set error reported on all associated errdef structures
4528 	 */
4529 	for (hp = shadow_list.next; hp != &shadow_list; hp = hp->next) {
4530 		(void) ddi_pathname(hp->dip, hppath);
4531 		if (strcmp(path, hppath) != 0)
4532 			continue;
4533 		for (lp = hp->link; lp != NULL; lp = lp->link) {
4534 			ep = lp->errentp;
4535 			ep->errstate.errmsg_count++;
4536 			if (!(ep->state & BOFI_DEV_ACTIVE))
4537 				continue;
4538 			if (ep->errstate.msg_time != 0)
4539 				continue;
4540 			if (service_ereport) {
4541 				ptr = class + strlen(service_class);
4542 				if (strcmp(ptr, DDI_FM_SERVICE_LOST) == 0)
4543 					impact = DDI_SERVICE_LOST;
4544 				else if (strcmp(ptr,
4545 				    DDI_FM_SERVICE_DEGRADED) == 0)
4546 					impact = DDI_SERVICE_DEGRADED;
4547 				else if (strcmp(ptr,
4548 				    DDI_FM_SERVICE_RESTORED) == 0)
4549 					impact = DDI_SERVICE_RESTORED;
4550 				else
4551 					impact = DDI_SERVICE_UNAFFECTED;
4552 				if (ep->errstate.severity > impact)
4553 					ep->errstate.severity = impact;
4554 			} else if (ep->errstate.buffer[0] == '\0') {
4555 				(void) strncpy(ep->errstate.buffer, class,
4556 				    ERRMSGSIZE);
4557 			}
4558 			if (ep->errstate.buffer[0] != '\0' &&
4559 			    ep->errstate.severity < DDI_SERVICE_RESTORED) {
4560 				ep->errstate.msg_time = bofi_gettime();
4561 				ddi_trigger_softintr(ep->softintr_id);
4562 			}
4563 		}
4564 	}
4565 	nvlist_free(nvlist);
4566 	mutex_exit(&bofi_mutex);
4567 	return (0);
4568 }
4569 
4570 /*
4571  * our intr_ops routine
4572  */
4573 static int
4574 bofi_intr_ops(dev_info_t *dip, dev_info_t *rdip, ddi_intr_op_t intr_op,
4575     ddi_intr_handle_impl_t *hdlp, void *result)
4576 {
4577 	int retval;
4578 	struct bofi_shadow *hp;
4579 	struct bofi_shadow *dhashp;
4580 	struct bofi_shadow *hhashp;
4581 	struct bofi_errent *ep;
4582 	struct bofi_link   *lp, *next_lp;
4583 
4584 	switch (intr_op) {
4585 	case DDI_INTROP_ADDISR:
4586 		/*
4587 		 * if driver_list is set, only intercept those drivers
4588 		 */
4589 		if (!driver_under_test(rdip))
4590 			return (save_bus_ops.bus_intr_op(dip, rdip,
4591 			    intr_op, hdlp, result));
4592 		/*
4593 		 * allocate shadow handle structure and fill in
4594 		 */
4595 		hp = kmem_zalloc(sizeof (struct bofi_shadow), KM_SLEEP);
4596 		(void) strncpy(hp->name, ddi_get_name(rdip), NAMESIZE);
4597 		hp->instance = ddi_get_instance(rdip);
4598 		hp->save.intr.int_handler = hdlp->ih_cb_func;
4599 		hp->save.intr.int_handler_arg1 = hdlp->ih_cb_arg1;
4600 		hdlp->ih_cb_func = (ddi_intr_handler_t *)bofi_intercept_intr;
4601 		hdlp->ih_cb_arg1 = (caddr_t)hp;
4602 		hp->bofi_inum = hdlp->ih_inum;
4603 		hp->dip = rdip;
4604 		hp->link = NULL;
4605 		hp->type = BOFI_INT_HDL;
4606 		/*
4607 		 * save whether hilevel or not
4608 		 */
4609 
4610 		if (hdlp->ih_pri >= ddi_intr_get_hilevel_pri())
4611 			hp->hilevel = 1;
4612 		else
4613 			hp->hilevel = 0;
4614 
4615 		/*
4616 		 * call nexus to do real work, but specifying our handler, and
4617 		 * our shadow handle as argument
4618 		 */
4619 		retval = save_bus_ops.bus_intr_op(dip, rdip,
4620 		    intr_op, hdlp, result);
4621 		if (retval != DDI_SUCCESS) {
4622 			kmem_free(hp, sizeof (struct bofi_shadow));
4623 			return (retval);
4624 		}
4625 		/*
4626 		 * add to dhash, hhash and inuse lists
4627 		 */
4628 		mutex_enter(&bofi_low_mutex);
4629 		mutex_enter(&bofi_mutex);
4630 		hp->next = shadow_list.next;
4631 		shadow_list.next->prev = hp;
4632 		hp->prev = &shadow_list;
4633 		shadow_list.next = hp;
4634 		hhashp = HDL_HHASH(hdlp->ih_inum);
4635 		hp->hnext = hhashp->hnext;
4636 		hhashp->hnext->hprev = hp;
4637 		hp->hprev = hhashp;
4638 		hhashp->hnext = hp;
4639 		dhashp = HDL_DHASH(hp->dip);
4640 		hp->dnext = dhashp->dnext;
4641 		dhashp->dnext->dprev = hp;
4642 		hp->dprev = dhashp;
4643 		dhashp->dnext = hp;
4644 		/*
4645 		 * chain on any pre-existing errdefs that apply to this
4646 		 * acc_handle
4647 		 */
4648 		for (ep = errent_listp; ep != NULL; ep = ep->next) {
4649 			if (ddi_name_to_major(hp->name) ==
4650 			    ddi_name_to_major(ep->name) &&
4651 			    hp->instance == ep->errdef.instance &&
4652 			    (ep->errdef.access_type & BOFI_INTR)) {
4653 				lp = bofi_link_freelist;
4654 				if (lp != NULL) {
4655 					bofi_link_freelist = lp->link;
4656 					lp->errentp = ep;
4657 					lp->link = hp->link;
4658 					hp->link = lp;
4659 				}
4660 			}
4661 		}
4662 		mutex_exit(&bofi_mutex);
4663 		mutex_exit(&bofi_low_mutex);
4664 		return (retval);
4665 	case DDI_INTROP_REMISR:
4666 		/*
4667 		 * call nexus routine first
4668 		 */
4669 		retval = save_bus_ops.bus_intr_op(dip, rdip,
4670 		    intr_op, hdlp, result);
4671 		/*
4672 		 * find shadow handle
4673 		 */
4674 		mutex_enter(&bofi_low_mutex);
4675 		mutex_enter(&bofi_mutex);
4676 		hhashp = HDL_HHASH(hdlp->ih_inum);
4677 		for (hp = hhashp->hnext; hp != hhashp; hp = hp->hnext) {
4678 			if (hp->dip == rdip &&
4679 			    hp->type == BOFI_INT_HDL &&
4680 			    hp->bofi_inum == hdlp->ih_inum) {
4681 				break;
4682 			}
4683 		}
4684 		if (hp == hhashp) {
4685 			mutex_exit(&bofi_mutex);
4686 			mutex_exit(&bofi_low_mutex);
4687 			return (retval);
4688 		}
4689 		/*
4690 		 * found one - remove from dhash, hhash and inuse lists
4691 		 */
4692 		hp->hnext->hprev = hp->hprev;
4693 		hp->hprev->hnext = hp->hnext;
4694 		hp->dnext->dprev = hp->dprev;
4695 		hp->dprev->dnext = hp->dnext;
4696 		hp->next->prev = hp->prev;
4697 		hp->prev->next = hp->next;
4698 		/*
4699 		 * free any errdef link structures
4700 		 * tagged on to this shadow handle
4701 		 */
4702 		for (lp = hp->link; lp != NULL; ) {
4703 			next_lp = lp->link;
4704 			lp->link = bofi_link_freelist;
4705 			bofi_link_freelist = lp;
4706 			lp = next_lp;
4707 		}
4708 		hp->link = NULL;
4709 		mutex_exit(&bofi_mutex);
4710 		mutex_exit(&bofi_low_mutex);
4711 		kmem_free(hp, sizeof (struct bofi_shadow));
4712 		return (retval);
4713 	default:
4714 		return (save_bus_ops.bus_intr_op(dip, rdip,
4715 		    intr_op, hdlp, result));
4716 	}
4717 }
4718