1 /*
2 * CDDL HEADER START
3 *
4 * The contents of this file are subject to the terms of the
5 * Common Development and Distribution License (the "License").
6 * You may not use this file except in compliance with the License.
7 *
8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9 * or http://www.opensolaris.org/os/licensing.
10 * See the License for the specific language governing permissions
11 * and limitations under the License.
12 *
13 * When distributing Covered Code, include this CDDL HEADER in each
14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15 * If applicable, add the following below this CDDL HEADER, with the
16 * fields enclosed by brackets "[]" replaced with your own identifying
17 * information: Portions Copyright [yyyy] [name of copyright owner]
18 *
19 * CDDL HEADER END
20 */
21 /*
22 * Copyright 2009 Sun Microsystems, Inc. All rights reserved.
23 * Use is subject to license terms.
24 */
25
26 #include <sys/types.h>
27 #include <sys/param.h>
28 #include <sys/systm.h>
29 #include <sys/buf.h>
30 #include <sys/uio.h>
31 #include <sys/cred.h>
32 #include <sys/poll.h>
33 #include <sys/mman.h>
34 #include <sys/kmem.h>
35 #include <sys/model.h>
36 #include <sys/file.h>
37 #include <sys/proc.h>
38 #include <sys/open.h>
39 #include <sys/user.h>
40 #include <sys/t_lock.h>
41 #include <sys/vm.h>
42 #include <sys/stat.h>
43 #include <vm/hat.h>
44 #include <vm/seg.h>
45 #include <vm/as.h>
46 #include <sys/cmn_err.h>
47 #include <sys/debug.h>
48 #include <sys/avintr.h>
49 #include <sys/autoconf.h>
50 #include <sys/sunddi.h>
51 #include <sys/esunddi.h>
52 #include <sys/sunndi.h>
53 #include <sys/ddi.h>
54 #include <sys/kstat.h>
55 #include <sys/conf.h>
56 #include <sys/ddi_impldefs.h> /* include implementation structure defs */
57 #include <sys/ndi_impldefs.h>
58 #include <sys/hwconf.h>
59 #include <sys/pathname.h>
60 #include <sys/modctl.h>
61 #include <sys/epm.h>
62 #include <sys/devctl.h>
63 #include <sys/callb.h>
64 #include <sys/bootconf.h>
65 #include <sys/dacf_impl.h>
66 #include <sys/nvpair.h>
67 #include <sys/sunmdi.h>
68 #include <sys/fs/dv_node.h>
69 #include <sys/sunldi_impl.h>
70
71 #ifdef __sparc
72 #include <sys/archsystm.h> /* getpil/setpil */
73 #include <sys/membar.h> /* membar_sync */
74 #endif
75
76 /*
77 * ndi property handling
78 */
79 int
ndi_prop_update_int(dev_t match_dev,dev_info_t * dip,char * name,int data)80 ndi_prop_update_int(dev_t match_dev, dev_info_t *dip,
81 char *name, int data)
82 {
83 return (ddi_prop_update_common(match_dev, dip,
84 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP,
85 name, &data, 1, ddi_prop_fm_encode_ints));
86 }
87
88 int
ndi_prop_update_int64(dev_t match_dev,dev_info_t * dip,char * name,int64_t data)89 ndi_prop_update_int64(dev_t match_dev, dev_info_t *dip,
90 char *name, int64_t data)
91 {
92 return (ddi_prop_update_common(match_dev, dip,
93 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP,
94 name, &data, 1, ddi_prop_fm_encode_int64));
95 }
96
97 int
ndi_prop_create_boolean(dev_t match_dev,dev_info_t * dip,char * name)98 ndi_prop_create_boolean(dev_t match_dev, dev_info_t *dip,
99 char *name)
100 {
101 return (ddi_prop_update_common(match_dev, dip,
102 DDI_PROP_HW_DEF | DDI_PROP_TYPE_ANY | DDI_PROP_DONTSLEEP,
103 name, NULL, 0, ddi_prop_fm_encode_bytes));
104 }
105
106 int
ndi_prop_update_int_array(dev_t match_dev,dev_info_t * dip,char * name,int * data,uint_t nelements)107 ndi_prop_update_int_array(dev_t match_dev, dev_info_t *dip,
108 char *name, int *data, uint_t nelements)
109 {
110 return (ddi_prop_update_common(match_dev, dip,
111 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP,
112 name, data, nelements, ddi_prop_fm_encode_ints));
113 }
114
115 int
ndi_prop_update_int64_array(dev_t match_dev,dev_info_t * dip,char * name,int64_t * data,uint_t nelements)116 ndi_prop_update_int64_array(dev_t match_dev, dev_info_t *dip,
117 char *name, int64_t *data, uint_t nelements)
118 {
119 return (ddi_prop_update_common(match_dev, dip,
120 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP,
121 name, data, nelements, ddi_prop_fm_encode_int64));
122 }
123
124 int
ndi_prop_update_string(dev_t match_dev,dev_info_t * dip,char * name,char * data)125 ndi_prop_update_string(dev_t match_dev, dev_info_t *dip,
126 char *name, char *data)
127 {
128 return (ddi_prop_update_common(match_dev, dip,
129 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP,
130 name, &data, 1, ddi_prop_fm_encode_string));
131 }
132
133 int
ndi_prop_update_string_array(dev_t match_dev,dev_info_t * dip,char * name,char ** data,uint_t nelements)134 ndi_prop_update_string_array(dev_t match_dev, dev_info_t *dip,
135 char *name, char **data, uint_t nelements)
136 {
137 return (ddi_prop_update_common(match_dev, dip,
138 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP,
139 name, data, nelements,
140 ddi_prop_fm_encode_strings));
141 }
142
143 int
ndi_prop_update_byte_array(dev_t match_dev,dev_info_t * dip,char * name,uchar_t * data,uint_t nelements)144 ndi_prop_update_byte_array(dev_t match_dev, dev_info_t *dip,
145 char *name, uchar_t *data, uint_t nelements)
146 {
147 if (nelements == 0)
148 return (DDI_PROP_INVAL_ARG);
149
150 return (ddi_prop_update_common(match_dev, dip,
151 DDI_PROP_HW_DEF | DDI_PROP_TYPE_BYTE | DDI_PROP_DONTSLEEP,
152 name, data, nelements, ddi_prop_fm_encode_bytes));
153 }
154
155 int
ndi_prop_remove(dev_t dev,dev_info_t * dip,char * name)156 ndi_prop_remove(dev_t dev, dev_info_t *dip, char *name)
157 {
158 return (ddi_prop_remove_common(dev, dip, name, DDI_PROP_HW_DEF));
159 }
160
161 void
ndi_prop_remove_all(dev_info_t * dip)162 ndi_prop_remove_all(dev_info_t *dip)
163 {
164 i_ddi_prop_dyn_parent_set(dip, NULL);
165 ddi_prop_remove_all_common(dip, (int)DDI_PROP_HW_DEF);
166 }
167
168 /*
169 * Post an event notification to nexus driver responsible for handling
170 * the event. The responsible nexus is defined in the cookie passed in as
171 * the third parameter.
172 * The dip parameter is an artifact of an older implementation in which all
173 * requests to remove an eventcall would bubble up the tree. Today, this
174 * parameter is ignored.
175 * Input Parameters:
176 * dip - Ignored.
177 * rdip - device driver posting the event
178 * cookie - valid ddi_eventcookie_t, obtained by caller prior to
179 * invocation of this routine
180 * impl_data - used by framework
181 */
182 /*ARGSUSED*/
183 int
ndi_post_event(dev_info_t * dip,dev_info_t * rdip,ddi_eventcookie_t cookie,void * impl_data)184 ndi_post_event(dev_info_t *dip, dev_info_t *rdip,
185 ddi_eventcookie_t cookie, void *impl_data)
186 {
187 dev_info_t *ddip;
188
189 ASSERT(cookie);
190 ddip = NDI_EVENT_DDIP(cookie);
191
192 /*
193 * perform sanity checks. These conditions should never be true.
194 */
195
196 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL);
197 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6);
198 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event != NULL);
199
200 /*
201 * post the event to the responsible ancestor
202 */
203 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event))
204 (ddip, rdip, cookie, impl_data));
205 }
206
207 /*
208 * Calls the bus nexus driver's implementation of the
209 * (*bus_remove_eventcall)() interface.
210 */
211 int
ndi_busop_remove_eventcall(dev_info_t * ddip,ddi_callback_id_t id)212 ndi_busop_remove_eventcall(dev_info_t *ddip, ddi_callback_id_t id)
213 {
214
215 ASSERT(id);
216 /* check for a correct revno before calling up the device tree. */
217 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL);
218 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6);
219
220 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall == NULL)
221 return (DDI_FAILURE);
222
223 /*
224 * request responsible nexus to remove the eventcall
225 */
226 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall))
227 (ddip, id));
228 }
229
230 /*
231 * Calls the bus nexus driver's implementation of the
232 * (*bus_add_eventcall)() interface. The dip parameter is an
233 * artifact of an older implementation in which all requests to
234 * add an eventcall would bubble up the tree. Today, this parameter is
235 * ignored.
236 */
237 /*ARGSUSED*/
238 int
ndi_busop_add_eventcall(dev_info_t * dip,dev_info_t * rdip,ddi_eventcookie_t cookie,void (* callback)(),void * arg,ddi_callback_id_t * cb_id)239 ndi_busop_add_eventcall(dev_info_t *dip, dev_info_t *rdip,
240 ddi_eventcookie_t cookie, void (*callback)(), void *arg,
241 ddi_callback_id_t *cb_id)
242 {
243 dev_info_t *ddip = (dev_info_t *)NDI_EVENT_DDIP(cookie);
244
245 /*
246 * check for a correct revno before calling up the device tree.
247 */
248 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL);
249 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6);
250
251 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall == NULL)
252 return (DDI_FAILURE);
253
254 /*
255 * request responsible ancestor to add the eventcall
256 */
257 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall))
258 (ddip, rdip, cookie, callback, arg, cb_id));
259 }
260
261 /*
262 * Calls the bus nexus driver's implementation of the
263 * (*bus_get_eventcookie)() interface up the device tree hierarchy.
264 */
265 int
ndi_busop_get_eventcookie(dev_info_t * dip,dev_info_t * rdip,char * name,ddi_eventcookie_t * event_cookiep)266 ndi_busop_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, char *name,
267 ddi_eventcookie_t *event_cookiep)
268 {
269 dev_info_t *pdip = (dev_info_t *)DEVI(dip)->devi_parent;
270
271 /* Can not be called from rootnex. */
272 ASSERT(pdip);
273
274 /*
275 * check for a correct revno before calling up the device tree.
276 */
277 ASSERT(DEVI(pdip)->devi_ops->devo_bus_ops != NULL);
278
279 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev < BUSO_REV_6) ||
280 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie == NULL)) {
281 #ifdef DEBUG
282 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev >=
283 BUSO_REV_3) &&
284 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie)) {
285 cmn_err(CE_WARN,
286 "Warning: %s%d busops_rev=%d no longer supported"
287 " by the NDI event framework.\nBUSO_REV_6 or "
288 "greater must be used.",
289 DEVI(pdip)->devi_binding_name,
290 DEVI(pdip)->devi_instance,
291 DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev);
292 }
293 #endif /* DEBUG */
294
295 return (ndi_busop_get_eventcookie(pdip, rdip, name,
296 event_cookiep));
297 }
298
299 return ((*(DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie))
300 (pdip, rdip, name, event_cookiep));
301 }
302
303 /*
304 * Copy in the devctl IOCTL data and return a handle to
305 * the data.
306 */
307 int
ndi_dc_allochdl(void * iocarg,struct devctl_iocdata ** rdcp)308 ndi_dc_allochdl(void *iocarg, struct devctl_iocdata **rdcp)
309 {
310 struct devctl_iocdata *dcp;
311 char *cpybuf;
312
313 ASSERT(rdcp != NULL);
314
315 dcp = kmem_zalloc(sizeof (*dcp), KM_SLEEP);
316
317 if (get_udatamodel() == DATAMODEL_NATIVE) {
318 if (copyin(iocarg, dcp, sizeof (*dcp)) != 0) {
319 kmem_free(dcp, sizeof (*dcp));
320 return (NDI_FAULT);
321 }
322 }
323 #ifdef _SYSCALL32_IMPL
324 else {
325 struct devctl_iocdata32 dcp32;
326
327 if (copyin(iocarg, &dcp32, sizeof (dcp32)) != 0) {
328 kmem_free(dcp, sizeof (*dcp));
329 return (NDI_FAULT);
330 }
331 dcp->cmd = (uint_t)dcp32.cmd;
332 dcp->flags = (uint_t)dcp32.flags;
333 dcp->cpyout_buf = (uint_t *)(uintptr_t)dcp32.cpyout_buf;
334 dcp->nvl_user = (nvlist_t *)(uintptr_t)dcp32.nvl_user;
335 dcp->nvl_usersz = (size_t)dcp32.nvl_usersz;
336 dcp->c_nodename = (char *)(uintptr_t)dcp32.c_nodename;
337 dcp->c_unitaddr = (char *)(uintptr_t)dcp32.c_unitaddr;
338 }
339 #endif
340 if (dcp->c_nodename != NULL) {
341 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP);
342 if (copyinstr(dcp->c_nodename, cpybuf, MAXNAMELEN, 0) != 0) {
343 kmem_free(cpybuf, MAXNAMELEN);
344 kmem_free(dcp, sizeof (*dcp));
345 return (NDI_FAULT);
346 }
347 cpybuf[MAXNAMELEN - 1] = '\0';
348 dcp->c_nodename = cpybuf;
349 }
350
351 if (dcp->c_unitaddr != NULL) {
352 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP);
353 if (copyinstr(dcp->c_unitaddr, cpybuf, MAXNAMELEN, 0) != 0) {
354 kmem_free(cpybuf, MAXNAMELEN);
355 if (dcp->c_nodename != NULL)
356 kmem_free(dcp->c_nodename, MAXNAMELEN);
357 kmem_free(dcp, sizeof (*dcp));
358 return (NDI_FAULT);
359 }
360 cpybuf[MAXNAMELEN - 1] = '\0';
361 dcp->c_unitaddr = cpybuf;
362 }
363
364 /*
365 * copyin and unpack a user defined nvlist if one was passed
366 */
367 if (dcp->nvl_user != NULL) {
368 if ((dcp->nvl_usersz == 0) ||
369 (dcp->nvl_usersz > DEVCTL_MAX_NVL_USERSZ)) {
370 if (dcp->c_nodename != NULL)
371 kmem_free(dcp->c_nodename, MAXNAMELEN);
372 if (dcp->c_unitaddr != NULL)
373 kmem_free(dcp->c_unitaddr, MAXNAMELEN);
374 kmem_free(dcp, sizeof (*dcp));
375 return (NDI_FAILURE);
376 }
377 cpybuf = kmem_alloc(dcp->nvl_usersz, KM_SLEEP);
378 if (copyin(dcp->nvl_user, cpybuf, dcp->nvl_usersz) != 0) {
379 kmem_free(cpybuf, dcp->nvl_usersz);
380 if (dcp->c_nodename != NULL)
381 kmem_free(dcp->c_nodename, MAXNAMELEN);
382 if (dcp->c_unitaddr != NULL)
383 kmem_free(dcp->c_unitaddr, MAXNAMELEN);
384 kmem_free(dcp, sizeof (*dcp));
385 return (NDI_FAULT);
386 }
387
388 if (nvlist_unpack(cpybuf, dcp->nvl_usersz, &dcp->nvl_user,
389 KM_SLEEP)) {
390 kmem_free(cpybuf, dcp->nvl_usersz);
391 if (dcp->c_nodename != NULL)
392 kmem_free(dcp->c_nodename, MAXNAMELEN);
393 if (dcp->c_unitaddr != NULL)
394 kmem_free(dcp->c_unitaddr, MAXNAMELEN);
395 kmem_free(dcp, sizeof (*dcp));
396 return (NDI_FAULT);
397 }
398 /*
399 * free the buffer containing the packed nvlist
400 */
401 kmem_free(cpybuf, dcp->nvl_usersz);
402
403 }
404
405 *rdcp = dcp;
406 return (NDI_SUCCESS);
407 }
408
409 /*
410 * free all space allocated to a handle.
411 */
412 void
ndi_dc_freehdl(struct devctl_iocdata * dcp)413 ndi_dc_freehdl(struct devctl_iocdata *dcp)
414 {
415 ASSERT(dcp != NULL);
416
417 if (dcp->c_nodename != NULL)
418 kmem_free(dcp->c_nodename, MAXNAMELEN);
419
420 if (dcp->c_unitaddr != NULL)
421 kmem_free(dcp->c_unitaddr, MAXNAMELEN);
422
423 if (dcp->nvl_user != NULL)
424 nvlist_free(dcp->nvl_user);
425
426 kmem_free(dcp, sizeof (*dcp));
427 }
428
429 char *
ndi_dc_getname(struct devctl_iocdata * dcp)430 ndi_dc_getname(struct devctl_iocdata *dcp)
431 {
432 ASSERT(dcp != NULL);
433 return (dcp->c_nodename);
434
435 }
436
437 char *
ndi_dc_getaddr(struct devctl_iocdata * dcp)438 ndi_dc_getaddr(struct devctl_iocdata *dcp)
439 {
440 ASSERT(dcp != NULL);
441 return (dcp->c_unitaddr);
442 }
443
444 nvlist_t *
ndi_dc_get_ap_data(struct devctl_iocdata * dcp)445 ndi_dc_get_ap_data(struct devctl_iocdata *dcp)
446 {
447 ASSERT(dcp != NULL);
448
449 return (dcp->nvl_user);
450 }
451
452 /*
453 * Transition the child named by "devname@devaddr" to the online state.
454 * For use by a driver's DEVCTL_DEVICE_ONLINE handler.
455 */
456 int
ndi_devctl_device_online(dev_info_t * dip,struct devctl_iocdata * dcp,uint_t flags)457 ndi_devctl_device_online(dev_info_t *dip, struct devctl_iocdata *dcp,
458 uint_t flags)
459 {
460 int rval;
461 char *name;
462 dev_info_t *rdip;
463
464 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL)
465 return (EINVAL);
466
467 name = kmem_alloc(MAXNAMELEN, KM_SLEEP);
468 (void) snprintf(name, MAXNAMELEN, "%s@%s",
469 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp));
470
471 if ((rval = ndi_devi_config_one(dip, name, &rdip,
472 flags | NDI_DEVI_ONLINE | NDI_CONFIG)) == NDI_SUCCESS) {
473 ndi_rele_devi(rdip);
474
475 /*
476 * Invalidate devfs cached directory contents. For the checks
477 * in the "if" condition see the comment in ndi_devi_online().
478 */
479 if (i_ddi_devi_attached(dip) && !DEVI_BUSY_OWNED(dip))
480 (void) devfs_clean(dip, NULL, 0);
481
482 } else if (rval == NDI_BUSY) {
483 rval = EBUSY;
484 } else if (rval == NDI_FAILURE) {
485 rval = EIO;
486 }
487
488 NDI_DEBUG(flags, (CE_CONT, "%s%d: online: %s: %s\n",
489 ddi_driver_name(dip), ddi_get_instance(dip), name,
490 ((rval == NDI_SUCCESS) ? "ok" : "failed")));
491
492 kmem_free(name, MAXNAMELEN);
493
494 return (rval);
495 }
496
497 /*
498 * Transition the child named by "devname@devaddr" to the offline state.
499 * For use by a driver's DEVCTL_DEVICE_OFFLINE handler.
500 */
501 int
ndi_devctl_device_offline(dev_info_t * dip,struct devctl_iocdata * dcp,uint_t flags)502 ndi_devctl_device_offline(dev_info_t *dip, struct devctl_iocdata *dcp,
503 uint_t flags)
504 {
505 int rval;
506 char *name;
507
508 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL)
509 return (EINVAL);
510
511 name = kmem_alloc(MAXNAMELEN, KM_SLEEP);
512 (void) snprintf(name, MAXNAMELEN, "%s@%s",
513 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp));
514
515 (void) devfs_clean(dip, name, DV_CLEAN_FORCE);
516 rval = ndi_devi_unconfig_one(dip, name, NULL,
517 flags | NDI_DEVI_OFFLINE);
518
519 if (rval == NDI_BUSY) {
520 rval = EBUSY;
521 } else if (rval == NDI_FAILURE) {
522 rval = EIO;
523 }
524
525 NDI_DEBUG(flags, (CE_CONT, "%s%d: offline: %s: %s\n",
526 ddi_driver_name(dip), ddi_get_instance(dip), name,
527 (rval == NDI_SUCCESS) ? "ok" : "failed"));
528
529 kmem_free(name, MAXNAMELEN);
530
531 return (rval);
532 }
533
534 /*
535 * Remove the child named by "devname@devaddr".
536 * For use by a driver's DEVCTL_DEVICE_REMOVE handler.
537 */
538 int
ndi_devctl_device_remove(dev_info_t * dip,struct devctl_iocdata * dcp,uint_t flags)539 ndi_devctl_device_remove(dev_info_t *dip, struct devctl_iocdata *dcp,
540 uint_t flags)
541 {
542 int rval;
543 char *name;
544
545 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL)
546 return (EINVAL);
547
548 name = kmem_alloc(MAXNAMELEN, KM_SLEEP);
549 (void) snprintf(name, MAXNAMELEN, "%s@%s",
550 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp));
551
552 (void) devfs_clean(dip, name, DV_CLEAN_FORCE);
553
554 rval = ndi_devi_unconfig_one(dip, name, NULL, flags | NDI_DEVI_REMOVE);
555
556 if (rval == NDI_BUSY) {
557 rval = EBUSY;
558 } else if (rval == NDI_FAILURE) {
559 rval = EIO;
560 }
561
562 NDI_DEBUG(flags, (CE_CONT, "%s%d: remove: %s: %s\n",
563 ddi_driver_name(dip), ddi_get_instance(dip), name,
564 (rval == NDI_SUCCESS) ? "ok" : "failed"));
565
566 kmem_free(name, MAXNAMELEN);
567
568 return (rval);
569 }
570
571 /*
572 * Return devctl state of the child named by "name@addr".
573 * For use by a driver's DEVCTL_DEVICE_GETSTATE handler.
574 */
575 int
ndi_devctl_device_getstate(dev_info_t * parent,struct devctl_iocdata * dcp,uint_t * state)576 ndi_devctl_device_getstate(dev_info_t *parent, struct devctl_iocdata *dcp,
577 uint_t *state)
578 {
579 dev_info_t *dip;
580 char *name, *addr;
581 char *devname;
582 int devnamelen;
583 int circ;
584
585 if (parent == NULL ||
586 ((name = ndi_dc_getname(dcp)) == NULL) ||
587 ((addr = ndi_dc_getaddr(dcp)) == NULL))
588 return (NDI_FAILURE);
589
590 devnamelen = strlen(name) + strlen(addr) + 2;
591 devname = kmem_alloc(devnamelen, KM_SLEEP);
592 if (strlen(addr) > 0) {
593 (void) snprintf(devname, devnamelen, "%s@%s", name, addr);
594 } else {
595 (void) snprintf(devname, devnamelen, "%s", name);
596 }
597
598 ndi_devi_enter(parent, &circ);
599
600 dip = ndi_devi_findchild(parent, devname);
601 kmem_free(devname, devnamelen);
602
603 if (dip == NULL) {
604 ndi_devi_exit(parent, circ);
605 return (NDI_FAILURE);
606 }
607
608 mutex_enter(&(DEVI(dip)->devi_lock));
609 if (DEVI_IS_DEVICE_OFFLINE(dip)) {
610 *state = DEVICE_OFFLINE;
611 } else if (DEVI_IS_DEVICE_DOWN(dip)) {
612 *state = DEVICE_DOWN;
613 } else {
614 *state = DEVICE_ONLINE;
615 if (devi_stillreferenced(dip) == DEVI_REFERENCED)
616 *state |= DEVICE_BUSY;
617 }
618
619 mutex_exit(&(DEVI(dip)->devi_lock));
620 ndi_devi_exit(parent, circ);
621
622 return (NDI_SUCCESS);
623 }
624
625 /*
626 * return the current state of the device "dip"
627 *
628 * recommend using ndi_devctl_ioctl() or
629 * ndi_devctl_device_getstate() instead
630 */
631 int
ndi_dc_return_dev_state(dev_info_t * dip,struct devctl_iocdata * dcp)632 ndi_dc_return_dev_state(dev_info_t *dip, struct devctl_iocdata *dcp)
633 {
634 dev_info_t *pdip;
635 uint_t devstate = 0;
636 int circ;
637
638 if ((dip == NULL) || (dcp == NULL))
639 return (NDI_FAILURE);
640
641 pdip = ddi_get_parent(dip);
642
643 ndi_devi_enter(pdip, &circ);
644 mutex_enter(&(DEVI(dip)->devi_lock));
645 if (DEVI_IS_DEVICE_OFFLINE(dip)) {
646 devstate = DEVICE_OFFLINE;
647 } else if (DEVI_IS_DEVICE_DOWN(dip)) {
648 devstate = DEVICE_DOWN;
649 } else {
650 devstate = DEVICE_ONLINE;
651 if (devi_stillreferenced(dip) == DEVI_REFERENCED)
652 devstate |= DEVICE_BUSY;
653 }
654
655 mutex_exit(&(DEVI(dip)->devi_lock));
656 ndi_devi_exit(pdip, circ);
657
658 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0)
659 return (NDI_FAULT);
660
661 return (NDI_SUCCESS);
662 }
663
664 /*
665 * Return device's bus state
666 * For use by a driver's DEVCTL_BUS_GETSTATE handler.
667 */
668 int
ndi_devctl_bus_getstate(dev_info_t * dip,struct devctl_iocdata * dcp,uint_t * state)669 ndi_devctl_bus_getstate(dev_info_t *dip, struct devctl_iocdata *dcp,
670 uint_t *state)
671 {
672 if ((dip == NULL) || (dcp == NULL))
673 return (NDI_FAILURE);
674
675 return (ndi_get_bus_state(dip, state));
676 }
677
678 /*
679 * Generic devctl ioctl handler
680 */
681 int
ndi_devctl_ioctl(dev_info_t * dip,int cmd,intptr_t arg,int mode,uint_t flags)682 ndi_devctl_ioctl(dev_info_t *dip, int cmd, intptr_t arg, int mode, uint_t flags)
683 {
684 _NOTE(ARGUNUSED(mode))
685 struct devctl_iocdata *dcp;
686 uint_t state;
687 int rval = ENOTTY;
688
689 /*
690 * read devctl ioctl data
691 */
692 if (ndi_dc_allochdl((void *)arg, &dcp) != NDI_SUCCESS)
693 return (EFAULT);
694
695 switch (cmd) {
696
697 case DEVCTL_BUS_GETSTATE:
698 rval = ndi_devctl_bus_getstate(dip, dcp, &state);
699 if (rval == NDI_SUCCESS) {
700 if (copyout(&state, dcp->cpyout_buf,
701 sizeof (uint_t)) != 0)
702 rval = NDI_FAULT;
703 }
704 break;
705
706 case DEVCTL_DEVICE_ONLINE:
707 rval = ndi_devctl_device_online(dip, dcp, flags);
708 break;
709
710 case DEVCTL_DEVICE_OFFLINE:
711 rval = ndi_devctl_device_offline(dip, dcp, flags);
712 break;
713
714 case DEVCTL_DEVICE_GETSTATE:
715 rval = ndi_devctl_device_getstate(dip, dcp, &state);
716 if (rval == NDI_SUCCESS) {
717 if (copyout(&state, dcp->cpyout_buf,
718 sizeof (uint_t)) != 0)
719 rval = NDI_FAULT;
720 }
721 break;
722
723 case DEVCTL_DEVICE_REMOVE:
724 rval = ndi_devctl_device_remove(dip, dcp, flags);
725 break;
726
727 case DEVCTL_BUS_DEV_CREATE:
728 rval = ndi_dc_devi_create(dcp, dip, 0, NULL);
729 break;
730
731 /*
732 * ioctls for which a generic implementation makes no sense
733 */
734 case DEVCTL_BUS_RESET:
735 case DEVCTL_BUS_RESETALL:
736 case DEVCTL_DEVICE_RESET:
737 case DEVCTL_AP_CONNECT:
738 case DEVCTL_AP_DISCONNECT:
739 case DEVCTL_AP_INSERT:
740 case DEVCTL_AP_REMOVE:
741 case DEVCTL_AP_CONFIGURE:
742 case DEVCTL_AP_UNCONFIGURE:
743 case DEVCTL_AP_GETSTATE:
744 case DEVCTL_AP_CONTROL:
745 case DEVCTL_BUS_QUIESCE:
746 case DEVCTL_BUS_UNQUIESCE:
747 rval = ENOTSUP;
748 break;
749 }
750
751 ndi_dc_freehdl(dcp);
752 return (rval);
753 }
754
755 /*
756 * Copyout the state of the Attachment Point "ap" to the requesting
757 * user process.
758 */
759 int
ndi_dc_return_ap_state(devctl_ap_state_t * ap,struct devctl_iocdata * dcp)760 ndi_dc_return_ap_state(devctl_ap_state_t *ap, struct devctl_iocdata *dcp)
761 {
762 if ((ap == NULL) || (dcp == NULL))
763 return (NDI_FAILURE);
764
765
766 if (get_udatamodel() == DATAMODEL_NATIVE) {
767 if (copyout(ap, dcp->cpyout_buf,
768 sizeof (devctl_ap_state_t)) != 0)
769 return (NDI_FAULT);
770 }
771 #ifdef _SYSCALL32_IMPL
772 else {
773 struct devctl_ap_state32 ap_state32;
774
775 ap_state32.ap_rstate = ap->ap_rstate;
776 ap_state32.ap_ostate = ap->ap_ostate;
777 ap_state32.ap_condition = ap->ap_condition;
778 ap_state32.ap_error_code = ap->ap_error_code;
779 ap_state32.ap_in_transition = ap->ap_in_transition;
780 ap_state32.ap_last_change = (time32_t)ap->ap_last_change;
781 if (copyout(&ap_state32, dcp->cpyout_buf,
782 sizeof (devctl_ap_state32_t)) != 0)
783 return (NDI_FAULT);
784 }
785 #endif
786
787 return (NDI_SUCCESS);
788 }
789
790 /*
791 * Copyout the bus state of the bus nexus device "dip" to the requesting
792 * user process.
793 */
794 int
ndi_dc_return_bus_state(dev_info_t * dip,struct devctl_iocdata * dcp)795 ndi_dc_return_bus_state(dev_info_t *dip, struct devctl_iocdata *dcp)
796 {
797 uint_t devstate = 0;
798
799 if ((dip == NULL) || (dcp == NULL))
800 return (NDI_FAILURE);
801
802 if (ndi_get_bus_state(dip, &devstate) != NDI_SUCCESS)
803 return (NDI_FAILURE);
804
805 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0)
806 return (NDI_FAULT);
807
808 return (NDI_SUCCESS);
809 }
810
811 static int
812 i_dc_devi_create(struct devctl_iocdata *, dev_info_t *, dev_info_t **);
813
814 /*
815 * create a child device node given the property definitions
816 * supplied by the userland process
817 */
818 int
ndi_dc_devi_create(struct devctl_iocdata * dcp,dev_info_t * pdip,int flags,dev_info_t ** rdip)819 ndi_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip, int flags,
820 dev_info_t **rdip)
821 {
822 dev_info_t *cdip;
823 int rv, circular = 0;
824 char devnm[MAXNAMELEN];
825 int nmlen;
826
827 /*
828 * The child device may have been pre-constructed by an earlier
829 * call to this function with the flag DEVCTL_CONSTRUCT set.
830 */
831
832 if ((cdip = (rdip != NULL) ? *rdip : NULL) == NULL)
833 if ((rv = i_dc_devi_create(dcp, pdip, &cdip)) != 0)
834 return (rv);
835
836 ASSERT(cdip != NULL);
837
838 /*
839 * Return the device node partially constructed if the
840 * DEVCTL_CONSTRUCT flag is set.
841 */
842 if (flags & DEVCTL_CONSTRUCT) {
843 if (rdip == NULL) {
844 (void) ndi_devi_free(cdip);
845 return (EINVAL);
846 }
847 *rdip = cdip;
848 return (0);
849 }
850
851 /*
852 * Bring the node up to a named but OFFLINE state. The calling
853 * application will need to manage the node from here on.
854 */
855 if (dcp->flags & DEVCTL_OFFLINE) {
856 /*
857 * In the unlikely event that the dip was somehow attached by
858 * the userland process (and device contracts or LDI opens
859 * were registered against the dip) after it was created by
860 * a previous DEVCTL_CONSTRUCT call, we start notify
861 * proceedings on this dip. Note that we don't need to
862 * return the dip after a failure of the notify since
863 * for a contract or LDI handle to be created the dip was
864 * already available to the user.
865 */
866 if (e_ddi_offline_notify(cdip) == DDI_FAILURE) {
867 return (EBUSY);
868 }
869
870 /*
871 * hand set the OFFLINE flag to prevent any asynchronous
872 * autoconfiguration operations from attaching this node.
873 */
874 mutex_enter(&(DEVI(cdip)->devi_lock));
875 DEVI_SET_DEVICE_OFFLINE(cdip);
876 mutex_exit(&(DEVI(cdip)->devi_lock));
877
878 e_ddi_offline_finalize(cdip, DDI_SUCCESS);
879
880 rv = ndi_devi_bind_driver(cdip, flags);
881 if (rv != NDI_SUCCESS) {
882 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
883 return (ENXIO);
884 }
885
886 /*
887 * remove the dev_info node if it failed to bind to a
888 * driver above.
889 */
890 if (i_ddi_node_state(cdip) < DS_BOUND) {
891 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
892 return (ENXIO);
893 }
894
895 /*
896 * add the node to the per-driver list and INITCHILD it
897 * to give it a name.
898 */
899 ndi_devi_enter(pdip, &circular);
900 if ((rv = ddi_initchild(pdip, cdip)) != DDI_SUCCESS) {
901 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
902 ndi_devi_exit(pdip, circular);
903 return (EINVAL);
904 }
905 ndi_devi_exit(pdip, circular);
906
907 } else {
908 /*
909 * Attempt to bring the device ONLINE. If the request to
910 * fails, remove the dev_info node.
911 */
912 if (ndi_devi_online(cdip, NDI_ONLINE_ATTACH) != NDI_SUCCESS) {
913 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
914 return (ENXIO);
915 }
916
917 /*
918 * if the node was successfully added but there was
919 * no driver available for the device, remove the node
920 */
921 if (i_ddi_node_state(cdip) < DS_BOUND) {
922 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
923 return (ENODEV);
924 }
925 }
926
927 /*
928 * return a handle to the child device
929 * copy out the name of the newly attached child device if
930 * the IOCTL request has provided a copyout buffer.
931 */
932 if (rdip != NULL)
933 *rdip = cdip;
934
935 if (dcp->cpyout_buf == NULL)
936 return (0);
937
938 ASSERT(ddi_node_name(cdip) != NULL);
939 ASSERT(ddi_get_name_addr(cdip) != NULL);
940
941 nmlen = snprintf(devnm, MAXNAMELEN, "%s@%s",
942 ddi_node_name(cdip), ddi_get_name_addr(cdip));
943
944 if (copyout(&devnm, dcp->cpyout_buf, nmlen) != 0) {
945 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE);
946 return (EFAULT);
947 }
948 return (0);
949 }
950
951 static int
i_dc_devi_create(struct devctl_iocdata * dcp,dev_info_t * pdip,dev_info_t ** rdip)952 i_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip,
953 dev_info_t **rdip)
954 {
955
956 dev_info_t *cdip;
957 char *cname = NULL;
958 nvlist_t *nvlp = dcp->nvl_user;
959 nvpair_t *npp;
960 char *np;
961 int rv = 0;
962
963 ASSERT(rdip != NULL && *rdip == NULL);
964
965 if ((nvlp == NULL) ||
966 (nvlist_lookup_string(nvlp, DC_DEVI_NODENAME, &cname) != 0))
967 return (EINVAL);
968
969 /*
970 * construct a new dev_info node with a user-provided nodename
971 */
972 ndi_devi_alloc_sleep(pdip, cname, (pnode_t)DEVI_SID_NODEID, &cdip);
973
974 /*
975 * create hardware properties for each member in the property
976 * list.
977 */
978 for (npp = nvlist_next_nvpair(nvlp, NULL); (npp != NULL && !rv);
979 npp = nvlist_next_nvpair(nvlp, npp)) {
980
981 np = nvpair_name(npp);
982
983 /*
984 * skip the nodename property
985 */
986 if (strcmp(np, DC_DEVI_NODENAME) == 0)
987 continue;
988
989 switch (nvpair_type(npp)) {
990
991 case DATA_TYPE_INT32: {
992 int32_t prop_val;
993
994 if ((rv = nvpair_value_int32(npp, &prop_val)) != 0)
995 break;
996
997 (void) ndi_prop_update_int(DDI_DEV_T_NONE, cdip, np,
998 (int)prop_val);
999 break;
1000 }
1001
1002 case DATA_TYPE_STRING: {
1003 char *prop_val;
1004
1005 if ((rv = nvpair_value_string(npp, &prop_val)) != 0)
1006 break;
1007
1008 (void) ndi_prop_update_string(DDI_DEV_T_NONE, cdip,
1009 np, prop_val);
1010 break;
1011 }
1012
1013 case DATA_TYPE_BYTE_ARRAY: {
1014 uchar_t *val;
1015 uint_t nelms;
1016
1017 if ((rv = nvpair_value_byte_array(npp, &val,
1018 &nelms)) != 0)
1019 break;
1020
1021 (void) ndi_prop_update_byte_array(DDI_DEV_T_NONE,
1022 cdip, np, (uchar_t *)val, nelms);
1023 break;
1024 }
1025
1026 case DATA_TYPE_INT32_ARRAY: {
1027 int32_t *val;
1028 uint_t nelms;
1029
1030 if ((rv = nvpair_value_int32_array(npp, &val,
1031 &nelms)) != 0)
1032 break;
1033
1034 (void) ndi_prop_update_int_array(DDI_DEV_T_NONE,
1035 cdip, np, val, nelms);
1036 break;
1037 }
1038
1039 case DATA_TYPE_STRING_ARRAY: {
1040 char **val;
1041 uint_t nelms;
1042
1043 if ((rv = nvpair_value_string_array(npp, &val,
1044 &nelms)) != 0)
1045 break;
1046
1047 (void) ndi_prop_update_string_array(DDI_DEV_T_NONE,
1048 cdip, np, val, nelms);
1049 break;
1050 }
1051
1052 /*
1053 * unsupported property data type
1054 */
1055 default:
1056 rv = EINVAL;
1057 }
1058 }
1059
1060 /*
1061 * something above failed
1062 * destroy the partially child device and abort the request
1063 */
1064 if (rv != 0) {
1065 (void) ndi_devi_free(cdip);
1066 return (rv);
1067 }
1068
1069 *rdip = cdip;
1070 return (0);
1071 }
1072
1073 /*
1074 * return current soft bus state of bus nexus "dip"
1075 */
1076 int
ndi_get_bus_state(dev_info_t * dip,uint_t * rstate)1077 ndi_get_bus_state(dev_info_t *dip, uint_t *rstate)
1078 {
1079 if (dip == NULL || rstate == NULL)
1080 return (NDI_FAILURE);
1081
1082 if (DEVI(dip)->devi_ops->devo_bus_ops == NULL)
1083 return (NDI_FAILURE);
1084
1085 mutex_enter(&(DEVI(dip)->devi_lock));
1086 if (DEVI_IS_BUS_QUIESCED(dip))
1087 *rstate = BUS_QUIESCED;
1088 else if (DEVI_IS_BUS_DOWN(dip))
1089 *rstate = BUS_SHUTDOWN;
1090 else
1091 *rstate = BUS_ACTIVE;
1092 mutex_exit(&(DEVI(dip)->devi_lock));
1093 return (NDI_SUCCESS);
1094 }
1095
1096 /*
1097 * Set the soft state of bus nexus "dip"
1098 */
1099 int
ndi_set_bus_state(dev_info_t * dip,uint_t state)1100 ndi_set_bus_state(dev_info_t *dip, uint_t state)
1101 {
1102 int rv = NDI_SUCCESS;
1103
1104 if (dip == NULL)
1105 return (NDI_FAILURE);
1106
1107 mutex_enter(&(DEVI(dip)->devi_lock));
1108
1109 switch (state) {
1110 case BUS_QUIESCED:
1111 DEVI_SET_BUS_QUIESCE(dip);
1112 break;
1113
1114 case BUS_ACTIVE:
1115 DEVI_SET_BUS_ACTIVE(dip);
1116 DEVI_SET_BUS_UP(dip);
1117 break;
1118
1119 case BUS_SHUTDOWN:
1120 DEVI_SET_BUS_DOWN(dip);
1121 break;
1122
1123 default:
1124 rv = NDI_FAILURE;
1125 }
1126
1127 mutex_exit(&(DEVI(dip)->devi_lock));
1128 return (rv);
1129 }
1130
1131 /*
1132 * These dummy functions are obsolete and may be removed.
1133 * Retained for existing driver compatibility only.
1134 * Drivers should be fixed not to use these functions.
1135 * Don't write new code using these obsolete interfaces.
1136 */
1137 /*ARGSUSED*/
1138 void
i_ndi_block_device_tree_changes(uint_t * lkcnt)1139 i_ndi_block_device_tree_changes(uint_t *lkcnt) /* obsolete */
1140 {
1141 /* obsolete dummy function */
1142 }
1143
1144 /*ARGSUSED*/
1145 void
i_ndi_allow_device_tree_changes(uint_t lkcnt)1146 i_ndi_allow_device_tree_changes(uint_t lkcnt) /* obsolete */
1147 {
1148 /* obsolete dummy function */
1149 }
1150
1151 /*
1152 * Single thread entry into per-driver list
1153 */
1154 /*ARGSUSED*/
1155 void
e_ddi_enter_driver_list(struct devnames * dnp,int * listcnt)1156 e_ddi_enter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */
1157 {
1158 /* obsolete dummy function */
1159 }
1160
1161 /*
1162 * release the per-driver list
1163 */
1164 /*ARGSUSED*/
1165 void
e_ddi_exit_driver_list(struct devnames * dnp,int listcnt)1166 e_ddi_exit_driver_list(struct devnames *dnp, int listcnt) /* obsolete */
1167 {
1168 /* obsolete dummy function */
1169 }
1170
1171 /*
1172 * Attempt to enter driver list
1173 */
1174 /*ARGSUSED*/
1175 int
e_ddi_tryenter_driver_list(struct devnames * dnp,int * listcnt)1176 e_ddi_tryenter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */
1177 {
1178 return (1); /* obsolete dummy function */
1179 }
1180
1181 /*
1182 * ndi event handling support functions:
1183 * The NDI event support model is as follows:
1184 *
1185 * The nexus driver defines a set of events using some static structures (so
1186 * these structures can be shared by all instances of the nexus driver).
1187 * The nexus driver allocates an event handle and binds the event set
1188 * to this handle. The nexus driver's event busop functions can just
1189 * call the appropriate NDI event support function using this handle
1190 * as the first argument.
1191 *
1192 * The reasoning for tying events to the device tree is that the entity
1193 * generating the callback will typically be one of the device driver's
1194 * ancestors in the tree.
1195 */
1196 static int ndi_event_debug = 0;
1197
1198 #ifdef DEBUG
1199 #define NDI_EVENT_DEBUG ndi_event_debug
1200 #endif /* DEBUG */
1201
1202 /*
1203 * allocate a new ndi event handle
1204 */
1205 int
ndi_event_alloc_hdl(dev_info_t * dip,ddi_iblock_cookie_t cookie,ndi_event_hdl_t * handle,uint_t flag)1206 ndi_event_alloc_hdl(dev_info_t *dip, ddi_iblock_cookie_t cookie,
1207 ndi_event_hdl_t *handle, uint_t flag)
1208 {
1209 struct ndi_event_hdl *ndi_event_hdl;
1210
1211 ndi_event_hdl = kmem_zalloc(sizeof (struct ndi_event_hdl),
1212 ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP));
1213
1214 if (!ndi_event_hdl) {
1215 return (NDI_FAILURE);
1216 }
1217
1218 ndi_event_hdl->ndi_evthdl_dip = dip;
1219 ndi_event_hdl->ndi_evthdl_iblock_cookie = cookie;
1220 mutex_init(&ndi_event_hdl->ndi_evthdl_mutex, NULL,
1221 MUTEX_DRIVER, (void *)cookie);
1222
1223 mutex_init(&ndi_event_hdl->ndi_evthdl_cb_mutex, NULL,
1224 MUTEX_DRIVER, (void *)cookie);
1225
1226 *handle = (ndi_event_hdl_t)ndi_event_hdl;
1227
1228 return (NDI_SUCCESS);
1229 }
1230
1231 /*
1232 * free the ndi event handle
1233 */
1234 int
ndi_event_free_hdl(ndi_event_hdl_t handle)1235 ndi_event_free_hdl(ndi_event_hdl_t handle)
1236 {
1237 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1238 ndi_event_cookie_t *cookie;
1239 ndi_event_cookie_t *free;
1240
1241 ASSERT(handle);
1242
1243 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1244 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1245
1246 cookie = ndi_event_hdl->ndi_evthdl_cookie_list;
1247
1248 /* deallocate all defined cookies */
1249 while (cookie != NULL) {
1250 ASSERT(cookie->callback_list == NULL);
1251 free = cookie;
1252 cookie = cookie->next_cookie;
1253
1254 kmem_free(free, sizeof (ndi_event_cookie_t));
1255 }
1256
1257
1258 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1259 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1260
1261 /* destroy mutexes */
1262 mutex_destroy(&ndi_event_hdl->ndi_evthdl_mutex);
1263 mutex_destroy(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1264
1265 /* free event handle */
1266 kmem_free(ndi_event_hdl, sizeof (struct ndi_event_hdl));
1267
1268 return (NDI_SUCCESS);
1269 }
1270
1271
1272 /*
1273 * ndi_event_bind_set() adds a set of events to the NDI event
1274 * handle.
1275 *
1276 * Events generated by high level interrupts should not
1277 * be mixed in the same event set with events generated by
1278 * normal interrupts or kernel events.
1279 *
1280 * This function can be called multiple times to bind
1281 * additional sets to the event handle.
1282 * However, events generated by high level interrupts cannot
1283 * be bound to a handle that already has bound events generated
1284 * by normal interrupts or from kernel context and vice versa.
1285 */
1286 int
ndi_event_bind_set(ndi_event_hdl_t handle,ndi_event_set_t * ndi_events,uint_t flag)1287 ndi_event_bind_set(ndi_event_hdl_t handle,
1288 ndi_event_set_t *ndi_events,
1289 uint_t flag)
1290 {
1291 struct ndi_event_hdl *ndi_event_hdl;
1292 ndi_event_cookie_t *next, *prev, *new_cookie;
1293 uint_t i, len;
1294 uint_t dup = 0;
1295 uint_t high_plevels, other_plevels;
1296 ndi_event_definition_t *ndi_event_defs;
1297
1298 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP);
1299
1300 ASSERT(handle);
1301 ASSERT(ndi_events);
1302
1303 /*
1304 * binding must be performed during attach/detach
1305 */
1306 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) &&
1307 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) {
1308 cmn_err(CE_WARN, "ndi_event_bind_set must be called within "
1309 "attach or detach");
1310 return (NDI_FAILURE);
1311 }
1312
1313 /*
1314 * if it is not the correct version or the event set is
1315 * empty, bail out
1316 */
1317 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1)
1318 return (NDI_FAILURE);
1319
1320 ndi_event_hdl = (struct ndi_event_hdl *)handle;
1321 ndi_event_defs = ndi_events->ndi_event_defs;
1322 high_plevels = other_plevels = 0;
1323
1324 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1325
1326 /* check for mixing events at high level with the other types */
1327 for (i = 0; i < ndi_events->ndi_n_events; i++) {
1328 if (ndi_event_defs[i].ndi_event_plevel == EPL_HIGHLEVEL) {
1329 high_plevels++;
1330 } else {
1331 other_plevels++;
1332 }
1333 }
1334
1335 /*
1336 * bail out if high level events are mixed with other types in this
1337 * event set or the set is incompatible with the set in the handle
1338 */
1339 if ((high_plevels && other_plevels) ||
1340 (other_plevels && ndi_event_hdl->ndi_evthdl_high_plevels) ||
1341 (high_plevels && ndi_event_hdl->ndi_evthdl_other_plevels)) {
1342 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1343
1344 return (NDI_FAILURE);
1345 }
1346
1347 /*
1348 * check for duplicate events in both the existing handle
1349 * and the event set, add events if not duplicates
1350 */
1351 next = ndi_event_hdl->ndi_evthdl_cookie_list;
1352 for (i = 0; i < ndi_events->ndi_n_events; i++) {
1353 while (next != NULL) {
1354 len = strlen(NDI_EVENT_NAME(next)) + 1;
1355 if (strncmp(NDI_EVENT_NAME(next),
1356 ndi_event_defs[i].ndi_event_name, len) == 0) {
1357 dup = 1;
1358 break;
1359 }
1360
1361 prev = next;
1362 next = next->next_cookie;
1363 }
1364
1365 if (dup == 0) {
1366 new_cookie = kmem_zalloc(sizeof (ndi_event_cookie_t),
1367 km_flag);
1368
1369 if (!new_cookie)
1370 return (NDI_FAILURE);
1371
1372 if (ndi_event_hdl->ndi_evthdl_n_events == 0) {
1373 ndi_event_hdl->ndi_evthdl_cookie_list =
1374 new_cookie;
1375 } else {
1376 prev->next_cookie = new_cookie;
1377 }
1378
1379 ndi_event_hdl->ndi_evthdl_n_events++;
1380
1381 /*
1382 * set up new cookie
1383 */
1384 new_cookie->definition = &ndi_event_defs[i];
1385 new_cookie->ddip = ndi_event_hdl->ndi_evthdl_dip;
1386
1387 } else {
1388 /*
1389 * event not added, must correct plevel numbers
1390 */
1391 if (ndi_event_defs[i].ndi_event_plevel ==
1392 EPL_HIGHLEVEL) {
1393 high_plevels--;
1394 } else {
1395 other_plevels--;
1396 }
1397 }
1398
1399 dup = 0;
1400 next = ndi_event_hdl->ndi_evthdl_cookie_list;
1401 prev = NULL;
1402
1403 }
1404
1405 ndi_event_hdl->ndi_evthdl_high_plevels += high_plevels;
1406 ndi_event_hdl->ndi_evthdl_other_plevels += other_plevels;
1407
1408 ASSERT((ndi_event_hdl->ndi_evthdl_high_plevels == 0) ||
1409 (ndi_event_hdl->ndi_evthdl_other_plevels == 0));
1410
1411 #ifdef NDI_EVENT_DEBUG
1412 if (ndi_event_debug) {
1413 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_bind_set");
1414 }
1415 #endif /* NDI_EVENT_DEBUG */
1416
1417 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1418
1419 return (NDI_SUCCESS);
1420 }
1421
1422 /*
1423 * ndi_event_unbind_set() unbinds a set of events, previously
1424 * bound using ndi_event_bind_set(), from the NDI event
1425 * handle.
1426 *
1427 * This routine will unbind all events in the event set. If an event,
1428 * specified in the event set, is not found in the handle, this
1429 * routine will proceed onto the next member of the set as if the event
1430 * was never specified.
1431 *
1432 * The event set may be a subset of the set of events that
1433 * was previously bound to the handle. For example, events
1434 * can be individually unbound.
1435 *
1436 * An event cannot be unbound if callbacks are still
1437 * registered against the event.
1438 */
1439 /*ARGSUSED*/
1440 int
ndi_event_unbind_set(ndi_event_hdl_t handle,ndi_event_set_t * ndi_events,uint_t flag)1441 ndi_event_unbind_set(ndi_event_hdl_t handle, ndi_event_set_t *ndi_events,
1442 uint_t flag)
1443 {
1444 ndi_event_definition_t *ndi_event_defs;
1445 int len;
1446 uint_t i;
1447 int rval;
1448 ndi_event_cookie_t *cookie_list;
1449 ndi_event_cookie_t *prev = NULL;
1450
1451 ASSERT(ndi_events);
1452 ASSERT(handle);
1453
1454 /*
1455 * binding must be performed during attach/detac
1456 */
1457 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) &&
1458 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) {
1459 cmn_err(CE_WARN, "ndi_event_bind_set must be called within "
1460 "attach or detach");
1461 return (NDI_FAILURE);
1462 }
1463
1464 /* bail out if ndi_event_set is outdated */
1465 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1) {
1466 return (NDI_FAILURE);
1467 }
1468
1469 ASSERT(ndi_events->ndi_event_defs);
1470
1471 ndi_event_defs = ndi_events->ndi_event_defs;
1472
1473 mutex_enter(&handle->ndi_evthdl_mutex);
1474 mutex_enter(&handle->ndi_evthdl_cb_mutex);
1475
1476 /*
1477 * Verify that all events in the event set are eligible
1478 * for unbinding(ie. there are no outstanding callbacks).
1479 * If any one of the events are ineligible, fail entire
1480 * operation.
1481 */
1482
1483 for (i = 0; i < ndi_events->ndi_n_events; i++) {
1484 cookie_list = handle->ndi_evthdl_cookie_list;
1485 while (cookie_list != NULL) {
1486 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1;
1487 if (strncmp(NDI_EVENT_NAME(cookie_list),
1488 ndi_event_defs[i].ndi_event_name, len) == 0) {
1489
1490 ASSERT(cookie_list->callback_list == NULL);
1491 if (cookie_list->callback_list) {
1492 rval = NDI_FAILURE;
1493 goto done;
1494 }
1495 break;
1496 } else {
1497 cookie_list = cookie_list->next_cookie;
1498 }
1499 }
1500 }
1501
1502 /*
1503 * remove all events found within the handle
1504 * If an event is not found, this function will proceed as if the event
1505 * was never specified.
1506 */
1507
1508 for (i = 0; i < ndi_events->ndi_n_events; i++) {
1509 cookie_list = handle->ndi_evthdl_cookie_list;
1510 prev = NULL;
1511 while (cookie_list != NULL) {
1512 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1;
1513 if (strncmp(NDI_EVENT_NAME(cookie_list),
1514 ndi_event_defs[i].ndi_event_name, len) == 0) {
1515
1516 /*
1517 * can not unbind an event definition with
1518 * outstanding callbacks
1519 */
1520 if (cookie_list->callback_list) {
1521 rval = NDI_FAILURE;
1522 goto done;
1523 }
1524
1525 /* remove this cookie from the list */
1526 if (prev != NULL) {
1527 prev->next_cookie =
1528 cookie_list->next_cookie;
1529 } else {
1530 handle->ndi_evthdl_cookie_list =
1531 cookie_list->next_cookie;
1532 }
1533
1534 /* adjust plevel counts */
1535 if (NDI_EVENT_PLEVEL(cookie_list) ==
1536 EPL_HIGHLEVEL) {
1537 handle->ndi_evthdl_high_plevels--;
1538 } else {
1539 handle->ndi_evthdl_other_plevels--;
1540 }
1541
1542 /* adjust cookie count */
1543 handle->ndi_evthdl_n_events--;
1544
1545 /* free the cookie */
1546 kmem_free(cookie_list,
1547 sizeof (ndi_event_cookie_t));
1548
1549 cookie_list = handle->ndi_evthdl_cookie_list;
1550 break;
1551
1552 } else {
1553 prev = cookie_list;
1554 cookie_list = cookie_list->next_cookie;
1555 }
1556
1557 }
1558
1559 }
1560
1561 #ifdef NDI_EVENT_DEBUG
1562 if (ndi_event_debug) {
1563 ndi_event_dump_hdl(handle, "ndi_event_unbind_set");
1564 }
1565 #endif /* NDI_EVENT_DEBUG */
1566
1567 rval = NDI_SUCCESS;
1568
1569 done:
1570 mutex_exit(&handle->ndi_evthdl_cb_mutex);
1571 mutex_exit(&handle->ndi_evthdl_mutex);
1572
1573 return (rval);
1574 }
1575
1576 /*
1577 * ndi_event_retrieve_cookie():
1578 * Return an event cookie for eventname if this nexus driver
1579 * has defined the named event. The event cookie returned
1580 * by this function is used to register callback handlers
1581 * for the event.
1582 *
1583 * ndi_event_retrieve_cookie() is intended to be used in the
1584 * nexus driver's bus_get_eventcookie busop routine.
1585 *
1586 * If the event is not defined by this bus nexus driver, and flag
1587 * does not include NDI_EVENT_NOPASS, then ndi_event_retrieve_cookie()
1588 * will pass the request up the device tree hierarchy by calling
1589 * ndi_busop_get_eventcookie(9N).
1590 * If the event is not defined by this bus nexus driver, and flag
1591 * does include NDI_EVENT_NOPASS, ndi_event_retrieve_cookie()
1592 * will return NDI_FAILURE. The caller may then determine what further
1593 * action to take, such as using a different handle, passing the
1594 * request up the device tree using ndi_busop_get_eventcookie(9N),
1595 * or returning the failure to the caller, thus blocking the
1596 * progress of the request up the tree.
1597 */
1598 int
ndi_event_retrieve_cookie(ndi_event_hdl_t handle,dev_info_t * rdip,char * eventname,ddi_eventcookie_t * cookiep,uint_t flag)1599 ndi_event_retrieve_cookie(ndi_event_hdl_t handle,
1600 dev_info_t *rdip,
1601 char *eventname,
1602 ddi_eventcookie_t *cookiep,
1603 uint_t flag)
1604 {
1605 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1606 int len;
1607 ndi_event_cookie_t *cookie_list;
1608
1609 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1610
1611 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list;
1612 /*
1613 * search the cookie list for the event name and return
1614 * cookie if found.
1615 */
1616 while (cookie_list != NULL) {
1617
1618 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1;
1619 if (strncmp(NDI_EVENT_NAME(cookie_list), eventname,
1620 len) == 0) {
1621 *cookiep = (ddi_eventcookie_t)cookie_list;
1622
1623 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1624 return (NDI_SUCCESS);
1625 }
1626
1627 cookie_list = cookie_list->next_cookie;
1628 }
1629
1630 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1631 /*
1632 * event was not found, pass up or return failure
1633 */
1634 if ((flag & NDI_EVENT_NOPASS) == 0) {
1635 return (ndi_busop_get_eventcookie(
1636 ndi_event_hdl->ndi_evthdl_dip, rdip, eventname, cookiep));
1637 } else {
1638 return (NDI_FAILURE);
1639 }
1640 }
1641
1642 /*
1643 * check whether this nexus defined this event and look up attributes
1644 */
1645 static int
ndi_event_is_defined(ndi_event_hdl_t handle,ddi_eventcookie_t cookie,int * attributes)1646 ndi_event_is_defined(ndi_event_hdl_t handle,
1647 ddi_eventcookie_t cookie, int *attributes)
1648 {
1649
1650 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1651 ndi_event_cookie_t *cookie_list;
1652
1653 ASSERT(mutex_owned(&handle->ndi_evthdl_mutex));
1654
1655 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list;
1656 while (cookie_list != NULL) {
1657 if (cookie_list == NDI_EVENT(cookie)) {
1658 if (attributes)
1659 *attributes =
1660 NDI_EVENT_ATTRIBUTES(cookie_list);
1661
1662 return (NDI_SUCCESS);
1663 }
1664
1665 cookie_list = cookie_list->next_cookie;
1666 }
1667
1668 return (NDI_FAILURE);
1669 }
1670
1671 /*
1672 * ndi_event_add_callback(): adds an event callback registration
1673 * to the event cookie defining this event.
1674 *
1675 * Refer also to bus_add_eventcall(9n) and ndi_busop_add_eventcall(9n).
1676 *
1677 * ndi_event_add_callback(9n) is intended to be used in
1678 * the nexus driver's bus_add_eventcall(9n) busop function.
1679 *
1680 * If the event is not defined by this bus nexus driver,
1681 * ndi_event_add_callback() will return NDI_FAILURE.
1682 */
1683 int
ndi_event_add_callback(ndi_event_hdl_t handle,dev_info_t * child_dip,ddi_eventcookie_t cookie,void (* event_callback)(dev_info_t *,ddi_eventcookie_t,void * arg,void * impldata),void * arg,uint_t flag,ddi_callback_id_t * cb_id)1684 ndi_event_add_callback(ndi_event_hdl_t handle, dev_info_t *child_dip,
1685 ddi_eventcookie_t cookie,
1686 void (*event_callback)(dev_info_t *,
1687 ddi_eventcookie_t, void *arg, void *impldata),
1688 void *arg,
1689 uint_t flag,
1690 ddi_callback_id_t *cb_id)
1691 {
1692 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1693 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP);
1694 ndi_event_callbacks_t *cb;
1695
1696 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1697
1698 /*
1699 * if the event was not bound to this handle, return failure
1700 */
1701 if (ndi_event_is_defined(handle, cookie, NULL) != NDI_SUCCESS) {
1702
1703 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1704 return (NDI_FAILURE);
1705
1706 }
1707
1708 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1709
1710 /*
1711 * allocate space for a callback structure
1712 */
1713 cb = kmem_zalloc(sizeof (ndi_event_callbacks_t), km_flag);
1714 if (cb == NULL) {
1715 return (NDI_FAILURE);
1716 }
1717
1718 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1719
1720 /* initialize callback structure */
1721 cb->ndi_evtcb_dip = child_dip;
1722 cb->ndi_evtcb_callback = event_callback;
1723 cb->ndi_evtcb_arg = arg;
1724 cb->ndi_evtcb_cookie = cookie;
1725 cb->devname = (char *)ddi_driver_name(child_dip);
1726
1727 *cb_id = (ddi_callback_id_t)cb;
1728 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1729
1730 /* add this callback structure to the list */
1731 if (NDI_EVENT(cookie)->callback_list) {
1732 cb->ndi_evtcb_next = NDI_EVENT(cookie)->callback_list;
1733 NDI_EVENT(cookie)->callback_list->ndi_evtcb_prev = cb;
1734 NDI_EVENT(cookie)->callback_list = cb;
1735 } else {
1736 NDI_EVENT(cookie)->callback_list = cb;
1737 }
1738 #ifdef NDI_EVENT_DEBUG
1739 if (ndi_event_debug) {
1740 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_add_callback");
1741 }
1742 #endif /* NDI_EVENT_DEBUG */
1743
1744 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1745 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1746
1747 return (NDI_SUCCESS);
1748 }
1749
1750 /*
1751 * ndi_event_remove_callback():
1752 *
1753 * ndi_event_remove_callback() removes a callback that was
1754 * previously registered using ndi_event_add_callback(9N).
1755 * Refer also to bus_remove_eventcall(9n) and
1756 * ndi_busop_remove_eventcall(9n).
1757 * ndi_event_remove_callback(9n) is intended to be used in
1758 * the nexus driver's bus_remove_eventcall (9n) busop function.
1759 * If the event is not defined by this bus nexus driver,
1760 * ndi_event_remove_callback() will return NDI_FAILURE.
1761 */
1762 static void do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl,
1763 ddi_callback_id_t cb_id);
1764
1765 int
ndi_event_remove_callback(ndi_event_hdl_t handle,ddi_callback_id_t cb_id)1766 ndi_event_remove_callback(ndi_event_hdl_t handle, ddi_callback_id_t cb_id)
1767 {
1768 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1769
1770 ASSERT(cb_id);
1771
1772 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1773 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1774
1775 do_ndi_event_remove_callback(ndi_event_hdl, cb_id);
1776
1777 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1778 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1779
1780 return (NDI_SUCCESS);
1781 }
1782
1783 /*ARGSUSED*/
1784 static void
do_ndi_event_remove_callback(struct ndi_event_hdl * ndi_event_hdl,ddi_callback_id_t cb_id)1785 do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl,
1786 ddi_callback_id_t cb_id)
1787 {
1788 ndi_event_callbacks_t *cb = (ndi_event_callbacks_t *)cb_id;
1789 ASSERT(cb);
1790
1791 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_mutex));
1792 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_cb_mutex));
1793
1794 /* remove from callback linked list */
1795 if (cb->ndi_evtcb_prev) {
1796 cb->ndi_evtcb_prev->ndi_evtcb_next = cb->ndi_evtcb_next;
1797 }
1798
1799 if (cb->ndi_evtcb_next) {
1800 cb->ndi_evtcb_next->ndi_evtcb_prev = cb->ndi_evtcb_prev;
1801 }
1802
1803 if (NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list == cb) {
1804 NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list =
1805 cb->ndi_evtcb_next;
1806 }
1807
1808 kmem_free(cb, sizeof (ndi_event_callbacks_t));
1809 }
1810
1811 /*
1812 * ndi_event_run_callbacks() performs event callbacks for the event
1813 * specified by cookie, if this is among those bound to the
1814 * supplied handle.
1815 * If the event is among those bound to the handle, none,
1816 * some, or all of the handlers registered for the event
1817 * will be called, according to the delivery attributes of
1818 * the event.
1819 * If the event attributes include NDI_EVENT_POST_TO_ALL
1820 * (the default), all the handlers for the event will be
1821 * called in an unspecified order.
1822 * If the event attributes include NDI_EVENT_POST_TO_TGT, only
1823 * the handlers (if any) registered by the driver identified by
1824 * rdip will be called.
1825 * If the event identified by cookie is not bound to the handle,
1826 * NDI_FAILURE will be returned.
1827 */
1828 int
ndi_event_run_callbacks(ndi_event_hdl_t handle,dev_info_t * child_dip,ddi_eventcookie_t cookie,void * bus_impldata)1829 ndi_event_run_callbacks(ndi_event_hdl_t handle, dev_info_t *child_dip,
1830 ddi_eventcookie_t cookie, void *bus_impldata)
1831 {
1832 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1833 ndi_event_callbacks_t *next, *cb;
1834 int attributes;
1835
1836 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1837
1838 /* if this is not our event, fail */
1839 if (ndi_event_is_defined(handle, cookie, &attributes) !=
1840 NDI_SUCCESS) {
1841
1842 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1843 return (NDI_FAILURE);
1844 }
1845
1846 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1847
1848 #ifdef NDI_EVENT_DEBUG
1849 if (ndi_event_debug) {
1850 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t"
1851 "producer dip=%p (%s%d): cookie = %p, name = %s\n",
1852 (void *)ndi_event_hdl->ndi_evthdl_dip,
1853 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip),
1854 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip),
1855 (void *)cookie,
1856 ndi_event_cookie_to_name(handle, cookie));
1857 }
1858 #endif /* #ifdef NDI_EVENT_DEBUG */
1859
1860
1861 /*
1862 * The callback handlers may call conversion functions. The conversion
1863 * functions may hold the ndi_evthdl_mutex during execution. Thus, to
1864 * avoid a recursive mutex problem, only the ndi_evthdl_cb_mutex is
1865 * held. The ndi_evthdl_mutex is not held when running the callbacks.
1866 */
1867 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1868
1869 /* perform callbacks */
1870 next = NDI_EVENT(cookie)->callback_list;
1871 while (next != NULL) {
1872
1873 cb = next;
1874 next = next->ndi_evtcb_next;
1875
1876 ASSERT(cb->ndi_evtcb_cookie == cookie);
1877
1878 if (attributes == NDI_EVENT_POST_TO_TGT &&
1879 child_dip != cb->ndi_evtcb_dip) {
1880 continue;
1881 }
1882
1883 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip, cb->ndi_evtcb_cookie,
1884 cb->ndi_evtcb_arg, bus_impldata);
1885
1886 #ifdef NDI_EVENT_DEBUG
1887 if (ndi_event_debug) {
1888 cmn_err(CE_CONT,
1889 "\t\tconsumer dip=%p (%s%d)\n",
1890 (void *)cb->ndi_evtcb_dip,
1891 ddi_node_name(cb->ndi_evtcb_dip),
1892 ddi_get_instance(cb->ndi_evtcb_dip));
1893 }
1894 #endif
1895
1896 }
1897
1898 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1899
1900 #ifdef NDI_EVENT_DEBUG
1901 if (ndi_event_debug) {
1902 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1903 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks");
1904 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1905 }
1906 #endif /* NDI_EVENT_DEBUG */
1907
1908 return (NDI_SUCCESS);
1909 }
1910
1911
1912 /*
1913 * perform one callback for a specified cookie and just one target
1914 */
1915 int
ndi_event_do_callback(ndi_event_hdl_t handle,dev_info_t * child_dip,ddi_eventcookie_t cookie,void * bus_impldata)1916 ndi_event_do_callback(ndi_event_hdl_t handle, dev_info_t *child_dip,
1917 ddi_eventcookie_t cookie, void *bus_impldata)
1918 {
1919 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
1920 ndi_event_callbacks_t *next, *cb;
1921 int attributes;
1922
1923 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1924
1925 /* if this is not our event, fail */
1926 if (ndi_event_is_defined(handle, cookie, &attributes) !=
1927 NDI_SUCCESS) {
1928
1929 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1930
1931 return (NDI_FAILURE);
1932 }
1933
1934 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1935
1936 #ifdef NDI_EVENT_DEBUG
1937 if (ndi_event_debug) {
1938 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t"
1939 "producer dip=%p (%s%d): cookie = %p, name = %s\n",
1940 (void *)ndi_event_hdl->ndi_evthdl_dip,
1941 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip),
1942 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip),
1943 (void *)cookie,
1944 ndi_event_cookie_to_name(handle, cookie));
1945 }
1946 #endif
1947
1948
1949 /*
1950 * we only grab the cb mutex because the callback handlers
1951 * may call the conversion functions which would cause a recursive
1952 * mutex problem
1953 */
1954 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1955
1956 /* perform callbacks */
1957 for (next = NDI_EVENT(cookie)->callback_list; next != NULL; ) {
1958 cb = next;
1959 next = next->ndi_evtcb_next;
1960
1961 if (cb->ndi_evtcb_dip == child_dip) {
1962 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip,
1963 cb->ndi_evtcb_cookie, cb->ndi_evtcb_arg,
1964 bus_impldata);
1965
1966 #ifdef NDI_EVENT_DEBUG
1967 if (ndi_event_debug) {
1968 cmn_err(CE_CONT,
1969 "\t\tconsumer dip=%p (%s%d)\n",
1970 (void *)cb->ndi_evtcb_dip,
1971 ddi_node_name(cb->ndi_evtcb_dip),
1972 ddi_get_instance(cb->ndi_evtcb_dip));
1973 }
1974 #endif
1975 break;
1976 }
1977 }
1978
1979 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex);
1980
1981 #ifdef NDI_EVENT_DEBUG
1982 if (ndi_event_debug) {
1983 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
1984 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks");
1985 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
1986 }
1987 #endif /* NDI_EVENT_DEBUG */
1988
1989 return (NDI_SUCCESS);
1990 }
1991
1992
1993 /*
1994 * ndi_event_tag_to_cookie: utility function to find an event cookie
1995 * given an event tag
1996 */
1997 ddi_eventcookie_t
ndi_event_tag_to_cookie(ndi_event_hdl_t handle,int event_tag)1998 ndi_event_tag_to_cookie(ndi_event_hdl_t handle, int event_tag)
1999 {
2000 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
2001 ndi_event_cookie_t *list;
2002
2003 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
2004
2005 list = ndi_event_hdl->ndi_evthdl_cookie_list;
2006 while (list != NULL) {
2007 if (NDI_EVENT_TAG(list) == event_tag) {
2008 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2009 return ((ddi_eventcookie_t)list);
2010 }
2011
2012 list = list->next_cookie;
2013 }
2014
2015 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2016 return (NULL);
2017 }
2018
2019 /*
2020 * ndi_event_cookie_to_tag: utility function to find a event tag
2021 * given an event_cookie
2022 */
2023 int
ndi_event_cookie_to_tag(ndi_event_hdl_t handle,ddi_eventcookie_t cookie)2024 ndi_event_cookie_to_tag(ndi_event_hdl_t handle, ddi_eventcookie_t cookie)
2025 {
2026 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
2027 ndi_event_cookie_t *list;
2028
2029 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
2030
2031 list = ndi_event_hdl->ndi_evthdl_cookie_list;
2032
2033 while (list != NULL) {
2034 if ((ddi_eventcookie_t)list == cookie) {
2035 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2036 return (NDI_EVENT_TAG(list));
2037 }
2038
2039 list = list->next_cookie;
2040 }
2041
2042 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2043 return (NDI_FAILURE);
2044
2045 }
2046
2047 /*
2048 * ndi_event_cookie_to_name: utility function to find an event name
2049 * given an event_cookie
2050 */
2051 char *
ndi_event_cookie_to_name(ndi_event_hdl_t handle,ddi_eventcookie_t cookie)2052 ndi_event_cookie_to_name(ndi_event_hdl_t handle, ddi_eventcookie_t cookie)
2053 {
2054 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
2055 ndi_event_cookie_t *list;
2056
2057 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
2058
2059 list = ndi_event_hdl->ndi_evthdl_cookie_list;
2060
2061 while (list != NULL) {
2062 if (list == NDI_EVENT(cookie)) {
2063 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2064 return (NDI_EVENT_NAME(list));
2065 }
2066
2067 list = list->next_cookie;
2068 }
2069
2070 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2071 return (NULL);
2072 }
2073
2074 /*
2075 * ndi_event_tag_to_name: utility function to find an event name
2076 * given an event tag
2077 */
2078 char *
ndi_event_tag_to_name(ndi_event_hdl_t handle,int event_tag)2079 ndi_event_tag_to_name(ndi_event_hdl_t handle, int event_tag)
2080 {
2081 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle;
2082 ndi_event_cookie_t *list;
2083
2084 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex);
2085
2086 list = ndi_event_hdl->ndi_evthdl_cookie_list;
2087
2088 while (list) {
2089 if (NDI_EVENT_TAG(list) == event_tag) {
2090 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2091 return (NDI_EVENT_NAME(list));
2092 }
2093
2094 list = list->next_cookie;
2095 }
2096
2097 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex);
2098
2099 return (NULL);
2100 }
2101
2102 #ifdef NDI_EVENT_DEBUG
2103 void
ndi_event_dump_hdl(struct ndi_event_hdl * hdl,char * location)2104 ndi_event_dump_hdl(struct ndi_event_hdl *hdl, char *location)
2105 {
2106
2107
2108 ndi_event_callbacks_t *next;
2109 ndi_event_cookie_t *list;
2110
2111 ASSERT(mutex_owned(&hdl->ndi_evthdl_mutex));
2112 list = hdl->ndi_evthdl_cookie_list;
2113
2114 cmn_err(CE_CONT, "%s: event handle (%p): dip = %p (%s%d)\n",
2115 location, (void *)hdl, (void *)hdl->ndi_evthdl_dip,
2116 ddi_node_name(hdl->ndi_evthdl_dip),
2117 ddi_get_instance(hdl->ndi_evthdl_dip));
2118 cmn_err(CE_CONT, "\thigh=%d other=%d n=%d\n",
2119 hdl->ndi_evthdl_high_plevels, hdl->ndi_evthdl_other_plevels,
2120 hdl->ndi_evthdl_n_events);
2121
2122 cmn_err(CE_CONT, "\tevent cookies:\n");
2123 while (list) {
2124 cmn_err(CE_CONT, "\t\ttag=%d name=%s p=%d a=%x dd=%p\n",
2125 NDI_EVENT_TAG(list), NDI_EVENT_NAME(list),
2126 NDI_EVENT_PLEVEL(list), NDI_EVENT_ATTRIBUTES(list),
2127 (void *)NDI_EVENT_DDIP(list));
2128 cmn_err(CE_CONT, "\t\tcallbacks:\n");
2129 for (next = list->callback_list; next != NULL;
2130 next = next->ndi_evtcb_next) {
2131 cmn_err(CE_CONT,
2132 "\t\t dip=%p (%s%d) cookie=%p arg=%p\n",
2133 (void*)next->ndi_evtcb_dip,
2134 ddi_driver_name(next->ndi_evtcb_dip),
2135 ddi_get_instance(next->ndi_evtcb_dip),
2136 (void *)next->ndi_evtcb_cookie,
2137 next->ndi_evtcb_arg);
2138 }
2139
2140 list = list->next_cookie;
2141 }
2142
2143 cmn_err(CE_CONT, "\n");
2144 }
2145 #endif
2146
2147 int
ndi_dev_is_prom_node(dev_info_t * dip)2148 ndi_dev_is_prom_node(dev_info_t *dip)
2149 {
2150 return (DEVI(dip)->devi_node_class == DDI_NC_PROM);
2151 }
2152
2153 int
ndi_dev_is_pseudo_node(dev_info_t * dip)2154 ndi_dev_is_pseudo_node(dev_info_t *dip)
2155 {
2156 /*
2157 * NOTE: this does NOT mean the pseudo branch of the device tree,
2158 * it means the node was created by software (DEVI_SID_NODEID ||
2159 * DEVI_PSEUDO_NODEID || DEVI_SID_HIDDEN_NODEID) instead of being
2160 * generated from a PROM node.
2161 */
2162 return (DEVI(dip)->devi_node_class == DDI_NC_PSEUDO);
2163 }
2164
2165 int
ndi_dev_is_persistent_node(dev_info_t * dip)2166 ndi_dev_is_persistent_node(dev_info_t *dip)
2167 {
2168 return ((DEVI(dip)->devi_node_attributes & DDI_PERSISTENT) != 0);
2169 }
2170
2171 int
ndi_dev_is_hidden_node(dev_info_t * dip)2172 ndi_dev_is_hidden_node(dev_info_t *dip)
2173 {
2174 return ((DEVI(dip)->devi_node_attributes & DDI_HIDDEN_NODE) != 0);
2175 }
2176
2177 int
ndi_dev_is_hotplug_node(dev_info_t * dip)2178 ndi_dev_is_hotplug_node(dev_info_t *dip)
2179 {
2180 return ((DEVI(dip)->devi_node_attributes & DDI_HOTPLUG_NODE) != 0);
2181 }
2182
2183 void
ndi_devi_set_hidden(dev_info_t * dip)2184 ndi_devi_set_hidden(dev_info_t *dip)
2185 {
2186 DEVI(dip)->devi_node_attributes |= DDI_HIDDEN_NODE;
2187 }
2188
2189 void
ndi_devi_clr_hidden(dev_info_t * dip)2190 ndi_devi_clr_hidden(dev_info_t *dip)
2191 {
2192 DEVI(dip)->devi_node_attributes &= ~DDI_HIDDEN_NODE;
2193 }
2194
2195 int
i_ndi_dev_is_auto_assigned_node(dev_info_t * dip)2196 i_ndi_dev_is_auto_assigned_node(dev_info_t *dip)
2197 {
2198 return ((DEVI(dip)->devi_node_attributes &
2199 DDI_AUTO_ASSIGNED_NODEID) != 0);
2200 }
2201
2202 void
i_ndi_set_node_class(dev_info_t * dip,ddi_node_class_t c)2203 i_ndi_set_node_class(dev_info_t *dip, ddi_node_class_t c)
2204 {
2205 DEVI(dip)->devi_node_class = c;
2206 }
2207
2208 ddi_node_class_t
i_ndi_get_node_class(dev_info_t * dip)2209 i_ndi_get_node_class(dev_info_t *dip)
2210 {
2211 return (DEVI(dip)->devi_node_class);
2212 }
2213
2214 void
i_ndi_set_node_attributes(dev_info_t * dip,int p)2215 i_ndi_set_node_attributes(dev_info_t *dip, int p)
2216 {
2217 DEVI(dip)->devi_node_attributes = p;
2218 }
2219
2220 int
i_ndi_get_node_attributes(dev_info_t * dip)2221 i_ndi_get_node_attributes(dev_info_t *dip)
2222 {
2223 return (DEVI(dip)->devi_node_attributes);
2224 }
2225
2226 void
i_ndi_set_nodeid(dev_info_t * dip,int n)2227 i_ndi_set_nodeid(dev_info_t *dip, int n)
2228 {
2229 DEVI(dip)->devi_nodeid = n;
2230 }
2231
2232 void
ndi_set_acc_fault(ddi_acc_handle_t ah)2233 ndi_set_acc_fault(ddi_acc_handle_t ah)
2234 {
2235 i_ddi_acc_set_fault(ah);
2236 }
2237
2238 void
ndi_clr_acc_fault(ddi_acc_handle_t ah)2239 ndi_clr_acc_fault(ddi_acc_handle_t ah)
2240 {
2241 i_ddi_acc_clr_fault(ah);
2242 }
2243
2244 void
ndi_set_dma_fault(ddi_dma_handle_t dh)2245 ndi_set_dma_fault(ddi_dma_handle_t dh)
2246 {
2247 i_ddi_dma_set_fault(dh);
2248 }
2249
2250 void
ndi_clr_dma_fault(ddi_dma_handle_t dh)2251 ndi_clr_dma_fault(ddi_dma_handle_t dh)
2252 {
2253 i_ddi_dma_clr_fault(dh);
2254 }
2255
2256 /*
2257 * The default fault-handler, called when the event posted by
2258 * ddi_dev_report_fault() reaches rootnex.
2259 */
2260 static void
i_ddi_fault_handler(dev_info_t * dip,struct ddi_fault_event_data * fedp)2261 i_ddi_fault_handler(dev_info_t *dip, struct ddi_fault_event_data *fedp)
2262 {
2263 ASSERT(fedp);
2264
2265 mutex_enter(&(DEVI(dip)->devi_lock));
2266 if (!DEVI_IS_DEVICE_OFFLINE(dip)) {
2267 switch (fedp->f_impact) {
2268 case DDI_SERVICE_LOST:
2269 DEVI_SET_DEVICE_DOWN(dip);
2270 break;
2271
2272 case DDI_SERVICE_DEGRADED:
2273 DEVI_SET_DEVICE_DEGRADED(dip);
2274 break;
2275
2276 case DDI_SERVICE_UNAFFECTED:
2277 default:
2278 break;
2279
2280 case DDI_SERVICE_RESTORED:
2281 DEVI_SET_DEVICE_UP(dip);
2282 break;
2283 }
2284 }
2285 mutex_exit(&(DEVI(dip)->devi_lock));
2286 }
2287
2288 /*
2289 * The default fault-logger, called when the event posted by
2290 * ddi_dev_report_fault() reaches rootnex.
2291 */
2292 /*ARGSUSED*/
2293 static void
i_ddi_fault_logger(dev_info_t * rdip,struct ddi_fault_event_data * fedp)2294 i_ddi_fault_logger(dev_info_t *rdip, struct ddi_fault_event_data *fedp)
2295 {
2296 ddi_devstate_t newstate;
2297 const char *action;
2298 const char *servstate;
2299 const char *location;
2300 int bad;
2301 int changed;
2302 int level;
2303 int still;
2304
2305 ASSERT(fedp);
2306
2307 bad = 0;
2308 switch (fedp->f_location) {
2309 case DDI_DATAPATH_FAULT:
2310 location = "in datapath to";
2311 break;
2312 case DDI_DEVICE_FAULT:
2313 location = "in";
2314 break;
2315 case DDI_EXTERNAL_FAULT:
2316 location = "external to";
2317 break;
2318 default:
2319 location = "somewhere near";
2320 bad = 1;
2321 break;
2322 }
2323
2324 newstate = ddi_get_devstate(fedp->f_dip);
2325 switch (newstate) {
2326 case DDI_DEVSTATE_OFFLINE:
2327 servstate = "unavailable";
2328 break;
2329 case DDI_DEVSTATE_DOWN:
2330 servstate = "unavailable";
2331 break;
2332 case DDI_DEVSTATE_QUIESCED:
2333 servstate = "suspended";
2334 break;
2335 case DDI_DEVSTATE_DEGRADED:
2336 servstate = "degraded";
2337 break;
2338 default:
2339 servstate = "available";
2340 break;
2341 }
2342
2343 changed = (newstate != fedp->f_oldstate);
2344 level = (newstate < fedp->f_oldstate) ? CE_WARN : CE_NOTE;
2345 switch (fedp->f_impact) {
2346 case DDI_SERVICE_LOST:
2347 case DDI_SERVICE_DEGRADED:
2348 case DDI_SERVICE_UNAFFECTED:
2349 /* fault detected; service [still] <servstate> */
2350 action = "fault detected";
2351 still = !changed;
2352 break;
2353
2354 case DDI_SERVICE_RESTORED:
2355 if (newstate != DDI_DEVSTATE_UP) {
2356 /* fault cleared; service still <servstate> */
2357 action = "fault cleared";
2358 still = 1;
2359 } else if (changed) {
2360 /* fault cleared; service <servstate> */
2361 action = "fault cleared";
2362 still = 0;
2363 } else {
2364 /* no fault; service <servstate> */
2365 action = "no fault";
2366 still = 0;
2367 }
2368 break;
2369
2370 default:
2371 bad = 1;
2372 break;
2373 }
2374
2375 cmn_err(level, "!%s%d: %s %s device; service %s%s"+(bad|changed),
2376 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip),
2377 bad ? "invalid report of fault" : action,
2378 location, still ? "still " : "", servstate);
2379
2380 cmn_err(level, "!%s%d: %s"+(bad|changed),
2381 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip),
2382 fedp->f_message);
2383 }
2384
2385 /*
2386 * Platform-settable pointers to fault handler and logger functions.
2387 * These are called by the default rootnex event-posting code when
2388 * a fault event reaches rootnex.
2389 */
2390 void (*plat_fault_handler)(dev_info_t *, struct ddi_fault_event_data *) =
2391 i_ddi_fault_handler;
2392 void (*plat_fault_logger)(dev_info_t *, struct ddi_fault_event_data *) =
2393 i_ddi_fault_logger;
2394
2395 /*
2396 * Rootnex event definitions ...
2397 */
2398 enum rootnex_event_tags {
2399 ROOTNEX_FAULT_EVENT
2400 };
2401 static ndi_event_hdl_t rootnex_event_hdl;
2402 static ndi_event_definition_t rootnex_event_set[] = {
2403 {
2404 ROOTNEX_FAULT_EVENT,
2405 DDI_DEVI_FAULT_EVENT,
2406 EPL_INTERRUPT,
2407 NDI_EVENT_POST_TO_ALL
2408 }
2409 };
2410 static ndi_event_set_t rootnex_events = {
2411 NDI_EVENTS_REV1,
2412 sizeof (rootnex_event_set) / sizeof (rootnex_event_set[0]),
2413 rootnex_event_set
2414 };
2415
2416 /*
2417 * Initialize rootnex event handle
2418 */
2419 void
i_ddi_rootnex_init_events(dev_info_t * dip)2420 i_ddi_rootnex_init_events(dev_info_t *dip)
2421 {
2422 if (ndi_event_alloc_hdl(dip, (ddi_iblock_cookie_t)(LOCK_LEVEL-1),
2423 &rootnex_event_hdl, NDI_SLEEP) == NDI_SUCCESS) {
2424 if (ndi_event_bind_set(rootnex_event_hdl,
2425 &rootnex_events, NDI_SLEEP) != NDI_SUCCESS) {
2426 (void) ndi_event_free_hdl(rootnex_event_hdl);
2427 rootnex_event_hdl = NULL;
2428 }
2429 }
2430 }
2431
2432 /*
2433 * Event-handling functions for rootnex
2434 * These provide the standard implementation of fault handling
2435 */
2436 /*ARGSUSED*/
2437 int
i_ddi_rootnex_get_eventcookie(dev_info_t * dip,dev_info_t * rdip,char * eventname,ddi_eventcookie_t * cookiep)2438 i_ddi_rootnex_get_eventcookie(dev_info_t *dip, dev_info_t *rdip,
2439 char *eventname, ddi_eventcookie_t *cookiep)
2440 {
2441 if (rootnex_event_hdl == NULL)
2442 return (NDI_FAILURE);
2443 return (ndi_event_retrieve_cookie(rootnex_event_hdl, rdip, eventname,
2444 cookiep, NDI_EVENT_NOPASS));
2445 }
2446
2447 /*ARGSUSED*/
2448 int
i_ddi_rootnex_add_eventcall(dev_info_t * dip,dev_info_t * rdip,ddi_eventcookie_t eventid,void (* handler)(dev_info_t * dip,ddi_eventcookie_t event,void * arg,void * impl_data),void * arg,ddi_callback_id_t * cb_id)2449 i_ddi_rootnex_add_eventcall(dev_info_t *dip, dev_info_t *rdip,
2450 ddi_eventcookie_t eventid, void (*handler)(dev_info_t *dip,
2451 ddi_eventcookie_t event, void *arg, void *impl_data), void *arg,
2452 ddi_callback_id_t *cb_id)
2453 {
2454 if (rootnex_event_hdl == NULL)
2455 return (NDI_FAILURE);
2456 return (ndi_event_add_callback(rootnex_event_hdl, rdip,
2457 eventid, handler, arg, NDI_SLEEP, cb_id));
2458 }
2459
2460 /*ARGSUSED*/
2461 int
i_ddi_rootnex_remove_eventcall(dev_info_t * dip,ddi_callback_id_t cb_id)2462 i_ddi_rootnex_remove_eventcall(dev_info_t *dip, ddi_callback_id_t cb_id)
2463 {
2464 if (rootnex_event_hdl == NULL)
2465 return (NDI_FAILURE);
2466
2467 return (ndi_event_remove_callback(rootnex_event_hdl, cb_id));
2468 }
2469
2470 /*ARGSUSED*/
2471 int
i_ddi_rootnex_post_event(dev_info_t * dip,dev_info_t * rdip,ddi_eventcookie_t eventid,void * impl_data)2472 i_ddi_rootnex_post_event(dev_info_t *dip, dev_info_t *rdip,
2473 ddi_eventcookie_t eventid, void *impl_data)
2474 {
2475 int tag;
2476
2477 if (rootnex_event_hdl == NULL)
2478 return (NDI_FAILURE);
2479
2480 tag = ndi_event_cookie_to_tag(rootnex_event_hdl, eventid);
2481 if (tag == ROOTNEX_FAULT_EVENT) {
2482 (*plat_fault_handler)(rdip, impl_data);
2483 (*plat_fault_logger)(rdip, impl_data);
2484 }
2485 return (ndi_event_run_callbacks(rootnex_event_hdl, rdip,
2486 eventid, impl_data));
2487 }
2488
2489 /*
2490 * ndi_set_bus_private/ndi_get_bus_private:
2491 * Get/set device bus private data in devinfo.
2492 */
2493 void
ndi_set_bus_private(dev_info_t * dip,boolean_t up,uint32_t port_type,void * data)2494 ndi_set_bus_private(dev_info_t *dip, boolean_t up, uint32_t port_type,
2495 void *data)
2496 {
2497 if (up) {
2498 DEVI(dip)->devi_bus.port_up.info.port.type = port_type;
2499 DEVI(dip)->devi_bus.port_up.priv_p = data;
2500 } else {
2501 DEVI(dip)->devi_bus.port_down.info.port.type = port_type;
2502 DEVI(dip)->devi_bus.port_down.priv_p = data;
2503 }
2504 }
2505
2506 void *
ndi_get_bus_private(dev_info_t * dip,boolean_t up)2507 ndi_get_bus_private(dev_info_t *dip, boolean_t up)
2508 {
2509 if (up)
2510 return (DEVI(dip)->devi_bus.port_up.priv_p);
2511 else
2512 return (DEVI(dip)->devi_bus.port_down.priv_p);
2513 }
2514
2515 boolean_t
ndi_port_type(dev_info_t * dip,boolean_t up,uint32_t port_type)2516 ndi_port_type(dev_info_t *dip, boolean_t up, uint32_t port_type)
2517 {
2518 if (up) {
2519 return ((DEVI(dip)->devi_bus.port_up.info.port.type) ==
2520 port_type);
2521 } else {
2522 return ((DEVI(dip)->devi_bus.port_down.info.port.type) ==
2523 port_type);
2524 }
2525 }
2526
2527 /* Interfaces for 'self' to set/get a child's flavor */
2528 void
ndi_flavor_set(dev_info_t * child,ndi_flavor_t child_flavor)2529 ndi_flavor_set(dev_info_t *child, ndi_flavor_t child_flavor)
2530 {
2531 DEVI(child)->devi_flavor = child_flavor;
2532 }
2533
2534 ndi_flavor_t
ndi_flavor_get(dev_info_t * child)2535 ndi_flavor_get(dev_info_t *child)
2536 {
2537 return (DEVI(child)->devi_flavor);
2538 }
2539
2540 /*
2541 * Interfaces to maintain flavor-specific private data of flavored
2542 * children of self.
2543 *
2544 * The flavor count always includes the default (0) vanilla flavor,
2545 * but storage for the vanilla flavor data pointer is in the same
2546 * place that ddi_[sg]et_driver_private uses, so the flavorv
2547 * storage is just for flavors 1..{nflavors-1}.
2548 */
2549 void
ndi_flavorv_alloc(dev_info_t * self,int nflavors)2550 ndi_flavorv_alloc(dev_info_t *self, int nflavors)
2551 {
2552 ASSERT(nflavors > 0 && (DEVI(self)->devi_flavorv == NULL ||
2553 nflavors == DEVI(self)->devi_flavorv_n));
2554 if (nflavors <= 1 || (DEVI(self)->devi_flavorv)) {
2555 return;
2556 }
2557 DEVI(self)->devi_flavorv =
2558 kmem_zalloc((nflavors - 1) * sizeof (void *), KM_SLEEP);
2559 DEVI(self)->devi_flavorv_n = nflavors;
2560 }
2561
2562 void
ndi_flavorv_set(dev_info_t * self,ndi_flavor_t child_flavor,void * v)2563 ndi_flavorv_set(dev_info_t *self, ndi_flavor_t child_flavor, void *v)
2564 {
2565 if (child_flavor == NDI_FLAVOR_VANILLA) {
2566 ddi_set_driver_private(self, v);
2567 } else {
2568 ASSERT(child_flavor < DEVI(self)->devi_flavorv_n &&
2569 DEVI(self)->devi_flavorv != NULL);
2570 if (child_flavor > DEVI(self)->devi_flavorv_n ||
2571 DEVI(self)->devi_flavorv == NULL) {
2572 return;
2573 }
2574 DEVI(self)->devi_flavorv[child_flavor - 1] = v;
2575 }
2576 }
2577
2578 void *
ndi_flavorv_get(dev_info_t * self,ndi_flavor_t child_flavor)2579 ndi_flavorv_get(dev_info_t *self, ndi_flavor_t child_flavor)
2580 {
2581 if (child_flavor == NDI_FLAVOR_VANILLA) {
2582 return (ddi_get_driver_private(self));
2583 } else {
2584 ASSERT(child_flavor < DEVI(self)->devi_flavorv_n &&
2585 DEVI(self)->devi_flavorv != NULL);
2586 if (child_flavor > DEVI(self)->devi_flavorv_n ||
2587 DEVI(self)->devi_flavorv == NULL) {
2588 return (NULL);
2589 }
2590 return (DEVI(self)->devi_flavorv[child_flavor - 1]);
2591 }
2592 }
2593