1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2008 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 #include <sys/types.h> 29 #include <sys/param.h> 30 #include <sys/systm.h> 31 #include <sys/buf.h> 32 #include <sys/uio.h> 33 #include <sys/cred.h> 34 #include <sys/poll.h> 35 #include <sys/mman.h> 36 #include <sys/kmem.h> 37 #include <sys/model.h> 38 #include <sys/file.h> 39 #include <sys/proc.h> 40 #include <sys/open.h> 41 #include <sys/user.h> 42 #include <sys/t_lock.h> 43 #include <sys/vm.h> 44 #include <sys/stat.h> 45 #include <vm/hat.h> 46 #include <vm/seg.h> 47 #include <vm/as.h> 48 #include <sys/cmn_err.h> 49 #include <sys/debug.h> 50 #include <sys/avintr.h> 51 #include <sys/autoconf.h> 52 #include <sys/sunddi.h> 53 #include <sys/esunddi.h> 54 #include <sys/sunndi.h> 55 #include <sys/ddi.h> 56 #include <sys/kstat.h> 57 #include <sys/conf.h> 58 #include <sys/ddi_impldefs.h> /* include implementation structure defs */ 59 #include <sys/ndi_impldefs.h> 60 #include <sys/hwconf.h> 61 #include <sys/pathname.h> 62 #include <sys/modctl.h> 63 #include <sys/epm.h> 64 #include <sys/devctl.h> 65 #include <sys/callb.h> 66 #include <sys/bootconf.h> 67 #include <sys/dacf_impl.h> 68 #include <sys/nvpair.h> 69 #include <sys/sunmdi.h> 70 #include <sys/fs/dv_node.h> 71 #include <sys/sunldi_impl.h> 72 73 #ifdef __sparc 74 #include <sys/archsystm.h> /* getpil/setpil */ 75 #include <sys/membar.h> /* membar_sync */ 76 #endif 77 78 /* 79 * ndi property handling 80 */ 81 int 82 ndi_prop_update_int(dev_t match_dev, dev_info_t *dip, 83 char *name, int data) 84 { 85 return (ddi_prop_update_common(match_dev, dip, 86 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP, 87 name, &data, 1, ddi_prop_fm_encode_ints)); 88 } 89 90 int 91 ndi_prop_update_int64(dev_t match_dev, dev_info_t *dip, 92 char *name, int64_t data) 93 { 94 return (ddi_prop_update_common(match_dev, dip, 95 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP, 96 name, &data, 1, ddi_prop_fm_encode_int64)); 97 } 98 99 int 100 ndi_prop_create_boolean(dev_t match_dev, dev_info_t *dip, 101 char *name) 102 { 103 return (ddi_prop_update_common(match_dev, dip, 104 DDI_PROP_HW_DEF | DDI_PROP_TYPE_ANY | DDI_PROP_DONTSLEEP, 105 name, NULL, 0, ddi_prop_fm_encode_bytes)); 106 } 107 108 int 109 ndi_prop_update_int_array(dev_t match_dev, dev_info_t *dip, 110 char *name, int *data, uint_t nelements) 111 { 112 return (ddi_prop_update_common(match_dev, dip, 113 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP, 114 name, data, nelements, ddi_prop_fm_encode_ints)); 115 } 116 117 int 118 ndi_prop_update_int64_array(dev_t match_dev, dev_info_t *dip, 119 char *name, int64_t *data, uint_t nelements) 120 { 121 return (ddi_prop_update_common(match_dev, dip, 122 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP, 123 name, data, nelements, ddi_prop_fm_encode_int64)); 124 } 125 126 int 127 ndi_prop_update_string(dev_t match_dev, dev_info_t *dip, 128 char *name, char *data) 129 { 130 return (ddi_prop_update_common(match_dev, dip, 131 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP, 132 name, &data, 1, ddi_prop_fm_encode_string)); 133 } 134 135 int 136 ndi_prop_update_string_array(dev_t match_dev, dev_info_t *dip, 137 char *name, char **data, uint_t nelements) 138 { 139 return (ddi_prop_update_common(match_dev, dip, 140 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP, 141 name, data, nelements, 142 ddi_prop_fm_encode_strings)); 143 } 144 145 int 146 ndi_prop_update_byte_array(dev_t match_dev, dev_info_t *dip, 147 char *name, uchar_t *data, uint_t nelements) 148 { 149 if (nelements == 0) 150 return (DDI_PROP_INVAL_ARG); 151 152 return (ddi_prop_update_common(match_dev, dip, 153 DDI_PROP_HW_DEF | DDI_PROP_TYPE_BYTE | DDI_PROP_DONTSLEEP, 154 name, data, nelements, ddi_prop_fm_encode_bytes)); 155 } 156 157 int 158 ndi_prop_remove(dev_t dev, dev_info_t *dip, char *name) 159 { 160 return (ddi_prop_remove_common(dev, dip, name, DDI_PROP_HW_DEF)); 161 } 162 163 void 164 ndi_prop_remove_all(dev_info_t *dip) 165 { 166 i_ddi_prop_dyn_parent_set(dip, NULL); 167 ddi_prop_remove_all_common(dip, (int)DDI_PROP_HW_DEF); 168 } 169 170 /* 171 * Post an event notification to nexus driver responsible for handling 172 * the event. The responsible nexus is defined in the cookie passed in as 173 * the third parameter. 174 * The dip parameter is an artifact of an older implementation in which all 175 * requests to remove an eventcall would bubble up the tree. Today, this 176 * parameter is ignored. 177 * Input Parameters: 178 * dip - Ignored. 179 * rdip - device driver posting the event 180 * cookie - valid ddi_eventcookie_t, obtained by caller prior to 181 * invocation of this routine 182 * impl_data - used by framework 183 */ 184 /*ARGSUSED*/ 185 int 186 ndi_post_event(dev_info_t *dip, dev_info_t *rdip, 187 ddi_eventcookie_t cookie, void *impl_data) 188 { 189 dev_info_t *ddip; 190 191 ASSERT(cookie); 192 ddip = NDI_EVENT_DDIP(cookie); 193 194 /* 195 * perform sanity checks. These conditions should never be true. 196 */ 197 198 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 199 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 200 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event != NULL); 201 202 /* 203 * post the event to the responsible ancestor 204 */ 205 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event)) 206 (ddip, rdip, cookie, impl_data)); 207 } 208 209 /* 210 * Calls the bus nexus driver's implementation of the 211 * (*bus_remove_eventcall)() interface. 212 */ 213 int 214 ndi_busop_remove_eventcall(dev_info_t *ddip, ddi_callback_id_t id) 215 { 216 217 ASSERT(id); 218 /* check for a correct revno before calling up the device tree. */ 219 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 220 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 221 222 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall == NULL) 223 return (DDI_FAILURE); 224 225 /* 226 * request responsible nexus to remove the eventcall 227 */ 228 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall)) 229 (ddip, id)); 230 } 231 232 /* 233 * Calls the bus nexus driver's implementation of the 234 * (*bus_add_eventcall)() interface. The dip parameter is an 235 * artifact of an older implementation in which all requests to 236 * add an eventcall would bubble up the tree. Today, this parameter is 237 * ignored. 238 */ 239 /*ARGSUSED*/ 240 int 241 ndi_busop_add_eventcall(dev_info_t *dip, dev_info_t *rdip, 242 ddi_eventcookie_t cookie, void (*callback)(), void *arg, 243 ddi_callback_id_t *cb_id) 244 { 245 dev_info_t *ddip = (dev_info_t *)NDI_EVENT_DDIP(cookie); 246 247 /* 248 * check for a correct revno before calling up the device tree. 249 */ 250 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 251 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 252 253 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall == NULL) 254 return (DDI_FAILURE); 255 256 /* 257 * request responsible ancestor to add the eventcall 258 */ 259 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall)) 260 (ddip, rdip, cookie, callback, arg, cb_id)); 261 } 262 263 /* 264 * Calls the bus nexus driver's implementation of the 265 * (*bus_get_eventcookie)() interface up the device tree hierarchy. 266 */ 267 int 268 ndi_busop_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, char *name, 269 ddi_eventcookie_t *event_cookiep) 270 { 271 dev_info_t *pdip = (dev_info_t *)DEVI(dip)->devi_parent; 272 273 /* Can not be called from rootnex. */ 274 ASSERT(pdip); 275 276 /* 277 * check for a correct revno before calling up the device tree. 278 */ 279 ASSERT(DEVI(pdip)->devi_ops->devo_bus_ops != NULL); 280 281 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev < BUSO_REV_6) || 282 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie == NULL)) { 283 #ifdef DEBUG 284 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev >= 285 BUSO_REV_3) && 286 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie)) { 287 cmn_err(CE_WARN, 288 "Warning: %s%d busops_rev=%d no longer supported" 289 " by the NDI event framework.\nBUSO_REV_6 or " 290 "greater must be used.", 291 DEVI(pdip)->devi_binding_name, 292 DEVI(pdip)->devi_instance, 293 DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev); 294 } 295 #endif /* DEBUG */ 296 297 return (ndi_busop_get_eventcookie(pdip, rdip, name, 298 event_cookiep)); 299 } 300 301 return ((*(DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie)) 302 (pdip, rdip, name, event_cookiep)); 303 } 304 305 /* 306 * Copy in the devctl IOCTL data and return a handle to 307 * the data. 308 */ 309 int 310 ndi_dc_allochdl(void *iocarg, struct devctl_iocdata **rdcp) 311 { 312 struct devctl_iocdata *dcp; 313 char *cpybuf; 314 315 ASSERT(rdcp != NULL); 316 317 dcp = kmem_zalloc(sizeof (*dcp), KM_SLEEP); 318 319 if (get_udatamodel() == DATAMODEL_NATIVE) { 320 if (copyin(iocarg, dcp, sizeof (*dcp)) != 0) { 321 kmem_free(dcp, sizeof (*dcp)); 322 return (NDI_FAULT); 323 } 324 } 325 #ifdef _SYSCALL32_IMPL 326 else { 327 struct devctl_iocdata32 dcp32; 328 329 if (copyin(iocarg, &dcp32, sizeof (dcp32)) != 0) { 330 kmem_free(dcp, sizeof (*dcp)); 331 return (NDI_FAULT); 332 } 333 dcp->cmd = (uint_t)dcp32.cmd; 334 dcp->flags = (uint_t)dcp32.flags; 335 dcp->cpyout_buf = (uint_t *)(uintptr_t)dcp32.cpyout_buf; 336 dcp->nvl_user = (nvlist_t *)(uintptr_t)dcp32.nvl_user; 337 dcp->nvl_usersz = (size_t)dcp32.nvl_usersz; 338 dcp->c_nodename = (char *)(uintptr_t)dcp32.c_nodename; 339 dcp->c_unitaddr = (char *)(uintptr_t)dcp32.c_unitaddr; 340 } 341 #endif 342 if (dcp->c_nodename != NULL) { 343 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP); 344 if (copyinstr(dcp->c_nodename, cpybuf, MAXNAMELEN, 0) != 0) { 345 kmem_free(cpybuf, MAXNAMELEN); 346 kmem_free(dcp, sizeof (*dcp)); 347 return (NDI_FAULT); 348 } 349 cpybuf[MAXNAMELEN - 1] = '\0'; 350 dcp->c_nodename = cpybuf; 351 } 352 353 if (dcp->c_unitaddr != NULL) { 354 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP); 355 if (copyinstr(dcp->c_unitaddr, cpybuf, MAXNAMELEN, 0) != 0) { 356 kmem_free(cpybuf, MAXNAMELEN); 357 if (dcp->c_nodename != NULL) 358 kmem_free(dcp->c_nodename, MAXNAMELEN); 359 kmem_free(dcp, sizeof (*dcp)); 360 return (NDI_FAULT); 361 } 362 cpybuf[MAXNAMELEN - 1] = '\0'; 363 dcp->c_unitaddr = cpybuf; 364 } 365 366 /* 367 * copyin and unpack a user defined nvlist if one was passed 368 */ 369 if (dcp->nvl_user != NULL) { 370 if ((dcp->nvl_usersz == 0) || 371 (dcp->nvl_usersz > DEVCTL_MAX_NVL_USERSZ)) { 372 if (dcp->c_nodename != NULL) 373 kmem_free(dcp->c_nodename, MAXNAMELEN); 374 if (dcp->c_unitaddr != NULL) 375 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 376 kmem_free(dcp, sizeof (*dcp)); 377 return (NDI_FAILURE); 378 } 379 cpybuf = kmem_alloc(dcp->nvl_usersz, KM_SLEEP); 380 if (copyin(dcp->nvl_user, cpybuf, dcp->nvl_usersz) != 0) { 381 kmem_free(cpybuf, dcp->nvl_usersz); 382 if (dcp->c_nodename != NULL) 383 kmem_free(dcp->c_nodename, MAXNAMELEN); 384 if (dcp->c_unitaddr != NULL) 385 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 386 kmem_free(dcp, sizeof (*dcp)); 387 return (NDI_FAULT); 388 } 389 390 if (nvlist_unpack(cpybuf, dcp->nvl_usersz, &dcp->nvl_user, 391 KM_SLEEP)) { 392 kmem_free(cpybuf, dcp->nvl_usersz); 393 if (dcp->c_nodename != NULL) 394 kmem_free(dcp->c_nodename, MAXNAMELEN); 395 if (dcp->c_unitaddr != NULL) 396 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 397 kmem_free(dcp, sizeof (*dcp)); 398 return (NDI_FAULT); 399 } 400 /* 401 * free the buffer containing the packed nvlist 402 */ 403 kmem_free(cpybuf, dcp->nvl_usersz); 404 405 } 406 407 *rdcp = dcp; 408 return (NDI_SUCCESS); 409 } 410 411 /* 412 * free all space allocated to a handle. 413 */ 414 void 415 ndi_dc_freehdl(struct devctl_iocdata *dcp) 416 { 417 ASSERT(dcp != NULL); 418 419 if (dcp->c_nodename != NULL) 420 kmem_free(dcp->c_nodename, MAXNAMELEN); 421 422 if (dcp->c_unitaddr != NULL) 423 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 424 425 if (dcp->nvl_user != NULL) 426 nvlist_free(dcp->nvl_user); 427 428 kmem_free(dcp, sizeof (*dcp)); 429 } 430 431 char * 432 ndi_dc_getname(struct devctl_iocdata *dcp) 433 { 434 ASSERT(dcp != NULL); 435 return (dcp->c_nodename); 436 437 } 438 439 char * 440 ndi_dc_getaddr(struct devctl_iocdata *dcp) 441 { 442 ASSERT(dcp != NULL); 443 return (dcp->c_unitaddr); 444 } 445 446 nvlist_t * 447 ndi_dc_get_ap_data(struct devctl_iocdata *dcp) 448 { 449 ASSERT(dcp != NULL); 450 451 return (dcp->nvl_user); 452 } 453 454 /* 455 * Transition the child named by "devname@devaddr" to the online state. 456 * For use by a driver's DEVCTL_DEVICE_ONLINE handler. 457 */ 458 int 459 ndi_devctl_device_online(dev_info_t *dip, struct devctl_iocdata *dcp, 460 uint_t flags) 461 { 462 int rval; 463 char *name; 464 dev_info_t *rdip; 465 466 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 467 return (EINVAL); 468 469 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 470 (void) snprintf(name, MAXNAMELEN, "%s@%s", 471 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 472 473 if ((rval = ndi_devi_config_one(dip, name, &rdip, 474 flags | NDI_DEVI_ONLINE | NDI_CONFIG)) == NDI_SUCCESS) { 475 ndi_rele_devi(rdip); 476 477 /* 478 * Invalidate devfs cached directory contents. For the checks 479 * in the "if" condition see the comment in ndi_devi_online(). 480 */ 481 if (i_ddi_devi_attached(dip) && !DEVI_BUSY_OWNED(dip)) 482 (void) devfs_clean(dip, NULL, 0); 483 484 } else if (rval == NDI_BUSY) { 485 rval = EBUSY; 486 } else if (rval == NDI_FAILURE) { 487 rval = EIO; 488 } 489 490 NDI_DEBUG(flags, (CE_CONT, "%s%d: online: %s: %s\n", 491 ddi_driver_name(dip), ddi_get_instance(dip), name, 492 ((rval == NDI_SUCCESS) ? "ok" : "failed"))); 493 494 kmem_free(name, MAXNAMELEN); 495 496 return (rval); 497 } 498 499 /* 500 * Transition the child named by "devname@devaddr" to the offline state. 501 * For use by a driver's DEVCTL_DEVICE_OFFLINE handler. 502 */ 503 int 504 ndi_devctl_device_offline(dev_info_t *dip, struct devctl_iocdata *dcp, 505 uint_t flags) 506 { 507 int rval; 508 char *name; 509 510 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 511 return (EINVAL); 512 513 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 514 (void) snprintf(name, MAXNAMELEN, "%s@%s", 515 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 516 517 (void) devfs_clean(dip, name, DV_CLEAN_FORCE); 518 rval = ndi_devi_unconfig_one(dip, name, NULL, 519 flags | NDI_DEVI_OFFLINE); 520 521 if (rval == NDI_BUSY) { 522 rval = EBUSY; 523 } else if (rval == NDI_FAILURE) { 524 rval = EIO; 525 } 526 527 NDI_DEBUG(flags, (CE_CONT, "%s%d: offline: %s: %s\n", 528 ddi_driver_name(dip), ddi_get_instance(dip), name, 529 (rval == NDI_SUCCESS) ? "ok" : "failed")); 530 531 kmem_free(name, MAXNAMELEN); 532 533 return (rval); 534 } 535 536 /* 537 * Remove the child named by "devname@devaddr". 538 * For use by a driver's DEVCTL_DEVICE_REMOVE handler. 539 */ 540 int 541 ndi_devctl_device_remove(dev_info_t *dip, struct devctl_iocdata *dcp, 542 uint_t flags) 543 { 544 int rval; 545 char *name; 546 547 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 548 return (EINVAL); 549 550 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 551 (void) snprintf(name, MAXNAMELEN, "%s@%s", 552 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 553 554 (void) devfs_clean(dip, name, DV_CLEAN_FORCE); 555 556 rval = ndi_devi_unconfig_one(dip, name, NULL, flags | NDI_DEVI_REMOVE); 557 558 if (rval == NDI_BUSY) { 559 rval = EBUSY; 560 } else if (rval == NDI_FAILURE) { 561 rval = EIO; 562 } 563 564 NDI_DEBUG(flags, (CE_CONT, "%s%d: remove: %s: %s\n", 565 ddi_driver_name(dip), ddi_get_instance(dip), name, 566 (rval == NDI_SUCCESS) ? "ok" : "failed")); 567 568 kmem_free(name, MAXNAMELEN); 569 570 return (rval); 571 } 572 573 /* 574 * Return devctl state of the child named by "name@addr". 575 * For use by a driver's DEVCTL_DEVICE_GETSTATE handler. 576 */ 577 int 578 ndi_devctl_device_getstate(dev_info_t *parent, struct devctl_iocdata *dcp, 579 uint_t *state) 580 { 581 dev_info_t *dip; 582 char *name, *addr; 583 char *devname; 584 int devnamelen; 585 int circ; 586 587 if (parent == NULL || 588 ((name = ndi_dc_getname(dcp)) == NULL) || 589 ((addr = ndi_dc_getaddr(dcp)) == NULL)) 590 return (NDI_FAILURE); 591 592 devnamelen = strlen(name) + strlen(addr) + 2; 593 devname = kmem_alloc(devnamelen, KM_SLEEP); 594 if (strlen(addr) > 0) { 595 (void) snprintf(devname, devnamelen, "%s@%s", name, addr); 596 } else { 597 (void) snprintf(devname, devnamelen, "%s", name); 598 } 599 600 ndi_devi_enter(parent, &circ); 601 602 dip = ndi_devi_findchild(parent, devname); 603 kmem_free(devname, devnamelen); 604 605 if (dip == NULL) { 606 ndi_devi_exit(parent, circ); 607 return (NDI_FAILURE); 608 } 609 610 mutex_enter(&(DEVI(dip)->devi_lock)); 611 if (DEVI_IS_DEVICE_OFFLINE(dip)) { 612 *state = DEVICE_OFFLINE; 613 } else if (DEVI_IS_DEVICE_DOWN(dip)) { 614 *state = DEVICE_DOWN; 615 } else { 616 *state = DEVICE_ONLINE; 617 if (devi_stillreferenced(dip) == DEVI_REFERENCED) 618 *state |= DEVICE_BUSY; 619 } 620 621 mutex_exit(&(DEVI(dip)->devi_lock)); 622 ndi_devi_exit(parent, circ); 623 624 return (NDI_SUCCESS); 625 } 626 627 /* 628 * return the current state of the device "dip" 629 * 630 * recommend using ndi_devctl_ioctl() or 631 * ndi_devctl_device_getstate() instead 632 */ 633 int 634 ndi_dc_return_dev_state(dev_info_t *dip, struct devctl_iocdata *dcp) 635 { 636 dev_info_t *pdip; 637 uint_t devstate = 0; 638 int circ; 639 640 if ((dip == NULL) || (dcp == NULL)) 641 return (NDI_FAILURE); 642 643 pdip = ddi_get_parent(dip); 644 645 ndi_devi_enter(pdip, &circ); 646 mutex_enter(&(DEVI(dip)->devi_lock)); 647 if (DEVI_IS_DEVICE_OFFLINE(dip)) { 648 devstate = DEVICE_OFFLINE; 649 } else if (DEVI_IS_DEVICE_DOWN(dip)) { 650 devstate = DEVICE_DOWN; 651 } else { 652 devstate = DEVICE_ONLINE; 653 if (devi_stillreferenced(dip) == DEVI_REFERENCED) 654 devstate |= DEVICE_BUSY; 655 } 656 657 mutex_exit(&(DEVI(dip)->devi_lock)); 658 ndi_devi_exit(pdip, circ); 659 660 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0) 661 return (NDI_FAULT); 662 663 return (NDI_SUCCESS); 664 } 665 666 /* 667 * Return device's bus state 668 * For use by a driver's DEVCTL_BUS_GETSTATE handler. 669 */ 670 int 671 ndi_devctl_bus_getstate(dev_info_t *dip, struct devctl_iocdata *dcp, 672 uint_t *state) 673 { 674 if ((dip == NULL) || (dcp == NULL)) 675 return (NDI_FAILURE); 676 677 return (ndi_get_bus_state(dip, state)); 678 } 679 680 /* 681 * Generic devctl ioctl handler 682 */ 683 int 684 ndi_devctl_ioctl(dev_info_t *dip, int cmd, intptr_t arg, int mode, uint_t flags) 685 { 686 _NOTE(ARGUNUSED(mode)) 687 struct devctl_iocdata *dcp; 688 uint_t state; 689 int rval = ENOTTY; 690 691 /* 692 * read devctl ioctl data 693 */ 694 if (ndi_dc_allochdl((void *)arg, &dcp) != NDI_SUCCESS) 695 return (EFAULT); 696 697 switch (cmd) { 698 699 case DEVCTL_BUS_GETSTATE: 700 rval = ndi_devctl_bus_getstate(dip, dcp, &state); 701 if (rval == NDI_SUCCESS) { 702 if (copyout(&state, dcp->cpyout_buf, 703 sizeof (uint_t)) != 0) 704 rval = NDI_FAULT; 705 } 706 break; 707 708 case DEVCTL_DEVICE_ONLINE: 709 rval = ndi_devctl_device_online(dip, dcp, flags); 710 break; 711 712 case DEVCTL_DEVICE_OFFLINE: 713 rval = ndi_devctl_device_offline(dip, dcp, flags); 714 break; 715 716 case DEVCTL_DEVICE_GETSTATE: 717 rval = ndi_devctl_device_getstate(dip, dcp, &state); 718 if (rval == NDI_SUCCESS) { 719 if (copyout(&state, dcp->cpyout_buf, 720 sizeof (uint_t)) != 0) 721 rval = NDI_FAULT; 722 } 723 break; 724 725 case DEVCTL_DEVICE_REMOVE: 726 rval = ndi_devctl_device_remove(dip, dcp, flags); 727 break; 728 729 case DEVCTL_BUS_DEV_CREATE: 730 rval = ndi_dc_devi_create(dcp, dip, 0, NULL); 731 break; 732 733 /* 734 * ioctls for which a generic implementation makes no sense 735 */ 736 case DEVCTL_BUS_RESET: 737 case DEVCTL_BUS_RESETALL: 738 case DEVCTL_DEVICE_RESET: 739 case DEVCTL_AP_CONNECT: 740 case DEVCTL_AP_DISCONNECT: 741 case DEVCTL_AP_INSERT: 742 case DEVCTL_AP_REMOVE: 743 case DEVCTL_AP_CONFIGURE: 744 case DEVCTL_AP_UNCONFIGURE: 745 case DEVCTL_AP_GETSTATE: 746 case DEVCTL_AP_CONTROL: 747 case DEVCTL_BUS_QUIESCE: 748 case DEVCTL_BUS_UNQUIESCE: 749 rval = ENOTSUP; 750 break; 751 } 752 753 ndi_dc_freehdl(dcp); 754 return (rval); 755 } 756 757 /* 758 * Copyout the state of the Attachment Point "ap" to the requesting 759 * user process. 760 */ 761 int 762 ndi_dc_return_ap_state(devctl_ap_state_t *ap, struct devctl_iocdata *dcp) 763 { 764 if ((ap == NULL) || (dcp == NULL)) 765 return (NDI_FAILURE); 766 767 768 if (get_udatamodel() == DATAMODEL_NATIVE) { 769 if (copyout(ap, dcp->cpyout_buf, 770 sizeof (devctl_ap_state_t)) != 0) 771 return (NDI_FAULT); 772 } 773 #ifdef _SYSCALL32_IMPL 774 else { 775 struct devctl_ap_state32 ap_state32; 776 777 ap_state32.ap_rstate = ap->ap_rstate; 778 ap_state32.ap_ostate = ap->ap_ostate; 779 ap_state32.ap_condition = ap->ap_condition; 780 ap_state32.ap_error_code = ap->ap_error_code; 781 ap_state32.ap_in_transition = ap->ap_in_transition; 782 ap_state32.ap_last_change = (time32_t)ap->ap_last_change; 783 if (copyout(&ap_state32, dcp->cpyout_buf, 784 sizeof (devctl_ap_state32_t)) != 0) 785 return (NDI_FAULT); 786 } 787 #endif 788 789 return (NDI_SUCCESS); 790 } 791 792 /* 793 * Copyout the bus state of the bus nexus device "dip" to the requesting 794 * user process. 795 */ 796 int 797 ndi_dc_return_bus_state(dev_info_t *dip, struct devctl_iocdata *dcp) 798 { 799 uint_t devstate = 0; 800 801 if ((dip == NULL) || (dcp == NULL)) 802 return (NDI_FAILURE); 803 804 if (ndi_get_bus_state(dip, &devstate) != NDI_SUCCESS) 805 return (NDI_FAILURE); 806 807 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0) 808 return (NDI_FAULT); 809 810 return (NDI_SUCCESS); 811 } 812 813 static int 814 i_dc_devi_create(struct devctl_iocdata *, dev_info_t *, dev_info_t **); 815 816 /* 817 * create a child device node given the property definitions 818 * supplied by the userland process 819 */ 820 int 821 ndi_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip, int flags, 822 dev_info_t **rdip) 823 { 824 dev_info_t *cdip; 825 int rv, circular = 0; 826 char devnm[MAXNAMELEN]; 827 int nmlen; 828 829 /* 830 * The child device may have been pre-constructed by an earlier 831 * call to this function with the flag DEVCTL_CONSTRUCT set. 832 */ 833 834 if ((cdip = (rdip != NULL) ? *rdip : NULL) == NULL) 835 if ((rv = i_dc_devi_create(dcp, pdip, &cdip)) != 0) 836 return (rv); 837 838 ASSERT(cdip != NULL); 839 840 /* 841 * Return the device node partially constructed if the 842 * DEVCTL_CONSTRUCT flag is set. 843 */ 844 if (flags & DEVCTL_CONSTRUCT) { 845 if (rdip == NULL) { 846 (void) ndi_devi_free(cdip); 847 return (EINVAL); 848 } 849 *rdip = cdip; 850 return (0); 851 } 852 853 /* 854 * Bring the node up to a named but OFFLINE state. The calling 855 * application will need to manage the node from here on. 856 */ 857 if (dcp->flags & DEVCTL_OFFLINE) { 858 /* 859 * In the unlikely event that the dip was somehow attached by 860 * the userland process (and device contracts or LDI opens 861 * were registered against the dip) after it was created by 862 * a previous DEVCTL_CONSTRUCT call, we start notify 863 * proceedings on this dip. Note that we don't need to 864 * return the dip after a failure of the notify since 865 * for a contract or LDI handle to be created the dip was 866 * already available to the user. 867 */ 868 if (e_ddi_offline_notify(cdip) == DDI_FAILURE) { 869 return (EBUSY); 870 } 871 872 /* 873 * hand set the OFFLINE flag to prevent any asynchronous 874 * autoconfiguration operations from attaching this node. 875 */ 876 mutex_enter(&(DEVI(cdip)->devi_lock)); 877 DEVI_SET_DEVICE_OFFLINE(cdip); 878 mutex_exit(&(DEVI(cdip)->devi_lock)); 879 880 e_ddi_offline_finalize(cdip, DDI_SUCCESS); 881 882 rv = ndi_devi_bind_driver(cdip, flags); 883 if (rv != NDI_SUCCESS) { 884 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 885 return (ENXIO); 886 } 887 888 /* 889 * remove the dev_info node if it failed to bind to a 890 * driver above. 891 */ 892 if (i_ddi_node_state(cdip) < DS_BOUND) { 893 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 894 return (ENXIO); 895 } 896 897 /* 898 * add the node to the per-driver list and INITCHILD it 899 * to give it a name. 900 */ 901 ndi_devi_enter(pdip, &circular); 902 if ((rv = ddi_initchild(pdip, cdip)) != DDI_SUCCESS) { 903 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 904 ndi_devi_exit(pdip, circular); 905 return (EINVAL); 906 } 907 ndi_devi_exit(pdip, circular); 908 909 } else { 910 /* 911 * Attempt to bring the device ONLINE. If the request to 912 * fails, remove the dev_info node. 913 */ 914 if (ndi_devi_online(cdip, NDI_ONLINE_ATTACH) != NDI_SUCCESS) { 915 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 916 return (ENXIO); 917 } 918 919 /* 920 * if the node was successfully added but there was 921 * no driver available for the device, remove the node 922 */ 923 if (i_ddi_node_state(cdip) < DS_BOUND) { 924 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 925 return (ENODEV); 926 } 927 } 928 929 /* 930 * return a handle to the child device 931 * copy out the name of the newly attached child device if 932 * the IOCTL request has provided a copyout buffer. 933 */ 934 if (rdip != NULL) 935 *rdip = cdip; 936 937 if (dcp->cpyout_buf == NULL) 938 return (0); 939 940 ASSERT(ddi_node_name(cdip) != NULL); 941 ASSERT(ddi_get_name_addr(cdip) != NULL); 942 943 nmlen = snprintf(devnm, MAXNAMELEN, "%s@%s", 944 ddi_node_name(cdip), ddi_get_name_addr(cdip)); 945 946 if (copyout(&devnm, dcp->cpyout_buf, nmlen) != 0) { 947 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 948 return (EFAULT); 949 } 950 return (0); 951 } 952 953 static int 954 i_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip, 955 dev_info_t **rdip) 956 { 957 958 dev_info_t *cdip; 959 char *cname = NULL; 960 nvlist_t *nvlp = dcp->nvl_user; 961 nvpair_t *npp; 962 char *np; 963 int rv = 0; 964 965 ASSERT(rdip != NULL && *rdip == NULL); 966 967 if ((nvlp == NULL) || 968 (nvlist_lookup_string(nvlp, DC_DEVI_NODENAME, &cname) != 0)) 969 return (EINVAL); 970 971 /* 972 * construct a new dev_info node with a user-provided nodename 973 */ 974 ndi_devi_alloc_sleep(pdip, cname, (pnode_t)DEVI_SID_NODEID, &cdip); 975 976 /* 977 * create hardware properties for each member in the property 978 * list. 979 */ 980 for (npp = nvlist_next_nvpair(nvlp, NULL); (npp != NULL && !rv); 981 npp = nvlist_next_nvpair(nvlp, npp)) { 982 983 np = nvpair_name(npp); 984 985 /* 986 * skip the nodename property 987 */ 988 if (strcmp(np, DC_DEVI_NODENAME) == 0) 989 continue; 990 991 switch (nvpair_type(npp)) { 992 993 case DATA_TYPE_INT32: { 994 int32_t prop_val; 995 996 if ((rv = nvpair_value_int32(npp, &prop_val)) != 0) 997 break; 998 999 (void) ndi_prop_update_int(DDI_DEV_T_NONE, cdip, np, 1000 (int)prop_val); 1001 break; 1002 } 1003 1004 case DATA_TYPE_STRING: { 1005 char *prop_val; 1006 1007 if ((rv = nvpair_value_string(npp, &prop_val)) != 0) 1008 break; 1009 1010 (void) ndi_prop_update_string(DDI_DEV_T_NONE, cdip, 1011 np, prop_val); 1012 break; 1013 } 1014 1015 case DATA_TYPE_BYTE_ARRAY: { 1016 uchar_t *val; 1017 uint_t nelms; 1018 1019 if ((rv = nvpair_value_byte_array(npp, &val, 1020 &nelms)) != 0) 1021 break; 1022 1023 (void) ndi_prop_update_byte_array(DDI_DEV_T_NONE, 1024 cdip, np, (uchar_t *)val, nelms); 1025 break; 1026 } 1027 1028 case DATA_TYPE_INT32_ARRAY: { 1029 int32_t *val; 1030 uint_t nelms; 1031 1032 if ((rv = nvpair_value_int32_array(npp, &val, 1033 &nelms)) != 0) 1034 break; 1035 1036 (void) ndi_prop_update_int_array(DDI_DEV_T_NONE, 1037 cdip, np, val, nelms); 1038 break; 1039 } 1040 1041 case DATA_TYPE_STRING_ARRAY: { 1042 char **val; 1043 uint_t nelms; 1044 1045 if ((rv = nvpair_value_string_array(npp, &val, 1046 &nelms)) != 0) 1047 break; 1048 1049 (void) ndi_prop_update_string_array(DDI_DEV_T_NONE, 1050 cdip, np, val, nelms); 1051 break; 1052 } 1053 1054 /* 1055 * unsupported property data type 1056 */ 1057 default: 1058 rv = EINVAL; 1059 } 1060 } 1061 1062 /* 1063 * something above failed 1064 * destroy the partially child device and abort the request 1065 */ 1066 if (rv != 0) { 1067 (void) ndi_devi_free(cdip); 1068 return (rv); 1069 } 1070 1071 *rdip = cdip; 1072 return (0); 1073 } 1074 1075 /* 1076 * return current soft bus state of bus nexus "dip" 1077 */ 1078 int 1079 ndi_get_bus_state(dev_info_t *dip, uint_t *rstate) 1080 { 1081 if (dip == NULL || rstate == NULL) 1082 return (NDI_FAILURE); 1083 1084 if (DEVI(dip)->devi_ops->devo_bus_ops == NULL) 1085 return (NDI_FAILURE); 1086 1087 mutex_enter(&(DEVI(dip)->devi_lock)); 1088 if (DEVI_IS_BUS_QUIESCED(dip)) 1089 *rstate = BUS_QUIESCED; 1090 else if (DEVI_IS_BUS_DOWN(dip)) 1091 *rstate = BUS_SHUTDOWN; 1092 else 1093 *rstate = BUS_ACTIVE; 1094 mutex_exit(&(DEVI(dip)->devi_lock)); 1095 return (NDI_SUCCESS); 1096 } 1097 1098 /* 1099 * Set the soft state of bus nexus "dip" 1100 */ 1101 int 1102 ndi_set_bus_state(dev_info_t *dip, uint_t state) 1103 { 1104 int rv = NDI_SUCCESS; 1105 1106 if (dip == NULL) 1107 return (NDI_FAILURE); 1108 1109 mutex_enter(&(DEVI(dip)->devi_lock)); 1110 1111 switch (state) { 1112 case BUS_QUIESCED: 1113 DEVI_SET_BUS_QUIESCE(dip); 1114 break; 1115 1116 case BUS_ACTIVE: 1117 DEVI_SET_BUS_ACTIVE(dip); 1118 DEVI_SET_BUS_UP(dip); 1119 break; 1120 1121 case BUS_SHUTDOWN: 1122 DEVI_SET_BUS_DOWN(dip); 1123 break; 1124 1125 default: 1126 rv = NDI_FAILURE; 1127 } 1128 1129 mutex_exit(&(DEVI(dip)->devi_lock)); 1130 return (rv); 1131 } 1132 1133 /* 1134 * These dummy functions are obsolete and may be removed. 1135 * Retained for existing driver compatibility only. 1136 * Drivers should be fixed not to use these functions. 1137 * Don't write new code using these obsolete interfaces. 1138 */ 1139 /*ARGSUSED*/ 1140 void 1141 i_ndi_block_device_tree_changes(uint_t *lkcnt) /* obsolete */ 1142 { 1143 /* obsolete dummy function */ 1144 } 1145 1146 /*ARGSUSED*/ 1147 void 1148 i_ndi_allow_device_tree_changes(uint_t lkcnt) /* obsolete */ 1149 { 1150 /* obsolete dummy function */ 1151 } 1152 1153 /* 1154 * Single thread entry into per-driver list 1155 */ 1156 /*ARGSUSED*/ 1157 void 1158 e_ddi_enter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */ 1159 { 1160 /* obsolete dummy function */ 1161 } 1162 1163 /* 1164 * release the per-driver list 1165 */ 1166 /*ARGSUSED*/ 1167 void 1168 e_ddi_exit_driver_list(struct devnames *dnp, int listcnt) /* obsolete */ 1169 { 1170 /* obsolete dummy function */ 1171 } 1172 1173 /* 1174 * Attempt to enter driver list 1175 */ 1176 /*ARGSUSED*/ 1177 int 1178 e_ddi_tryenter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */ 1179 { 1180 return (1); /* obsolete dummy function */ 1181 } 1182 1183 /* 1184 * ndi event handling support functions: 1185 * The NDI event support model is as follows: 1186 * 1187 * The nexus driver defines a set of events using some static structures (so 1188 * these structures can be shared by all instances of the nexus driver). 1189 * The nexus driver allocates an event handle and binds the event set 1190 * to this handle. The nexus driver's event busop functions can just 1191 * call the appropriate NDI event support function using this handle 1192 * as the first argument. 1193 * 1194 * The reasoning for tying events to the device tree is that the entity 1195 * generating the callback will typically be one of the device driver's 1196 * ancestors in the tree. 1197 */ 1198 static int ndi_event_debug = 0; 1199 1200 #ifdef DEBUG 1201 #define NDI_EVENT_DEBUG ndi_event_debug 1202 #endif /* DEBUG */ 1203 1204 /* 1205 * allocate a new ndi event handle 1206 */ 1207 int 1208 ndi_event_alloc_hdl(dev_info_t *dip, ddi_iblock_cookie_t cookie, 1209 ndi_event_hdl_t *handle, uint_t flag) 1210 { 1211 struct ndi_event_hdl *ndi_event_hdl; 1212 1213 ndi_event_hdl = kmem_zalloc(sizeof (struct ndi_event_hdl), 1214 ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP)); 1215 1216 if (!ndi_event_hdl) { 1217 return (NDI_FAILURE); 1218 } 1219 1220 ndi_event_hdl->ndi_evthdl_dip = dip; 1221 ndi_event_hdl->ndi_evthdl_iblock_cookie = cookie; 1222 mutex_init(&ndi_event_hdl->ndi_evthdl_mutex, NULL, 1223 MUTEX_DRIVER, (void *)cookie); 1224 1225 mutex_init(&ndi_event_hdl->ndi_evthdl_cb_mutex, NULL, 1226 MUTEX_DRIVER, (void *)cookie); 1227 1228 *handle = (ndi_event_hdl_t)ndi_event_hdl; 1229 1230 return (NDI_SUCCESS); 1231 } 1232 1233 /* 1234 * free the ndi event handle 1235 */ 1236 int 1237 ndi_event_free_hdl(ndi_event_hdl_t handle) 1238 { 1239 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1240 ndi_event_cookie_t *cookie; 1241 ndi_event_cookie_t *free; 1242 1243 ASSERT(handle); 1244 1245 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1246 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1247 1248 cookie = ndi_event_hdl->ndi_evthdl_cookie_list; 1249 1250 /* deallocate all defined cookies */ 1251 while (cookie != NULL) { 1252 ASSERT(cookie->callback_list == NULL); 1253 free = cookie; 1254 cookie = cookie->next_cookie; 1255 1256 kmem_free(free, sizeof (ndi_event_cookie_t)); 1257 } 1258 1259 1260 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1261 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1262 1263 /* destroy mutexes */ 1264 mutex_destroy(&ndi_event_hdl->ndi_evthdl_mutex); 1265 mutex_destroy(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1266 1267 /* free event handle */ 1268 kmem_free(ndi_event_hdl, sizeof (struct ndi_event_hdl)); 1269 1270 return (NDI_SUCCESS); 1271 } 1272 1273 1274 /* 1275 * ndi_event_bind_set() adds a set of events to the NDI event 1276 * handle. 1277 * 1278 * Events generated by high level interrupts should not 1279 * be mixed in the same event set with events generated by 1280 * normal interrupts or kernel events. 1281 * 1282 * This function can be called multiple times to bind 1283 * additional sets to the event handle. 1284 * However, events generated by high level interrupts cannot 1285 * be bound to a handle that already has bound events generated 1286 * by normal interrupts or from kernel context and vice versa. 1287 */ 1288 int 1289 ndi_event_bind_set(ndi_event_hdl_t handle, 1290 ndi_event_set_t *ndi_events, 1291 uint_t flag) 1292 { 1293 struct ndi_event_hdl *ndi_event_hdl; 1294 ndi_event_cookie_t *next, *prev, *new_cookie; 1295 uint_t i, len; 1296 uint_t dup = 0; 1297 uint_t high_plevels, other_plevels; 1298 ndi_event_definition_t *ndi_event_defs; 1299 1300 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP); 1301 1302 ASSERT(handle); 1303 ASSERT(ndi_events); 1304 1305 /* 1306 * binding must be performed during attach/detach 1307 */ 1308 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) && 1309 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) { 1310 cmn_err(CE_WARN, "ndi_event_bind_set must be called within " 1311 "attach or detach"); 1312 return (NDI_FAILURE); 1313 } 1314 1315 /* 1316 * if it is not the correct version or the event set is 1317 * empty, bail out 1318 */ 1319 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1) 1320 return (NDI_FAILURE); 1321 1322 ndi_event_hdl = (struct ndi_event_hdl *)handle; 1323 ndi_event_defs = ndi_events->ndi_event_defs; 1324 high_plevels = other_plevels = 0; 1325 1326 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1327 1328 /* check for mixing events at high level with the other types */ 1329 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1330 if (ndi_event_defs[i].ndi_event_plevel == EPL_HIGHLEVEL) { 1331 high_plevels++; 1332 } else { 1333 other_plevels++; 1334 } 1335 } 1336 1337 /* 1338 * bail out if high level events are mixed with other types in this 1339 * event set or the set is incompatible with the set in the handle 1340 */ 1341 if ((high_plevels && other_plevels) || 1342 (other_plevels && ndi_event_hdl->ndi_evthdl_high_plevels) || 1343 (high_plevels && ndi_event_hdl->ndi_evthdl_other_plevels)) { 1344 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1345 1346 return (NDI_FAILURE); 1347 } 1348 1349 /* 1350 * check for duplicate events in both the existing handle 1351 * and the event set, add events if not duplicates 1352 */ 1353 next = ndi_event_hdl->ndi_evthdl_cookie_list; 1354 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1355 while (next != NULL) { 1356 len = strlen(NDI_EVENT_NAME(next)) + 1; 1357 if (strncmp(NDI_EVENT_NAME(next), 1358 ndi_event_defs[i].ndi_event_name, len) == 0) { 1359 dup = 1; 1360 break; 1361 } 1362 1363 prev = next; 1364 next = next->next_cookie; 1365 } 1366 1367 if (dup == 0) { 1368 new_cookie = kmem_zalloc(sizeof (ndi_event_cookie_t), 1369 km_flag); 1370 1371 if (!new_cookie) 1372 return (NDI_FAILURE); 1373 1374 if (ndi_event_hdl->ndi_evthdl_n_events == 0) { 1375 ndi_event_hdl->ndi_evthdl_cookie_list = 1376 new_cookie; 1377 } else { 1378 prev->next_cookie = new_cookie; 1379 } 1380 1381 ndi_event_hdl->ndi_evthdl_n_events++; 1382 1383 /* 1384 * set up new cookie 1385 */ 1386 new_cookie->definition = &ndi_event_defs[i]; 1387 new_cookie->ddip = ndi_event_hdl->ndi_evthdl_dip; 1388 1389 } else { 1390 /* 1391 * event not added, must correct plevel numbers 1392 */ 1393 if (ndi_event_defs[i].ndi_event_plevel == 1394 EPL_HIGHLEVEL) { 1395 high_plevels--; 1396 } else { 1397 other_plevels--; 1398 } 1399 } 1400 1401 dup = 0; 1402 next = ndi_event_hdl->ndi_evthdl_cookie_list; 1403 prev = NULL; 1404 1405 } 1406 1407 ndi_event_hdl->ndi_evthdl_high_plevels += high_plevels; 1408 ndi_event_hdl->ndi_evthdl_other_plevels += other_plevels; 1409 1410 ASSERT((ndi_event_hdl->ndi_evthdl_high_plevels == 0) || 1411 (ndi_event_hdl->ndi_evthdl_other_plevels == 0)); 1412 1413 #ifdef NDI_EVENT_DEBUG 1414 if (ndi_event_debug) { 1415 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_bind_set"); 1416 } 1417 #endif /* NDI_EVENT_DEBUG */ 1418 1419 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1420 1421 return (NDI_SUCCESS); 1422 } 1423 1424 /* 1425 * ndi_event_unbind_set() unbinds a set of events, previously 1426 * bound using ndi_event_bind_set(), from the NDI event 1427 * handle. 1428 * 1429 * This routine will unbind all events in the event set. If an event, 1430 * specified in the event set, is not found in the handle, this 1431 * routine will proceed onto the next member of the set as if the event 1432 * was never specified. 1433 * 1434 * The event set may be a subset of the set of events that 1435 * was previously bound to the handle. For example, events 1436 * can be individually unbound. 1437 * 1438 * An event cannot be unbound if callbacks are still 1439 * registered against the event. 1440 */ 1441 /*ARGSUSED*/ 1442 int 1443 ndi_event_unbind_set(ndi_event_hdl_t handle, ndi_event_set_t *ndi_events, 1444 uint_t flag) 1445 { 1446 ndi_event_definition_t *ndi_event_defs; 1447 int len; 1448 uint_t i; 1449 int rval; 1450 ndi_event_cookie_t *cookie_list; 1451 ndi_event_cookie_t *prev = NULL; 1452 1453 ASSERT(ndi_events); 1454 ASSERT(handle); 1455 1456 /* 1457 * binding must be performed during attach/detac 1458 */ 1459 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) && 1460 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) { 1461 cmn_err(CE_WARN, "ndi_event_bind_set must be called within " 1462 "attach or detach"); 1463 return (NDI_FAILURE); 1464 } 1465 1466 /* bail out if ndi_event_set is outdated */ 1467 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1) { 1468 return (NDI_FAILURE); 1469 } 1470 1471 ASSERT(ndi_events->ndi_event_defs); 1472 1473 ndi_event_defs = ndi_events->ndi_event_defs; 1474 1475 mutex_enter(&handle->ndi_evthdl_mutex); 1476 mutex_enter(&handle->ndi_evthdl_cb_mutex); 1477 1478 /* 1479 * Verify that all events in the event set are eligible 1480 * for unbinding(ie. there are no outstanding callbacks). 1481 * If any one of the events are ineligible, fail entire 1482 * operation. 1483 */ 1484 1485 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1486 cookie_list = handle->ndi_evthdl_cookie_list; 1487 while (cookie_list != NULL) { 1488 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1489 if (strncmp(NDI_EVENT_NAME(cookie_list), 1490 ndi_event_defs[i].ndi_event_name, len) == 0) { 1491 1492 ASSERT(cookie_list->callback_list == NULL); 1493 if (cookie_list->callback_list) { 1494 rval = NDI_FAILURE; 1495 goto done; 1496 } 1497 break; 1498 } else { 1499 cookie_list = cookie_list->next_cookie; 1500 } 1501 } 1502 } 1503 1504 /* 1505 * remove all events found within the handle 1506 * If an event is not found, this function will proceed as if the event 1507 * was never specified. 1508 */ 1509 1510 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1511 cookie_list = handle->ndi_evthdl_cookie_list; 1512 prev = NULL; 1513 while (cookie_list != NULL) { 1514 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1515 if (strncmp(NDI_EVENT_NAME(cookie_list), 1516 ndi_event_defs[i].ndi_event_name, len) == 0) { 1517 1518 /* 1519 * can not unbind an event definition with 1520 * outstanding callbacks 1521 */ 1522 if (cookie_list->callback_list) { 1523 rval = NDI_FAILURE; 1524 goto done; 1525 } 1526 1527 /* remove this cookie from the list */ 1528 if (prev != NULL) { 1529 prev->next_cookie = 1530 cookie_list->next_cookie; 1531 } else { 1532 handle->ndi_evthdl_cookie_list = 1533 cookie_list->next_cookie; 1534 } 1535 1536 /* adjust plevel counts */ 1537 if (NDI_EVENT_PLEVEL(cookie_list) == 1538 EPL_HIGHLEVEL) { 1539 handle->ndi_evthdl_high_plevels--; 1540 } else { 1541 handle->ndi_evthdl_other_plevels--; 1542 } 1543 1544 /* adjust cookie count */ 1545 handle->ndi_evthdl_n_events--; 1546 1547 /* free the cookie */ 1548 kmem_free(cookie_list, 1549 sizeof (ndi_event_cookie_t)); 1550 1551 cookie_list = handle->ndi_evthdl_cookie_list; 1552 break; 1553 1554 } else { 1555 prev = cookie_list; 1556 cookie_list = cookie_list->next_cookie; 1557 } 1558 1559 } 1560 1561 } 1562 1563 #ifdef NDI_EVENT_DEBUG 1564 if (ndi_event_debug) { 1565 ndi_event_dump_hdl(handle, "ndi_event_unbind_set"); 1566 } 1567 #endif /* NDI_EVENT_DEBUG */ 1568 1569 rval = NDI_SUCCESS; 1570 1571 done: 1572 mutex_exit(&handle->ndi_evthdl_cb_mutex); 1573 mutex_exit(&handle->ndi_evthdl_mutex); 1574 1575 return (rval); 1576 } 1577 1578 /* 1579 * ndi_event_retrieve_cookie(): 1580 * Return an event cookie for eventname if this nexus driver 1581 * has defined the named event. The event cookie returned 1582 * by this function is used to register callback handlers 1583 * for the event. 1584 * 1585 * ndi_event_retrieve_cookie() is intended to be used in the 1586 * nexus driver's bus_get_eventcookie busop routine. 1587 * 1588 * If the event is not defined by this bus nexus driver, and flag 1589 * does not include NDI_EVENT_NOPASS, then ndi_event_retrieve_cookie() 1590 * will pass the request up the device tree hierarchy by calling 1591 * ndi_busop_get_eventcookie(9N). 1592 * If the event is not defined by this bus nexus driver, and flag 1593 * does include NDI_EVENT_NOPASS, ndi_event_retrieve_cookie() 1594 * will return NDI_FAILURE. The caller may then determine what further 1595 * action to take, such as using a different handle, passing the 1596 * request up the device tree using ndi_busop_get_eventcookie(9N), 1597 * or returning the failure to the caller, thus blocking the 1598 * progress of the request up the tree. 1599 */ 1600 int 1601 ndi_event_retrieve_cookie(ndi_event_hdl_t handle, 1602 dev_info_t *rdip, 1603 char *eventname, 1604 ddi_eventcookie_t *cookiep, 1605 uint_t flag) 1606 { 1607 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1608 int len; 1609 ndi_event_cookie_t *cookie_list; 1610 1611 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1612 1613 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list; 1614 /* 1615 * search the cookie list for the event name and return 1616 * cookie if found. 1617 */ 1618 while (cookie_list != NULL) { 1619 1620 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1621 if (strncmp(NDI_EVENT_NAME(cookie_list), eventname, 1622 len) == 0) { 1623 *cookiep = (ddi_eventcookie_t)cookie_list; 1624 1625 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1626 return (NDI_SUCCESS); 1627 } 1628 1629 cookie_list = cookie_list->next_cookie; 1630 } 1631 1632 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1633 /* 1634 * event was not found, pass up or return failure 1635 */ 1636 if ((flag & NDI_EVENT_NOPASS) == 0) { 1637 return (ndi_busop_get_eventcookie( 1638 ndi_event_hdl->ndi_evthdl_dip, rdip, eventname, cookiep)); 1639 } else { 1640 return (NDI_FAILURE); 1641 } 1642 } 1643 1644 /* 1645 * check whether this nexus defined this event and look up attributes 1646 */ 1647 static int 1648 ndi_event_is_defined(ndi_event_hdl_t handle, 1649 ddi_eventcookie_t cookie, int *attributes) 1650 { 1651 1652 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1653 ndi_event_cookie_t *cookie_list; 1654 1655 ASSERT(mutex_owned(&handle->ndi_evthdl_mutex)); 1656 1657 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list; 1658 while (cookie_list != NULL) { 1659 if (cookie_list == NDI_EVENT(cookie)) { 1660 if (attributes) 1661 *attributes = 1662 NDI_EVENT_ATTRIBUTES(cookie_list); 1663 1664 return (NDI_SUCCESS); 1665 } 1666 1667 cookie_list = cookie_list->next_cookie; 1668 } 1669 1670 return (NDI_FAILURE); 1671 } 1672 1673 /* 1674 * ndi_event_add_callback(): adds an event callback registration 1675 * to the event cookie defining this event. 1676 * 1677 * Refer also to bus_add_eventcall(9n) and ndi_busop_add_eventcall(9n). 1678 * 1679 * ndi_event_add_callback(9n) is intended to be used in 1680 * the nexus driver's bus_add_eventcall(9n) busop function. 1681 * 1682 * If the event is not defined by this bus nexus driver, 1683 * ndi_event_add_callback() will return NDI_FAILURE. 1684 */ 1685 int 1686 ndi_event_add_callback(ndi_event_hdl_t handle, dev_info_t *child_dip, 1687 ddi_eventcookie_t cookie, 1688 void (*event_callback)(dev_info_t *, 1689 ddi_eventcookie_t, void *arg, void *impldata), 1690 void *arg, 1691 uint_t flag, 1692 ddi_callback_id_t *cb_id) 1693 { 1694 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1695 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP); 1696 ndi_event_callbacks_t *cb; 1697 1698 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1699 1700 /* 1701 * if the event was not bound to this handle, return failure 1702 */ 1703 if (ndi_event_is_defined(handle, cookie, NULL) != NDI_SUCCESS) { 1704 1705 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1706 return (NDI_FAILURE); 1707 1708 } 1709 1710 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1711 1712 /* 1713 * allocate space for a callback structure 1714 */ 1715 cb = kmem_zalloc(sizeof (ndi_event_callbacks_t), km_flag); 1716 if (cb == NULL) { 1717 return (NDI_FAILURE); 1718 } 1719 1720 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1721 1722 /* initialize callback structure */ 1723 cb->ndi_evtcb_dip = child_dip; 1724 cb->ndi_evtcb_callback = event_callback; 1725 cb->ndi_evtcb_arg = arg; 1726 cb->ndi_evtcb_cookie = cookie; 1727 cb->devname = (char *)ddi_driver_name(child_dip); 1728 1729 *cb_id = (ddi_callback_id_t)cb; 1730 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1731 1732 /* add this callback structure to the list */ 1733 if (NDI_EVENT(cookie)->callback_list) { 1734 cb->ndi_evtcb_next = NDI_EVENT(cookie)->callback_list; 1735 NDI_EVENT(cookie)->callback_list->ndi_evtcb_prev = cb; 1736 NDI_EVENT(cookie)->callback_list = cb; 1737 } else { 1738 NDI_EVENT(cookie)->callback_list = cb; 1739 } 1740 #ifdef NDI_EVENT_DEBUG 1741 if (ndi_event_debug) { 1742 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_add_callback"); 1743 } 1744 #endif /* NDI_EVENT_DEBUG */ 1745 1746 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1747 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1748 1749 return (NDI_SUCCESS); 1750 } 1751 1752 /* 1753 * ndi_event_remove_callback(): 1754 * 1755 * ndi_event_remove_callback() removes a callback that was 1756 * previously registered using ndi_event_add_callback(9N). 1757 * Refer also to bus_remove_eventcall(9n) and 1758 * ndi_busop_remove_eventcall(9n). 1759 * ndi_event_remove_callback(9n) is intended to be used in 1760 * the nexus driver's bus_remove_eventcall (9n) busop function. 1761 * If the event is not defined by this bus nexus driver, 1762 * ndi_event_remove_callback() will return NDI_FAILURE. 1763 */ 1764 static void do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl, 1765 ddi_callback_id_t cb_id); 1766 1767 int 1768 ndi_event_remove_callback(ndi_event_hdl_t handle, ddi_callback_id_t cb_id) 1769 { 1770 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1771 1772 ASSERT(cb_id); 1773 1774 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1775 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1776 1777 do_ndi_event_remove_callback(ndi_event_hdl, cb_id); 1778 1779 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1780 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1781 1782 return (NDI_SUCCESS); 1783 } 1784 1785 /*ARGSUSED*/ 1786 static void 1787 do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl, 1788 ddi_callback_id_t cb_id) 1789 { 1790 ndi_event_callbacks_t *cb = (ndi_event_callbacks_t *)cb_id; 1791 ASSERT(cb); 1792 1793 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_mutex)); 1794 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_cb_mutex)); 1795 1796 /* remove from callback linked list */ 1797 if (cb->ndi_evtcb_prev) { 1798 cb->ndi_evtcb_prev->ndi_evtcb_next = cb->ndi_evtcb_next; 1799 } 1800 1801 if (cb->ndi_evtcb_next) { 1802 cb->ndi_evtcb_next->ndi_evtcb_prev = cb->ndi_evtcb_prev; 1803 } 1804 1805 if (NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list == cb) { 1806 NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list = 1807 cb->ndi_evtcb_next; 1808 } 1809 1810 kmem_free(cb, sizeof (ndi_event_callbacks_t)); 1811 } 1812 1813 /* 1814 * ndi_event_run_callbacks() performs event callbacks for the event 1815 * specified by cookie, if this is among those bound to the 1816 * supplied handle. 1817 * If the event is among those bound to the handle, none, 1818 * some, or all of the handlers registered for the event 1819 * will be called, according to the delivery attributes of 1820 * the event. 1821 * If the event attributes include NDI_EVENT_POST_TO_ALL 1822 * (the default), all the handlers for the event will be 1823 * called in an unspecified order. 1824 * If the event attributes include NDI_EVENT_POST_TO_TGT, only 1825 * the handlers (if any) registered by the driver identified by 1826 * rdip will be called. 1827 * If the event identified by cookie is not bound to the handle, 1828 * NDI_FAILURE will be returned. 1829 */ 1830 int 1831 ndi_event_run_callbacks(ndi_event_hdl_t handle, dev_info_t *child_dip, 1832 ddi_eventcookie_t cookie, void *bus_impldata) 1833 { 1834 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1835 ndi_event_callbacks_t *next, *cb; 1836 int attributes; 1837 1838 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1839 1840 /* if this is not our event, fail */ 1841 if (ndi_event_is_defined(handle, cookie, &attributes) != 1842 NDI_SUCCESS) { 1843 1844 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1845 return (NDI_FAILURE); 1846 } 1847 1848 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1849 1850 #ifdef NDI_EVENT_DEBUG 1851 if (ndi_event_debug) { 1852 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t" 1853 "producer dip=%p (%s%d): cookie = %p, name = %s\n", 1854 (void *)ndi_event_hdl->ndi_evthdl_dip, 1855 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip), 1856 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip), 1857 (void *)cookie, 1858 ndi_event_cookie_to_name(handle, cookie)); 1859 } 1860 #endif /* #ifdef NDI_EVENT_DEBUG */ 1861 1862 1863 /* 1864 * The callback handlers may call conversion functions. The conversion 1865 * functions may hold the ndi_evthdl_mutex during execution. Thus, to 1866 * avoid a recursive mutex problem, only the ndi_evthdl_cb_mutex is 1867 * held. The ndi_evthdl_mutex is not held when running the callbacks. 1868 */ 1869 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1870 1871 /* perform callbacks */ 1872 next = NDI_EVENT(cookie)->callback_list; 1873 while (next != NULL) { 1874 1875 cb = next; 1876 next = next->ndi_evtcb_next; 1877 1878 ASSERT(cb->ndi_evtcb_cookie == cookie); 1879 1880 if (attributes == NDI_EVENT_POST_TO_TGT && 1881 child_dip != cb->ndi_evtcb_dip) { 1882 continue; 1883 } 1884 1885 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip, cb->ndi_evtcb_cookie, 1886 cb->ndi_evtcb_arg, bus_impldata); 1887 1888 #ifdef NDI_EVENT_DEBUG 1889 if (ndi_event_debug) { 1890 cmn_err(CE_CONT, 1891 "\t\tconsumer dip=%p (%s%d)\n", 1892 (void *)cb->ndi_evtcb_dip, 1893 ddi_node_name(cb->ndi_evtcb_dip), 1894 ddi_get_instance(cb->ndi_evtcb_dip)); 1895 } 1896 #endif 1897 1898 } 1899 1900 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1901 1902 #ifdef NDI_EVENT_DEBUG 1903 if (ndi_event_debug) { 1904 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1905 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks"); 1906 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1907 } 1908 #endif /* NDI_EVENT_DEBUG */ 1909 1910 return (NDI_SUCCESS); 1911 } 1912 1913 1914 /* 1915 * perform one callback for a specified cookie and just one target 1916 */ 1917 int 1918 ndi_event_do_callback(ndi_event_hdl_t handle, dev_info_t *child_dip, 1919 ddi_eventcookie_t cookie, void *bus_impldata) 1920 { 1921 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1922 ndi_event_callbacks_t *next, *cb; 1923 int attributes; 1924 1925 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1926 1927 /* if this is not our event, fail */ 1928 if (ndi_event_is_defined(handle, cookie, &attributes) != 1929 NDI_SUCCESS) { 1930 1931 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1932 1933 return (NDI_FAILURE); 1934 } 1935 1936 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1937 1938 #ifdef NDI_EVENT_DEBUG 1939 if (ndi_event_debug) { 1940 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t" 1941 "producer dip=%p (%s%d): cookie = %p, name = %s\n", 1942 (void *)ndi_event_hdl->ndi_evthdl_dip, 1943 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip), 1944 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip), 1945 (void *)cookie, 1946 ndi_event_cookie_to_name(handle, cookie)); 1947 } 1948 #endif 1949 1950 1951 /* 1952 * we only grab the cb mutex because the callback handlers 1953 * may call the conversion functions which would cause a recursive 1954 * mutex problem 1955 */ 1956 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1957 1958 /* perform callbacks */ 1959 for (next = NDI_EVENT(cookie)->callback_list; next != NULL; ) { 1960 cb = next; 1961 next = next->ndi_evtcb_next; 1962 1963 if (cb->ndi_evtcb_dip == child_dip) { 1964 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip, 1965 cb->ndi_evtcb_cookie, cb->ndi_evtcb_arg, 1966 bus_impldata); 1967 1968 #ifdef NDI_EVENT_DEBUG 1969 if (ndi_event_debug) { 1970 cmn_err(CE_CONT, 1971 "\t\tconsumer dip=%p (%s%d)\n", 1972 (void *)cb->ndi_evtcb_dip, 1973 ddi_node_name(cb->ndi_evtcb_dip), 1974 ddi_get_instance(cb->ndi_evtcb_dip)); 1975 } 1976 #endif 1977 break; 1978 } 1979 } 1980 1981 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1982 1983 #ifdef NDI_EVENT_DEBUG 1984 if (ndi_event_debug) { 1985 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1986 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks"); 1987 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1988 } 1989 #endif /* NDI_EVENT_DEBUG */ 1990 1991 return (NDI_SUCCESS); 1992 } 1993 1994 1995 /* 1996 * ndi_event_tag_to_cookie: utility function to find an event cookie 1997 * given an event tag 1998 */ 1999 ddi_eventcookie_t 2000 ndi_event_tag_to_cookie(ndi_event_hdl_t handle, int event_tag) 2001 { 2002 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2003 ndi_event_cookie_t *list; 2004 2005 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2006 2007 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2008 while (list != NULL) { 2009 if (NDI_EVENT_TAG(list) == event_tag) { 2010 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2011 return ((ddi_eventcookie_t)list); 2012 } 2013 2014 list = list->next_cookie; 2015 } 2016 2017 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2018 return (NULL); 2019 } 2020 2021 /* 2022 * ndi_event_cookie_to_tag: utility function to find a event tag 2023 * given an event_cookie 2024 */ 2025 int 2026 ndi_event_cookie_to_tag(ndi_event_hdl_t handle, ddi_eventcookie_t cookie) 2027 { 2028 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2029 ndi_event_cookie_t *list; 2030 2031 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2032 2033 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2034 2035 while (list != NULL) { 2036 if ((ddi_eventcookie_t)list == cookie) { 2037 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2038 return (NDI_EVENT_TAG(list)); 2039 } 2040 2041 list = list->next_cookie; 2042 } 2043 2044 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2045 return (NDI_FAILURE); 2046 2047 } 2048 2049 /* 2050 * ndi_event_cookie_to_name: utility function to find an event name 2051 * given an event_cookie 2052 */ 2053 char * 2054 ndi_event_cookie_to_name(ndi_event_hdl_t handle, ddi_eventcookie_t cookie) 2055 { 2056 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2057 ndi_event_cookie_t *list; 2058 2059 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2060 2061 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2062 2063 while (list != NULL) { 2064 if (list == NDI_EVENT(cookie)) { 2065 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2066 return (NDI_EVENT_NAME(list)); 2067 } 2068 2069 list = list->next_cookie; 2070 } 2071 2072 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2073 return (NULL); 2074 } 2075 2076 /* 2077 * ndi_event_tag_to_name: utility function to find an event name 2078 * given an event tag 2079 */ 2080 char * 2081 ndi_event_tag_to_name(ndi_event_hdl_t handle, int event_tag) 2082 { 2083 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2084 ndi_event_cookie_t *list; 2085 2086 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2087 2088 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2089 2090 while (list) { 2091 if (NDI_EVENT_TAG(list) == event_tag) { 2092 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2093 return (NDI_EVENT_NAME(list)); 2094 } 2095 2096 list = list->next_cookie; 2097 } 2098 2099 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2100 2101 return (NULL); 2102 } 2103 2104 #ifdef NDI_EVENT_DEBUG 2105 void 2106 ndi_event_dump_hdl(struct ndi_event_hdl *hdl, char *location) 2107 { 2108 2109 2110 ndi_event_callbacks_t *next; 2111 ndi_event_cookie_t *list; 2112 2113 ASSERT(mutex_owned(&hdl->ndi_evthdl_mutex)); 2114 list = hdl->ndi_evthdl_cookie_list; 2115 2116 cmn_err(CE_CONT, "%s: event handle (%p): dip = %p (%s%d)\n", 2117 location, (void *)hdl, (void *)hdl->ndi_evthdl_dip, 2118 ddi_node_name(hdl->ndi_evthdl_dip), 2119 ddi_get_instance(hdl->ndi_evthdl_dip)); 2120 cmn_err(CE_CONT, "\thigh=%d other=%d n=%d\n", 2121 hdl->ndi_evthdl_high_plevels, hdl->ndi_evthdl_other_plevels, 2122 hdl->ndi_evthdl_n_events); 2123 2124 cmn_err(CE_CONT, "\tevent cookies:\n"); 2125 while (list) { 2126 cmn_err(CE_CONT, "\t\ttag=%d name=%s p=%d a=%x dd=%p\n", 2127 NDI_EVENT_TAG(list), NDI_EVENT_NAME(list), 2128 NDI_EVENT_PLEVEL(list), NDI_EVENT_ATTRIBUTES(list), 2129 (void *)NDI_EVENT_DDIP(list)); 2130 cmn_err(CE_CONT, "\t\tcallbacks:\n"); 2131 for (next = list->callback_list; next != NULL; 2132 next = next->ndi_evtcb_next) { 2133 cmn_err(CE_CONT, 2134 "\t\t dip=%p (%s%d) cookie=%p arg=%p\n", 2135 (void*)next->ndi_evtcb_dip, 2136 ddi_driver_name(next->ndi_evtcb_dip), 2137 ddi_get_instance(next->ndi_evtcb_dip), 2138 (void *)next->ndi_evtcb_cookie, 2139 next->ndi_evtcb_arg); 2140 } 2141 2142 list = list->next_cookie; 2143 } 2144 2145 cmn_err(CE_CONT, "\n"); 2146 } 2147 #endif 2148 2149 int 2150 ndi_dev_is_prom_node(dev_info_t *dip) 2151 { 2152 return (DEVI(dip)->devi_node_class == DDI_NC_PROM); 2153 } 2154 2155 int 2156 ndi_dev_is_pseudo_node(dev_info_t *dip) 2157 { 2158 /* 2159 * NOTE: this does NOT mean the pseudo branch of the device tree, 2160 * it means the node was created by software (DEVI_SID_NODEID | 2161 * DEVI_PSEUDO_NODEID) instead of being generated from a PROM node. 2162 */ 2163 return (DEVI(dip)->devi_node_class == DDI_NC_PSEUDO); 2164 } 2165 2166 int 2167 ndi_dev_is_persistent_node(dev_info_t *dip) 2168 { 2169 return ((DEVI(dip)->devi_node_attributes & DDI_PERSISTENT) != 0); 2170 } 2171 2172 int 2173 i_ndi_dev_is_auto_assigned_node(dev_info_t *dip) 2174 { 2175 return ((DEVI(dip)->devi_node_attributes & 2176 DDI_AUTO_ASSIGNED_NODEID) != 0); 2177 } 2178 2179 void 2180 i_ndi_set_node_class(dev_info_t *dip, ddi_node_class_t c) 2181 { 2182 DEVI(dip)->devi_node_class = c; 2183 } 2184 2185 ddi_node_class_t 2186 i_ndi_get_node_class(dev_info_t *dip) 2187 { 2188 return (DEVI(dip)->devi_node_class); 2189 } 2190 2191 void 2192 i_ndi_set_node_attributes(dev_info_t *dip, int p) 2193 { 2194 DEVI(dip)->devi_node_attributes = p; 2195 } 2196 2197 int 2198 i_ndi_get_node_attributes(dev_info_t *dip) 2199 { 2200 return (DEVI(dip)->devi_node_attributes); 2201 } 2202 2203 void 2204 i_ndi_set_nodeid(dev_info_t *dip, int n) 2205 { 2206 DEVI(dip)->devi_nodeid = n; 2207 } 2208 2209 void 2210 ndi_set_acc_fault(ddi_acc_handle_t ah) 2211 { 2212 i_ddi_acc_set_fault(ah); 2213 } 2214 2215 void 2216 ndi_clr_acc_fault(ddi_acc_handle_t ah) 2217 { 2218 i_ddi_acc_clr_fault(ah); 2219 } 2220 2221 void 2222 ndi_set_dma_fault(ddi_dma_handle_t dh) 2223 { 2224 i_ddi_dma_set_fault(dh); 2225 } 2226 2227 void 2228 ndi_clr_dma_fault(ddi_dma_handle_t dh) 2229 { 2230 i_ddi_dma_clr_fault(dh); 2231 } 2232 2233 /* 2234 * The default fault-handler, called when the event posted by 2235 * ddi_dev_report_fault() reaches rootnex. 2236 */ 2237 static void 2238 i_ddi_fault_handler(dev_info_t *dip, struct ddi_fault_event_data *fedp) 2239 { 2240 ASSERT(fedp); 2241 2242 mutex_enter(&(DEVI(dip)->devi_lock)); 2243 if (!DEVI_IS_DEVICE_OFFLINE(dip)) { 2244 switch (fedp->f_impact) { 2245 case DDI_SERVICE_LOST: 2246 DEVI_SET_DEVICE_DOWN(dip); 2247 break; 2248 2249 case DDI_SERVICE_DEGRADED: 2250 DEVI_SET_DEVICE_DEGRADED(dip); 2251 break; 2252 2253 case DDI_SERVICE_UNAFFECTED: 2254 default: 2255 break; 2256 2257 case DDI_SERVICE_RESTORED: 2258 DEVI_SET_DEVICE_UP(dip); 2259 break; 2260 } 2261 } 2262 mutex_exit(&(DEVI(dip)->devi_lock)); 2263 } 2264 2265 /* 2266 * The default fault-logger, called when the event posted by 2267 * ddi_dev_report_fault() reaches rootnex. 2268 */ 2269 /*ARGSUSED*/ 2270 static void 2271 i_ddi_fault_logger(dev_info_t *rdip, struct ddi_fault_event_data *fedp) 2272 { 2273 ddi_devstate_t newstate; 2274 const char *action; 2275 const char *servstate; 2276 const char *location; 2277 int bad; 2278 int changed; 2279 int level; 2280 int still; 2281 2282 ASSERT(fedp); 2283 2284 bad = 0; 2285 switch (fedp->f_location) { 2286 case DDI_DATAPATH_FAULT: 2287 location = "in datapath to"; 2288 break; 2289 case DDI_DEVICE_FAULT: 2290 location = "in"; 2291 break; 2292 case DDI_EXTERNAL_FAULT: 2293 location = "external to"; 2294 break; 2295 default: 2296 location = "somewhere near"; 2297 bad = 1; 2298 break; 2299 } 2300 2301 newstate = ddi_get_devstate(fedp->f_dip); 2302 switch (newstate) { 2303 case DDI_DEVSTATE_OFFLINE: 2304 servstate = "unavailable"; 2305 break; 2306 case DDI_DEVSTATE_DOWN: 2307 servstate = "unavailable"; 2308 break; 2309 case DDI_DEVSTATE_QUIESCED: 2310 servstate = "suspended"; 2311 break; 2312 case DDI_DEVSTATE_DEGRADED: 2313 servstate = "degraded"; 2314 break; 2315 default: 2316 servstate = "available"; 2317 break; 2318 } 2319 2320 changed = (newstate != fedp->f_oldstate); 2321 level = (newstate < fedp->f_oldstate) ? CE_WARN : CE_NOTE; 2322 switch (fedp->f_impact) { 2323 case DDI_SERVICE_LOST: 2324 case DDI_SERVICE_DEGRADED: 2325 case DDI_SERVICE_UNAFFECTED: 2326 /* fault detected; service [still] <servstate> */ 2327 action = "fault detected"; 2328 still = !changed; 2329 break; 2330 2331 case DDI_SERVICE_RESTORED: 2332 if (newstate != DDI_DEVSTATE_UP) { 2333 /* fault cleared; service still <servstate> */ 2334 action = "fault cleared"; 2335 still = 1; 2336 } else if (changed) { 2337 /* fault cleared; service <servstate> */ 2338 action = "fault cleared"; 2339 still = 0; 2340 } else { 2341 /* no fault; service <servstate> */ 2342 action = "no fault"; 2343 still = 0; 2344 } 2345 break; 2346 2347 default: 2348 bad = 1; 2349 break; 2350 } 2351 2352 cmn_err(level, "!%s%d: %s %s device; service %s%s"+(bad|changed), 2353 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip), 2354 bad ? "invalid report of fault" : action, 2355 location, still ? "still " : "", servstate); 2356 2357 cmn_err(level, "!%s%d: %s"+(bad|changed), 2358 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip), 2359 fedp->f_message); 2360 } 2361 2362 /* 2363 * Platform-settable pointers to fault handler and logger functions. 2364 * These are called by the default rootnex event-posting code when 2365 * a fault event reaches rootnex. 2366 */ 2367 void (*plat_fault_handler)(dev_info_t *, struct ddi_fault_event_data *) = 2368 i_ddi_fault_handler; 2369 void (*plat_fault_logger)(dev_info_t *, struct ddi_fault_event_data *) = 2370 i_ddi_fault_logger; 2371 2372 /* 2373 * Rootnex event definitions ... 2374 */ 2375 enum rootnex_event_tags { 2376 ROOTNEX_FAULT_EVENT 2377 }; 2378 static ndi_event_hdl_t rootnex_event_hdl; 2379 static ndi_event_definition_t rootnex_event_set[] = { 2380 { 2381 ROOTNEX_FAULT_EVENT, 2382 DDI_DEVI_FAULT_EVENT, 2383 EPL_INTERRUPT, 2384 NDI_EVENT_POST_TO_ALL 2385 } 2386 }; 2387 static ndi_event_set_t rootnex_events = { 2388 NDI_EVENTS_REV1, 2389 sizeof (rootnex_event_set) / sizeof (rootnex_event_set[0]), 2390 rootnex_event_set 2391 }; 2392 2393 /* 2394 * Initialize rootnex event handle 2395 */ 2396 void 2397 i_ddi_rootnex_init_events(dev_info_t *dip) 2398 { 2399 if (ndi_event_alloc_hdl(dip, (ddi_iblock_cookie_t)(LOCK_LEVEL-1), 2400 &rootnex_event_hdl, NDI_SLEEP) == NDI_SUCCESS) { 2401 if (ndi_event_bind_set(rootnex_event_hdl, 2402 &rootnex_events, NDI_SLEEP) != NDI_SUCCESS) { 2403 (void) ndi_event_free_hdl(rootnex_event_hdl); 2404 rootnex_event_hdl = NULL; 2405 } 2406 } 2407 } 2408 2409 /* 2410 * Event-handling functions for rootnex 2411 * These provide the standard implementation of fault handling 2412 */ 2413 /*ARGSUSED*/ 2414 int 2415 i_ddi_rootnex_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, 2416 char *eventname, ddi_eventcookie_t *cookiep) 2417 { 2418 if (rootnex_event_hdl == NULL) 2419 return (NDI_FAILURE); 2420 return (ndi_event_retrieve_cookie(rootnex_event_hdl, rdip, eventname, 2421 cookiep, NDI_EVENT_NOPASS)); 2422 } 2423 2424 /*ARGSUSED*/ 2425 int 2426 i_ddi_rootnex_add_eventcall(dev_info_t *dip, dev_info_t *rdip, 2427 ddi_eventcookie_t eventid, void (*handler)(dev_info_t *dip, 2428 ddi_eventcookie_t event, void *arg, void *impl_data), void *arg, 2429 ddi_callback_id_t *cb_id) 2430 { 2431 if (rootnex_event_hdl == NULL) 2432 return (NDI_FAILURE); 2433 return (ndi_event_add_callback(rootnex_event_hdl, rdip, 2434 eventid, handler, arg, NDI_SLEEP, cb_id)); 2435 } 2436 2437 /*ARGSUSED*/ 2438 int 2439 i_ddi_rootnex_remove_eventcall(dev_info_t *dip, ddi_callback_id_t cb_id) 2440 { 2441 if (rootnex_event_hdl == NULL) 2442 return (NDI_FAILURE); 2443 2444 return (ndi_event_remove_callback(rootnex_event_hdl, cb_id)); 2445 } 2446 2447 /*ARGSUSED*/ 2448 int 2449 i_ddi_rootnex_post_event(dev_info_t *dip, dev_info_t *rdip, 2450 ddi_eventcookie_t eventid, void *impl_data) 2451 { 2452 int tag; 2453 2454 if (rootnex_event_hdl == NULL) 2455 return (NDI_FAILURE); 2456 2457 tag = ndi_event_cookie_to_tag(rootnex_event_hdl, eventid); 2458 if (tag == ROOTNEX_FAULT_EVENT) { 2459 (*plat_fault_handler)(rdip, impl_data); 2460 (*plat_fault_logger)(rdip, impl_data); 2461 } 2462 return (ndi_event_run_callbacks(rootnex_event_hdl, rdip, 2463 eventid, impl_data)); 2464 } 2465 2466 /* 2467 * ndi_set_bus_private/ndi_get_bus_private: 2468 * Get/set device bus private data in devinfo. 2469 */ 2470 void 2471 ndi_set_bus_private(dev_info_t *dip, boolean_t up, uint32_t port_type, 2472 void *data) 2473 { 2474 if (up) { 2475 DEVI(dip)->devi_bus.port_up.info.port.type = port_type; 2476 DEVI(dip)->devi_bus.port_up.priv_p = data; 2477 } else { 2478 DEVI(dip)->devi_bus.port_down.info.port.type = port_type; 2479 DEVI(dip)->devi_bus.port_down.priv_p = data; 2480 } 2481 } 2482 2483 void * 2484 ndi_get_bus_private(dev_info_t *dip, boolean_t up) 2485 { 2486 if (up) 2487 return (DEVI(dip)->devi_bus.port_up.priv_p); 2488 else 2489 return (DEVI(dip)->devi_bus.port_down.priv_p); 2490 } 2491 2492 boolean_t 2493 ndi_port_type(dev_info_t *dip, boolean_t up, uint32_t port_type) 2494 { 2495 if (up) { 2496 return ((DEVI(dip)->devi_bus.port_up.info.port.type) == 2497 port_type); 2498 } else { 2499 return ((DEVI(dip)->devi_bus.port_down.info.port.type) == 2500 port_type); 2501 } 2502 } 2503