1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2009 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #include <sys/types.h> 27 #include <sys/param.h> 28 #include <sys/systm.h> 29 #include <sys/buf.h> 30 #include <sys/uio.h> 31 #include <sys/cred.h> 32 #include <sys/poll.h> 33 #include <sys/mman.h> 34 #include <sys/kmem.h> 35 #include <sys/model.h> 36 #include <sys/file.h> 37 #include <sys/proc.h> 38 #include <sys/open.h> 39 #include <sys/user.h> 40 #include <sys/t_lock.h> 41 #include <sys/vm.h> 42 #include <sys/stat.h> 43 #include <vm/hat.h> 44 #include <vm/seg.h> 45 #include <vm/as.h> 46 #include <sys/cmn_err.h> 47 #include <sys/debug.h> 48 #include <sys/avintr.h> 49 #include <sys/autoconf.h> 50 #include <sys/sunddi.h> 51 #include <sys/esunddi.h> 52 #include <sys/sunndi.h> 53 #include <sys/ddi.h> 54 #include <sys/kstat.h> 55 #include <sys/conf.h> 56 #include <sys/ddi_impldefs.h> /* include implementation structure defs */ 57 #include <sys/ndi_impldefs.h> 58 #include <sys/hwconf.h> 59 #include <sys/pathname.h> 60 #include <sys/modctl.h> 61 #include <sys/epm.h> 62 #include <sys/devctl.h> 63 #include <sys/callb.h> 64 #include <sys/bootconf.h> 65 #include <sys/dacf_impl.h> 66 #include <sys/nvpair.h> 67 #include <sys/sunmdi.h> 68 #include <sys/fs/dv_node.h> 69 #include <sys/sunldi_impl.h> 70 71 #ifdef __sparc 72 #include <sys/archsystm.h> /* getpil/setpil */ 73 #include <sys/membar.h> /* membar_sync */ 74 #endif 75 76 /* 77 * ndi property handling 78 */ 79 int 80 ndi_prop_update_int(dev_t match_dev, dev_info_t *dip, 81 char *name, int data) 82 { 83 return (ddi_prop_update_common(match_dev, dip, 84 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP, 85 name, &data, 1, ddi_prop_fm_encode_ints)); 86 } 87 88 int 89 ndi_prop_update_int64(dev_t match_dev, dev_info_t *dip, 90 char *name, int64_t data) 91 { 92 return (ddi_prop_update_common(match_dev, dip, 93 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP, 94 name, &data, 1, ddi_prop_fm_encode_int64)); 95 } 96 97 int 98 ndi_prop_create_boolean(dev_t match_dev, dev_info_t *dip, 99 char *name) 100 { 101 return (ddi_prop_update_common(match_dev, dip, 102 DDI_PROP_HW_DEF | DDI_PROP_TYPE_ANY | DDI_PROP_DONTSLEEP, 103 name, NULL, 0, ddi_prop_fm_encode_bytes)); 104 } 105 106 int 107 ndi_prop_update_int_array(dev_t match_dev, dev_info_t *dip, 108 char *name, int *data, uint_t nelements) 109 { 110 return (ddi_prop_update_common(match_dev, dip, 111 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT | DDI_PROP_DONTSLEEP, 112 name, data, nelements, ddi_prop_fm_encode_ints)); 113 } 114 115 int 116 ndi_prop_update_int64_array(dev_t match_dev, dev_info_t *dip, 117 char *name, int64_t *data, uint_t nelements) 118 { 119 return (ddi_prop_update_common(match_dev, dip, 120 DDI_PROP_HW_DEF | DDI_PROP_TYPE_INT64 | DDI_PROP_DONTSLEEP, 121 name, data, nelements, ddi_prop_fm_encode_int64)); 122 } 123 124 int 125 ndi_prop_update_string(dev_t match_dev, dev_info_t *dip, 126 char *name, char *data) 127 { 128 return (ddi_prop_update_common(match_dev, dip, 129 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP, 130 name, &data, 1, ddi_prop_fm_encode_string)); 131 } 132 133 int 134 ndi_prop_update_string_array(dev_t match_dev, dev_info_t *dip, 135 char *name, char **data, uint_t nelements) 136 { 137 return (ddi_prop_update_common(match_dev, dip, 138 DDI_PROP_HW_DEF | DDI_PROP_TYPE_STRING | DDI_PROP_DONTSLEEP, 139 name, data, nelements, 140 ddi_prop_fm_encode_strings)); 141 } 142 143 int 144 ndi_prop_update_byte_array(dev_t match_dev, dev_info_t *dip, 145 char *name, uchar_t *data, uint_t nelements) 146 { 147 if (nelements == 0) 148 return (DDI_PROP_INVAL_ARG); 149 150 return (ddi_prop_update_common(match_dev, dip, 151 DDI_PROP_HW_DEF | DDI_PROP_TYPE_BYTE | DDI_PROP_DONTSLEEP, 152 name, data, nelements, ddi_prop_fm_encode_bytes)); 153 } 154 155 int 156 ndi_prop_remove(dev_t dev, dev_info_t *dip, char *name) 157 { 158 return (ddi_prop_remove_common(dev, dip, name, DDI_PROP_HW_DEF)); 159 } 160 161 void 162 ndi_prop_remove_all(dev_info_t *dip) 163 { 164 i_ddi_prop_dyn_parent_set(dip, NULL); 165 ddi_prop_remove_all_common(dip, (int)DDI_PROP_HW_DEF); 166 } 167 168 /* 169 * Post an event notification to nexus driver responsible for handling 170 * the event. The responsible nexus is defined in the cookie passed in as 171 * the third parameter. 172 * The dip parameter is an artifact of an older implementation in which all 173 * requests to remove an eventcall would bubble up the tree. Today, this 174 * parameter is ignored. 175 * Input Parameters: 176 * dip - Ignored. 177 * rdip - device driver posting the event 178 * cookie - valid ddi_eventcookie_t, obtained by caller prior to 179 * invocation of this routine 180 * impl_data - used by framework 181 */ 182 /*ARGSUSED*/ 183 int 184 ndi_post_event(dev_info_t *dip, dev_info_t *rdip, 185 ddi_eventcookie_t cookie, void *impl_data) 186 { 187 dev_info_t *ddip; 188 189 ASSERT(cookie); 190 ddip = NDI_EVENT_DDIP(cookie); 191 192 /* 193 * perform sanity checks. These conditions should never be true. 194 */ 195 196 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 197 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 198 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event != NULL); 199 200 /* 201 * post the event to the responsible ancestor 202 */ 203 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_post_event)) 204 (ddip, rdip, cookie, impl_data)); 205 } 206 207 /* 208 * Calls the bus nexus driver's implementation of the 209 * (*bus_remove_eventcall)() interface. 210 */ 211 int 212 ndi_busop_remove_eventcall(dev_info_t *ddip, ddi_callback_id_t id) 213 { 214 215 ASSERT(id); 216 /* check for a correct revno before calling up the device tree. */ 217 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 218 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 219 220 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall == NULL) 221 return (DDI_FAILURE); 222 223 /* 224 * request responsible nexus to remove the eventcall 225 */ 226 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_remove_eventcall)) 227 (ddip, id)); 228 } 229 230 /* 231 * Calls the bus nexus driver's implementation of the 232 * (*bus_add_eventcall)() interface. The dip parameter is an 233 * artifact of an older implementation in which all requests to 234 * add an eventcall would bubble up the tree. Today, this parameter is 235 * ignored. 236 */ 237 /*ARGSUSED*/ 238 int 239 ndi_busop_add_eventcall(dev_info_t *dip, dev_info_t *rdip, 240 ddi_eventcookie_t cookie, void (*callback)(), void *arg, 241 ddi_callback_id_t *cb_id) 242 { 243 dev_info_t *ddip = (dev_info_t *)NDI_EVENT_DDIP(cookie); 244 245 /* 246 * check for a correct revno before calling up the device tree. 247 */ 248 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops != NULL); 249 ASSERT(DEVI(ddip)->devi_ops->devo_bus_ops->busops_rev >= BUSO_REV_6); 250 251 if (DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall == NULL) 252 return (DDI_FAILURE); 253 254 /* 255 * request responsible ancestor to add the eventcall 256 */ 257 return ((*(DEVI(ddip)->devi_ops->devo_bus_ops->bus_add_eventcall)) 258 (ddip, rdip, cookie, callback, arg, cb_id)); 259 } 260 261 /* 262 * Calls the bus nexus driver's implementation of the 263 * (*bus_get_eventcookie)() interface up the device tree hierarchy. 264 */ 265 int 266 ndi_busop_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, char *name, 267 ddi_eventcookie_t *event_cookiep) 268 { 269 dev_info_t *pdip = (dev_info_t *)DEVI(dip)->devi_parent; 270 271 /* Can not be called from rootnex. */ 272 ASSERT(pdip); 273 274 /* 275 * check for a correct revno before calling up the device tree. 276 */ 277 ASSERT(DEVI(pdip)->devi_ops->devo_bus_ops != NULL); 278 279 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev < BUSO_REV_6) || 280 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie == NULL)) { 281 #ifdef DEBUG 282 if ((DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev >= 283 BUSO_REV_3) && 284 (DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie)) { 285 cmn_err(CE_WARN, 286 "Warning: %s%d busops_rev=%d no longer supported" 287 " by the NDI event framework.\nBUSO_REV_6 or " 288 "greater must be used.", 289 DEVI(pdip)->devi_binding_name, 290 DEVI(pdip)->devi_instance, 291 DEVI(pdip)->devi_ops->devo_bus_ops->busops_rev); 292 } 293 #endif /* DEBUG */ 294 295 return (ndi_busop_get_eventcookie(pdip, rdip, name, 296 event_cookiep)); 297 } 298 299 return ((*(DEVI(pdip)->devi_ops->devo_bus_ops->bus_get_eventcookie)) 300 (pdip, rdip, name, event_cookiep)); 301 } 302 303 /* 304 * Copy in the devctl IOCTL data and return a handle to 305 * the data. 306 */ 307 int 308 ndi_dc_allochdl(void *iocarg, struct devctl_iocdata **rdcp) 309 { 310 struct devctl_iocdata *dcp; 311 char *cpybuf; 312 313 ASSERT(rdcp != NULL); 314 315 dcp = kmem_zalloc(sizeof (*dcp), KM_SLEEP); 316 317 if (get_udatamodel() == DATAMODEL_NATIVE) { 318 if (copyin(iocarg, dcp, sizeof (*dcp)) != 0) { 319 kmem_free(dcp, sizeof (*dcp)); 320 return (NDI_FAULT); 321 } 322 } 323 #ifdef _SYSCALL32_IMPL 324 else { 325 struct devctl_iocdata32 dcp32; 326 327 if (copyin(iocarg, &dcp32, sizeof (dcp32)) != 0) { 328 kmem_free(dcp, sizeof (*dcp)); 329 return (NDI_FAULT); 330 } 331 dcp->cmd = (uint_t)dcp32.cmd; 332 dcp->flags = (uint_t)dcp32.flags; 333 dcp->cpyout_buf = (uint_t *)(uintptr_t)dcp32.cpyout_buf; 334 dcp->nvl_user = (nvlist_t *)(uintptr_t)dcp32.nvl_user; 335 dcp->nvl_usersz = (size_t)dcp32.nvl_usersz; 336 dcp->c_nodename = (char *)(uintptr_t)dcp32.c_nodename; 337 dcp->c_unitaddr = (char *)(uintptr_t)dcp32.c_unitaddr; 338 } 339 #endif 340 if (dcp->c_nodename != NULL) { 341 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP); 342 if (copyinstr(dcp->c_nodename, cpybuf, MAXNAMELEN, 0) != 0) { 343 kmem_free(cpybuf, MAXNAMELEN); 344 kmem_free(dcp, sizeof (*dcp)); 345 return (NDI_FAULT); 346 } 347 cpybuf[MAXNAMELEN - 1] = '\0'; 348 dcp->c_nodename = cpybuf; 349 } 350 351 if (dcp->c_unitaddr != NULL) { 352 cpybuf = kmem_alloc(MAXNAMELEN, KM_SLEEP); 353 if (copyinstr(dcp->c_unitaddr, cpybuf, MAXNAMELEN, 0) != 0) { 354 kmem_free(cpybuf, MAXNAMELEN); 355 if (dcp->c_nodename != NULL) 356 kmem_free(dcp->c_nodename, MAXNAMELEN); 357 kmem_free(dcp, sizeof (*dcp)); 358 return (NDI_FAULT); 359 } 360 cpybuf[MAXNAMELEN - 1] = '\0'; 361 dcp->c_unitaddr = cpybuf; 362 } 363 364 /* 365 * copyin and unpack a user defined nvlist if one was passed 366 */ 367 if (dcp->nvl_user != NULL) { 368 if ((dcp->nvl_usersz == 0) || 369 (dcp->nvl_usersz > DEVCTL_MAX_NVL_USERSZ)) { 370 if (dcp->c_nodename != NULL) 371 kmem_free(dcp->c_nodename, MAXNAMELEN); 372 if (dcp->c_unitaddr != NULL) 373 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 374 kmem_free(dcp, sizeof (*dcp)); 375 return (NDI_FAILURE); 376 } 377 cpybuf = kmem_alloc(dcp->nvl_usersz, KM_SLEEP); 378 if (copyin(dcp->nvl_user, cpybuf, dcp->nvl_usersz) != 0) { 379 kmem_free(cpybuf, dcp->nvl_usersz); 380 if (dcp->c_nodename != NULL) 381 kmem_free(dcp->c_nodename, MAXNAMELEN); 382 if (dcp->c_unitaddr != NULL) 383 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 384 kmem_free(dcp, sizeof (*dcp)); 385 return (NDI_FAULT); 386 } 387 388 if (nvlist_unpack(cpybuf, dcp->nvl_usersz, &dcp->nvl_user, 389 KM_SLEEP)) { 390 kmem_free(cpybuf, dcp->nvl_usersz); 391 if (dcp->c_nodename != NULL) 392 kmem_free(dcp->c_nodename, MAXNAMELEN); 393 if (dcp->c_unitaddr != NULL) 394 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 395 kmem_free(dcp, sizeof (*dcp)); 396 return (NDI_FAULT); 397 } 398 /* 399 * free the buffer containing the packed nvlist 400 */ 401 kmem_free(cpybuf, dcp->nvl_usersz); 402 403 } 404 405 *rdcp = dcp; 406 return (NDI_SUCCESS); 407 } 408 409 /* 410 * free all space allocated to a handle. 411 */ 412 void 413 ndi_dc_freehdl(struct devctl_iocdata *dcp) 414 { 415 ASSERT(dcp != NULL); 416 417 if (dcp->c_nodename != NULL) 418 kmem_free(dcp->c_nodename, MAXNAMELEN); 419 420 if (dcp->c_unitaddr != NULL) 421 kmem_free(dcp->c_unitaddr, MAXNAMELEN); 422 423 if (dcp->nvl_user != NULL) 424 nvlist_free(dcp->nvl_user); 425 426 kmem_free(dcp, sizeof (*dcp)); 427 } 428 429 char * 430 ndi_dc_getname(struct devctl_iocdata *dcp) 431 { 432 ASSERT(dcp != NULL); 433 return (dcp->c_nodename); 434 435 } 436 437 char * 438 ndi_dc_getaddr(struct devctl_iocdata *dcp) 439 { 440 ASSERT(dcp != NULL); 441 return (dcp->c_unitaddr); 442 } 443 444 nvlist_t * 445 ndi_dc_get_ap_data(struct devctl_iocdata *dcp) 446 { 447 ASSERT(dcp != NULL); 448 449 return (dcp->nvl_user); 450 } 451 452 /* 453 * Transition the child named by "devname@devaddr" to the online state. 454 * For use by a driver's DEVCTL_DEVICE_ONLINE handler. 455 */ 456 int 457 ndi_devctl_device_online(dev_info_t *dip, struct devctl_iocdata *dcp, 458 uint_t flags) 459 { 460 int rval; 461 char *name; 462 dev_info_t *rdip; 463 464 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 465 return (EINVAL); 466 467 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 468 (void) snprintf(name, MAXNAMELEN, "%s@%s", 469 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 470 471 if ((rval = ndi_devi_config_one(dip, name, &rdip, 472 flags | NDI_DEVI_ONLINE | NDI_CONFIG)) == NDI_SUCCESS) { 473 ndi_rele_devi(rdip); 474 475 /* 476 * Invalidate devfs cached directory contents. For the checks 477 * in the "if" condition see the comment in ndi_devi_online(). 478 */ 479 if (i_ddi_devi_attached(dip) && !DEVI_BUSY_OWNED(dip)) 480 (void) devfs_clean(dip, NULL, 0); 481 482 } else if (rval == NDI_BUSY) { 483 rval = EBUSY; 484 } else if (rval == NDI_FAILURE) { 485 rval = EIO; 486 } 487 488 NDI_DEBUG(flags, (CE_CONT, "%s%d: online: %s: %s\n", 489 ddi_driver_name(dip), ddi_get_instance(dip), name, 490 ((rval == NDI_SUCCESS) ? "ok" : "failed"))); 491 492 kmem_free(name, MAXNAMELEN); 493 494 return (rval); 495 } 496 497 /* 498 * Transition the child named by "devname@devaddr" to the offline state. 499 * For use by a driver's DEVCTL_DEVICE_OFFLINE handler. 500 */ 501 int 502 ndi_devctl_device_offline(dev_info_t *dip, struct devctl_iocdata *dcp, 503 uint_t flags) 504 { 505 int rval; 506 char *name; 507 508 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 509 return (EINVAL); 510 511 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 512 (void) snprintf(name, MAXNAMELEN, "%s@%s", 513 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 514 515 (void) devfs_clean(dip, name, DV_CLEAN_FORCE); 516 rval = ndi_devi_unconfig_one(dip, name, NULL, 517 flags | NDI_DEVI_OFFLINE); 518 519 if (rval == NDI_BUSY) { 520 rval = EBUSY; 521 } else if (rval == NDI_FAILURE) { 522 rval = EIO; 523 } 524 525 NDI_DEBUG(flags, (CE_CONT, "%s%d: offline: %s: %s\n", 526 ddi_driver_name(dip), ddi_get_instance(dip), name, 527 (rval == NDI_SUCCESS) ? "ok" : "failed")); 528 529 kmem_free(name, MAXNAMELEN); 530 531 return (rval); 532 } 533 534 /* 535 * Remove the child named by "devname@devaddr". 536 * For use by a driver's DEVCTL_DEVICE_REMOVE handler. 537 */ 538 int 539 ndi_devctl_device_remove(dev_info_t *dip, struct devctl_iocdata *dcp, 540 uint_t flags) 541 { 542 int rval; 543 char *name; 544 545 if (ndi_dc_getname(dcp) == NULL || ndi_dc_getaddr(dcp) == NULL) 546 return (EINVAL); 547 548 name = kmem_alloc(MAXNAMELEN, KM_SLEEP); 549 (void) snprintf(name, MAXNAMELEN, "%s@%s", 550 ndi_dc_getname(dcp), ndi_dc_getaddr(dcp)); 551 552 (void) devfs_clean(dip, name, DV_CLEAN_FORCE); 553 554 rval = ndi_devi_unconfig_one(dip, name, NULL, flags | NDI_DEVI_REMOVE); 555 556 if (rval == NDI_BUSY) { 557 rval = EBUSY; 558 } else if (rval == NDI_FAILURE) { 559 rval = EIO; 560 } 561 562 NDI_DEBUG(flags, (CE_CONT, "%s%d: remove: %s: %s\n", 563 ddi_driver_name(dip), ddi_get_instance(dip), name, 564 (rval == NDI_SUCCESS) ? "ok" : "failed")); 565 566 kmem_free(name, MAXNAMELEN); 567 568 return (rval); 569 } 570 571 /* 572 * Return devctl state of the child named by "name@addr". 573 * For use by a driver's DEVCTL_DEVICE_GETSTATE handler. 574 */ 575 int 576 ndi_devctl_device_getstate(dev_info_t *parent, struct devctl_iocdata *dcp, 577 uint_t *state) 578 { 579 dev_info_t *dip; 580 char *name, *addr; 581 char *devname; 582 int devnamelen; 583 int circ; 584 585 if (parent == NULL || 586 ((name = ndi_dc_getname(dcp)) == NULL) || 587 ((addr = ndi_dc_getaddr(dcp)) == NULL)) 588 return (NDI_FAILURE); 589 590 devnamelen = strlen(name) + strlen(addr) + 2; 591 devname = kmem_alloc(devnamelen, KM_SLEEP); 592 if (strlen(addr) > 0) { 593 (void) snprintf(devname, devnamelen, "%s@%s", name, addr); 594 } else { 595 (void) snprintf(devname, devnamelen, "%s", name); 596 } 597 598 ndi_devi_enter(parent, &circ); 599 600 dip = ndi_devi_findchild(parent, devname); 601 kmem_free(devname, devnamelen); 602 603 if (dip == NULL) { 604 ndi_devi_exit(parent, circ); 605 return (NDI_FAILURE); 606 } 607 608 mutex_enter(&(DEVI(dip)->devi_lock)); 609 if (DEVI_IS_DEVICE_OFFLINE(dip)) { 610 *state = DEVICE_OFFLINE; 611 } else if (DEVI_IS_DEVICE_DOWN(dip)) { 612 *state = DEVICE_DOWN; 613 } else { 614 *state = DEVICE_ONLINE; 615 if (devi_stillreferenced(dip) == DEVI_REFERENCED) 616 *state |= DEVICE_BUSY; 617 } 618 619 mutex_exit(&(DEVI(dip)->devi_lock)); 620 ndi_devi_exit(parent, circ); 621 622 return (NDI_SUCCESS); 623 } 624 625 /* 626 * return the current state of the device "dip" 627 * 628 * recommend using ndi_devctl_ioctl() or 629 * ndi_devctl_device_getstate() instead 630 */ 631 int 632 ndi_dc_return_dev_state(dev_info_t *dip, struct devctl_iocdata *dcp) 633 { 634 dev_info_t *pdip; 635 uint_t devstate = 0; 636 int circ; 637 638 if ((dip == NULL) || (dcp == NULL)) 639 return (NDI_FAILURE); 640 641 pdip = ddi_get_parent(dip); 642 643 ndi_devi_enter(pdip, &circ); 644 mutex_enter(&(DEVI(dip)->devi_lock)); 645 if (DEVI_IS_DEVICE_OFFLINE(dip)) { 646 devstate = DEVICE_OFFLINE; 647 } else if (DEVI_IS_DEVICE_DOWN(dip)) { 648 devstate = DEVICE_DOWN; 649 } else { 650 devstate = DEVICE_ONLINE; 651 if (devi_stillreferenced(dip) == DEVI_REFERENCED) 652 devstate |= DEVICE_BUSY; 653 } 654 655 mutex_exit(&(DEVI(dip)->devi_lock)); 656 ndi_devi_exit(pdip, circ); 657 658 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0) 659 return (NDI_FAULT); 660 661 return (NDI_SUCCESS); 662 } 663 664 /* 665 * Return device's bus state 666 * For use by a driver's DEVCTL_BUS_GETSTATE handler. 667 */ 668 int 669 ndi_devctl_bus_getstate(dev_info_t *dip, struct devctl_iocdata *dcp, 670 uint_t *state) 671 { 672 if ((dip == NULL) || (dcp == NULL)) 673 return (NDI_FAILURE); 674 675 return (ndi_get_bus_state(dip, state)); 676 } 677 678 /* 679 * Generic devctl ioctl handler 680 */ 681 int 682 ndi_devctl_ioctl(dev_info_t *dip, int cmd, intptr_t arg, int mode, uint_t flags) 683 { 684 _NOTE(ARGUNUSED(mode)) 685 struct devctl_iocdata *dcp; 686 uint_t state; 687 int rval = ENOTTY; 688 689 /* 690 * read devctl ioctl data 691 */ 692 if (ndi_dc_allochdl((void *)arg, &dcp) != NDI_SUCCESS) 693 return (EFAULT); 694 695 switch (cmd) { 696 697 case DEVCTL_BUS_GETSTATE: 698 rval = ndi_devctl_bus_getstate(dip, dcp, &state); 699 if (rval == NDI_SUCCESS) { 700 if (copyout(&state, dcp->cpyout_buf, 701 sizeof (uint_t)) != 0) 702 rval = NDI_FAULT; 703 } 704 break; 705 706 case DEVCTL_DEVICE_ONLINE: 707 rval = ndi_devctl_device_online(dip, dcp, flags); 708 break; 709 710 case DEVCTL_DEVICE_OFFLINE: 711 rval = ndi_devctl_device_offline(dip, dcp, flags); 712 break; 713 714 case DEVCTL_DEVICE_GETSTATE: 715 rval = ndi_devctl_device_getstate(dip, dcp, &state); 716 if (rval == NDI_SUCCESS) { 717 if (copyout(&state, dcp->cpyout_buf, 718 sizeof (uint_t)) != 0) 719 rval = NDI_FAULT; 720 } 721 break; 722 723 case DEVCTL_DEVICE_REMOVE: 724 rval = ndi_devctl_device_remove(dip, dcp, flags); 725 break; 726 727 case DEVCTL_BUS_DEV_CREATE: 728 rval = ndi_dc_devi_create(dcp, dip, 0, NULL); 729 break; 730 731 /* 732 * ioctls for which a generic implementation makes no sense 733 */ 734 case DEVCTL_BUS_RESET: 735 case DEVCTL_BUS_RESETALL: 736 case DEVCTL_DEVICE_RESET: 737 case DEVCTL_AP_CONNECT: 738 case DEVCTL_AP_DISCONNECT: 739 case DEVCTL_AP_INSERT: 740 case DEVCTL_AP_REMOVE: 741 case DEVCTL_AP_CONFIGURE: 742 case DEVCTL_AP_UNCONFIGURE: 743 case DEVCTL_AP_GETSTATE: 744 case DEVCTL_AP_CONTROL: 745 case DEVCTL_BUS_QUIESCE: 746 case DEVCTL_BUS_UNQUIESCE: 747 rval = ENOTSUP; 748 break; 749 } 750 751 ndi_dc_freehdl(dcp); 752 return (rval); 753 } 754 755 /* 756 * Copyout the state of the Attachment Point "ap" to the requesting 757 * user process. 758 */ 759 int 760 ndi_dc_return_ap_state(devctl_ap_state_t *ap, struct devctl_iocdata *dcp) 761 { 762 if ((ap == NULL) || (dcp == NULL)) 763 return (NDI_FAILURE); 764 765 766 if (get_udatamodel() == DATAMODEL_NATIVE) { 767 if (copyout(ap, dcp->cpyout_buf, 768 sizeof (devctl_ap_state_t)) != 0) 769 return (NDI_FAULT); 770 } 771 #ifdef _SYSCALL32_IMPL 772 else { 773 struct devctl_ap_state32 ap_state32; 774 775 ap_state32.ap_rstate = ap->ap_rstate; 776 ap_state32.ap_ostate = ap->ap_ostate; 777 ap_state32.ap_condition = ap->ap_condition; 778 ap_state32.ap_error_code = ap->ap_error_code; 779 ap_state32.ap_in_transition = ap->ap_in_transition; 780 ap_state32.ap_last_change = (time32_t)ap->ap_last_change; 781 if (copyout(&ap_state32, dcp->cpyout_buf, 782 sizeof (devctl_ap_state32_t)) != 0) 783 return (NDI_FAULT); 784 } 785 #endif 786 787 return (NDI_SUCCESS); 788 } 789 790 /* 791 * Copyout the bus state of the bus nexus device "dip" to the requesting 792 * user process. 793 */ 794 int 795 ndi_dc_return_bus_state(dev_info_t *dip, struct devctl_iocdata *dcp) 796 { 797 uint_t devstate = 0; 798 799 if ((dip == NULL) || (dcp == NULL)) 800 return (NDI_FAILURE); 801 802 if (ndi_get_bus_state(dip, &devstate) != NDI_SUCCESS) 803 return (NDI_FAILURE); 804 805 if (copyout(&devstate, dcp->cpyout_buf, sizeof (uint_t)) != 0) 806 return (NDI_FAULT); 807 808 return (NDI_SUCCESS); 809 } 810 811 static int 812 i_dc_devi_create(struct devctl_iocdata *, dev_info_t *, dev_info_t **); 813 814 /* 815 * create a child device node given the property definitions 816 * supplied by the userland process 817 */ 818 int 819 ndi_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip, int flags, 820 dev_info_t **rdip) 821 { 822 dev_info_t *cdip; 823 int rv, circular = 0; 824 char devnm[MAXNAMELEN]; 825 int nmlen; 826 827 /* 828 * The child device may have been pre-constructed by an earlier 829 * call to this function with the flag DEVCTL_CONSTRUCT set. 830 */ 831 832 if ((cdip = (rdip != NULL) ? *rdip : NULL) == NULL) 833 if ((rv = i_dc_devi_create(dcp, pdip, &cdip)) != 0) 834 return (rv); 835 836 ASSERT(cdip != NULL); 837 838 /* 839 * Return the device node partially constructed if the 840 * DEVCTL_CONSTRUCT flag is set. 841 */ 842 if (flags & DEVCTL_CONSTRUCT) { 843 if (rdip == NULL) { 844 (void) ndi_devi_free(cdip); 845 return (EINVAL); 846 } 847 *rdip = cdip; 848 return (0); 849 } 850 851 /* 852 * Bring the node up to a named but OFFLINE state. The calling 853 * application will need to manage the node from here on. 854 */ 855 if (dcp->flags & DEVCTL_OFFLINE) { 856 /* 857 * In the unlikely event that the dip was somehow attached by 858 * the userland process (and device contracts or LDI opens 859 * were registered against the dip) after it was created by 860 * a previous DEVCTL_CONSTRUCT call, we start notify 861 * proceedings on this dip. Note that we don't need to 862 * return the dip after a failure of the notify since 863 * for a contract or LDI handle to be created the dip was 864 * already available to the user. 865 */ 866 if (e_ddi_offline_notify(cdip) == DDI_FAILURE) { 867 return (EBUSY); 868 } 869 870 /* 871 * hand set the OFFLINE flag to prevent any asynchronous 872 * autoconfiguration operations from attaching this node. 873 */ 874 mutex_enter(&(DEVI(cdip)->devi_lock)); 875 DEVI_SET_DEVICE_OFFLINE(cdip); 876 mutex_exit(&(DEVI(cdip)->devi_lock)); 877 878 e_ddi_offline_finalize(cdip, DDI_SUCCESS); 879 880 rv = ndi_devi_bind_driver(cdip, flags); 881 if (rv != NDI_SUCCESS) { 882 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 883 return (ENXIO); 884 } 885 886 /* 887 * remove the dev_info node if it failed to bind to a 888 * driver above. 889 */ 890 if (i_ddi_node_state(cdip) < DS_BOUND) { 891 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 892 return (ENXIO); 893 } 894 895 /* 896 * add the node to the per-driver list and INITCHILD it 897 * to give it a name. 898 */ 899 ndi_devi_enter(pdip, &circular); 900 if ((rv = ddi_initchild(pdip, cdip)) != DDI_SUCCESS) { 901 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 902 ndi_devi_exit(pdip, circular); 903 return (EINVAL); 904 } 905 ndi_devi_exit(pdip, circular); 906 907 } else { 908 /* 909 * Attempt to bring the device ONLINE. If the request to 910 * fails, remove the dev_info node. 911 */ 912 if (ndi_devi_online(cdip, NDI_ONLINE_ATTACH) != NDI_SUCCESS) { 913 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 914 return (ENXIO); 915 } 916 917 /* 918 * if the node was successfully added but there was 919 * no driver available for the device, remove the node 920 */ 921 if (i_ddi_node_state(cdip) < DS_BOUND) { 922 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 923 return (ENODEV); 924 } 925 } 926 927 /* 928 * return a handle to the child device 929 * copy out the name of the newly attached child device if 930 * the IOCTL request has provided a copyout buffer. 931 */ 932 if (rdip != NULL) 933 *rdip = cdip; 934 935 if (dcp->cpyout_buf == NULL) 936 return (0); 937 938 ASSERT(ddi_node_name(cdip) != NULL); 939 ASSERT(ddi_get_name_addr(cdip) != NULL); 940 941 nmlen = snprintf(devnm, MAXNAMELEN, "%s@%s", 942 ddi_node_name(cdip), ddi_get_name_addr(cdip)); 943 944 if (copyout(&devnm, dcp->cpyout_buf, nmlen) != 0) { 945 (void) ndi_devi_offline(cdip, NDI_DEVI_REMOVE); 946 return (EFAULT); 947 } 948 return (0); 949 } 950 951 static int 952 i_dc_devi_create(struct devctl_iocdata *dcp, dev_info_t *pdip, 953 dev_info_t **rdip) 954 { 955 956 dev_info_t *cdip; 957 char *cname = NULL; 958 nvlist_t *nvlp = dcp->nvl_user; 959 nvpair_t *npp; 960 char *np; 961 int rv = 0; 962 963 ASSERT(rdip != NULL && *rdip == NULL); 964 965 if ((nvlp == NULL) || 966 (nvlist_lookup_string(nvlp, DC_DEVI_NODENAME, &cname) != 0)) 967 return (EINVAL); 968 969 /* 970 * construct a new dev_info node with a user-provided nodename 971 */ 972 ndi_devi_alloc_sleep(pdip, cname, (pnode_t)DEVI_SID_NODEID, &cdip); 973 974 /* 975 * create hardware properties for each member in the property 976 * list. 977 */ 978 for (npp = nvlist_next_nvpair(nvlp, NULL); (npp != NULL && !rv); 979 npp = nvlist_next_nvpair(nvlp, npp)) { 980 981 np = nvpair_name(npp); 982 983 /* 984 * skip the nodename property 985 */ 986 if (strcmp(np, DC_DEVI_NODENAME) == 0) 987 continue; 988 989 switch (nvpair_type(npp)) { 990 991 case DATA_TYPE_INT32: { 992 int32_t prop_val; 993 994 if ((rv = nvpair_value_int32(npp, &prop_val)) != 0) 995 break; 996 997 (void) ndi_prop_update_int(DDI_DEV_T_NONE, cdip, np, 998 (int)prop_val); 999 break; 1000 } 1001 1002 case DATA_TYPE_STRING: { 1003 char *prop_val; 1004 1005 if ((rv = nvpair_value_string(npp, &prop_val)) != 0) 1006 break; 1007 1008 (void) ndi_prop_update_string(DDI_DEV_T_NONE, cdip, 1009 np, prop_val); 1010 break; 1011 } 1012 1013 case DATA_TYPE_BYTE_ARRAY: { 1014 uchar_t *val; 1015 uint_t nelms; 1016 1017 if ((rv = nvpair_value_byte_array(npp, &val, 1018 &nelms)) != 0) 1019 break; 1020 1021 (void) ndi_prop_update_byte_array(DDI_DEV_T_NONE, 1022 cdip, np, (uchar_t *)val, nelms); 1023 break; 1024 } 1025 1026 case DATA_TYPE_INT32_ARRAY: { 1027 int32_t *val; 1028 uint_t nelms; 1029 1030 if ((rv = nvpair_value_int32_array(npp, &val, 1031 &nelms)) != 0) 1032 break; 1033 1034 (void) ndi_prop_update_int_array(DDI_DEV_T_NONE, 1035 cdip, np, val, nelms); 1036 break; 1037 } 1038 1039 case DATA_TYPE_STRING_ARRAY: { 1040 char **val; 1041 uint_t nelms; 1042 1043 if ((rv = nvpair_value_string_array(npp, &val, 1044 &nelms)) != 0) 1045 break; 1046 1047 (void) ndi_prop_update_string_array(DDI_DEV_T_NONE, 1048 cdip, np, val, nelms); 1049 break; 1050 } 1051 1052 /* 1053 * unsupported property data type 1054 */ 1055 default: 1056 rv = EINVAL; 1057 } 1058 } 1059 1060 /* 1061 * something above failed 1062 * destroy the partially child device and abort the request 1063 */ 1064 if (rv != 0) { 1065 (void) ndi_devi_free(cdip); 1066 return (rv); 1067 } 1068 1069 *rdip = cdip; 1070 return (0); 1071 } 1072 1073 /* 1074 * return current soft bus state of bus nexus "dip" 1075 */ 1076 int 1077 ndi_get_bus_state(dev_info_t *dip, uint_t *rstate) 1078 { 1079 if (dip == NULL || rstate == NULL) 1080 return (NDI_FAILURE); 1081 1082 if (DEVI(dip)->devi_ops->devo_bus_ops == NULL) 1083 return (NDI_FAILURE); 1084 1085 mutex_enter(&(DEVI(dip)->devi_lock)); 1086 if (DEVI_IS_BUS_QUIESCED(dip)) 1087 *rstate = BUS_QUIESCED; 1088 else if (DEVI_IS_BUS_DOWN(dip)) 1089 *rstate = BUS_SHUTDOWN; 1090 else 1091 *rstate = BUS_ACTIVE; 1092 mutex_exit(&(DEVI(dip)->devi_lock)); 1093 return (NDI_SUCCESS); 1094 } 1095 1096 /* 1097 * Set the soft state of bus nexus "dip" 1098 */ 1099 int 1100 ndi_set_bus_state(dev_info_t *dip, uint_t state) 1101 { 1102 int rv = NDI_SUCCESS; 1103 1104 if (dip == NULL) 1105 return (NDI_FAILURE); 1106 1107 mutex_enter(&(DEVI(dip)->devi_lock)); 1108 1109 switch (state) { 1110 case BUS_QUIESCED: 1111 DEVI_SET_BUS_QUIESCE(dip); 1112 break; 1113 1114 case BUS_ACTIVE: 1115 DEVI_SET_BUS_ACTIVE(dip); 1116 DEVI_SET_BUS_UP(dip); 1117 break; 1118 1119 case BUS_SHUTDOWN: 1120 DEVI_SET_BUS_DOWN(dip); 1121 break; 1122 1123 default: 1124 rv = NDI_FAILURE; 1125 } 1126 1127 mutex_exit(&(DEVI(dip)->devi_lock)); 1128 return (rv); 1129 } 1130 1131 /* 1132 * These dummy functions are obsolete and may be removed. 1133 * Retained for existing driver compatibility only. 1134 * Drivers should be fixed not to use these functions. 1135 * Don't write new code using these obsolete interfaces. 1136 */ 1137 /*ARGSUSED*/ 1138 void 1139 i_ndi_block_device_tree_changes(uint_t *lkcnt) /* obsolete */ 1140 { 1141 /* obsolete dummy function */ 1142 } 1143 1144 /*ARGSUSED*/ 1145 void 1146 i_ndi_allow_device_tree_changes(uint_t lkcnt) /* obsolete */ 1147 { 1148 /* obsolete dummy function */ 1149 } 1150 1151 /* 1152 * Single thread entry into per-driver list 1153 */ 1154 /*ARGSUSED*/ 1155 void 1156 e_ddi_enter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */ 1157 { 1158 /* obsolete dummy function */ 1159 } 1160 1161 /* 1162 * release the per-driver list 1163 */ 1164 /*ARGSUSED*/ 1165 void 1166 e_ddi_exit_driver_list(struct devnames *dnp, int listcnt) /* obsolete */ 1167 { 1168 /* obsolete dummy function */ 1169 } 1170 1171 /* 1172 * Attempt to enter driver list 1173 */ 1174 /*ARGSUSED*/ 1175 int 1176 e_ddi_tryenter_driver_list(struct devnames *dnp, int *listcnt) /* obsolete */ 1177 { 1178 return (1); /* obsolete dummy function */ 1179 } 1180 1181 /* 1182 * ndi event handling support functions: 1183 * The NDI event support model is as follows: 1184 * 1185 * The nexus driver defines a set of events using some static structures (so 1186 * these structures can be shared by all instances of the nexus driver). 1187 * The nexus driver allocates an event handle and binds the event set 1188 * to this handle. The nexus driver's event busop functions can just 1189 * call the appropriate NDI event support function using this handle 1190 * as the first argument. 1191 * 1192 * The reasoning for tying events to the device tree is that the entity 1193 * generating the callback will typically be one of the device driver's 1194 * ancestors in the tree. 1195 */ 1196 static int ndi_event_debug = 0; 1197 1198 #ifdef DEBUG 1199 #define NDI_EVENT_DEBUG ndi_event_debug 1200 #endif /* DEBUG */ 1201 1202 /* 1203 * allocate a new ndi event handle 1204 */ 1205 int 1206 ndi_event_alloc_hdl(dev_info_t *dip, ddi_iblock_cookie_t cookie, 1207 ndi_event_hdl_t *handle, uint_t flag) 1208 { 1209 struct ndi_event_hdl *ndi_event_hdl; 1210 1211 ndi_event_hdl = kmem_zalloc(sizeof (struct ndi_event_hdl), 1212 ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP)); 1213 1214 if (!ndi_event_hdl) { 1215 return (NDI_FAILURE); 1216 } 1217 1218 ndi_event_hdl->ndi_evthdl_dip = dip; 1219 ndi_event_hdl->ndi_evthdl_iblock_cookie = cookie; 1220 mutex_init(&ndi_event_hdl->ndi_evthdl_mutex, NULL, 1221 MUTEX_DRIVER, (void *)cookie); 1222 1223 mutex_init(&ndi_event_hdl->ndi_evthdl_cb_mutex, NULL, 1224 MUTEX_DRIVER, (void *)cookie); 1225 1226 *handle = (ndi_event_hdl_t)ndi_event_hdl; 1227 1228 return (NDI_SUCCESS); 1229 } 1230 1231 /* 1232 * free the ndi event handle 1233 */ 1234 int 1235 ndi_event_free_hdl(ndi_event_hdl_t handle) 1236 { 1237 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1238 ndi_event_cookie_t *cookie; 1239 ndi_event_cookie_t *free; 1240 1241 ASSERT(handle); 1242 1243 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1244 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1245 1246 cookie = ndi_event_hdl->ndi_evthdl_cookie_list; 1247 1248 /* deallocate all defined cookies */ 1249 while (cookie != NULL) { 1250 ASSERT(cookie->callback_list == NULL); 1251 free = cookie; 1252 cookie = cookie->next_cookie; 1253 1254 kmem_free(free, sizeof (ndi_event_cookie_t)); 1255 } 1256 1257 1258 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1259 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1260 1261 /* destroy mutexes */ 1262 mutex_destroy(&ndi_event_hdl->ndi_evthdl_mutex); 1263 mutex_destroy(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1264 1265 /* free event handle */ 1266 kmem_free(ndi_event_hdl, sizeof (struct ndi_event_hdl)); 1267 1268 return (NDI_SUCCESS); 1269 } 1270 1271 1272 /* 1273 * ndi_event_bind_set() adds a set of events to the NDI event 1274 * handle. 1275 * 1276 * Events generated by high level interrupts should not 1277 * be mixed in the same event set with events generated by 1278 * normal interrupts or kernel events. 1279 * 1280 * This function can be called multiple times to bind 1281 * additional sets to the event handle. 1282 * However, events generated by high level interrupts cannot 1283 * be bound to a handle that already has bound events generated 1284 * by normal interrupts or from kernel context and vice versa. 1285 */ 1286 int 1287 ndi_event_bind_set(ndi_event_hdl_t handle, 1288 ndi_event_set_t *ndi_events, 1289 uint_t flag) 1290 { 1291 struct ndi_event_hdl *ndi_event_hdl; 1292 ndi_event_cookie_t *next, *prev, *new_cookie; 1293 uint_t i, len; 1294 uint_t dup = 0; 1295 uint_t high_plevels, other_plevels; 1296 ndi_event_definition_t *ndi_event_defs; 1297 1298 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP); 1299 1300 ASSERT(handle); 1301 ASSERT(ndi_events); 1302 1303 /* 1304 * binding must be performed during attach/detach 1305 */ 1306 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) && 1307 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) { 1308 cmn_err(CE_WARN, "ndi_event_bind_set must be called within " 1309 "attach or detach"); 1310 return (NDI_FAILURE); 1311 } 1312 1313 /* 1314 * if it is not the correct version or the event set is 1315 * empty, bail out 1316 */ 1317 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1) 1318 return (NDI_FAILURE); 1319 1320 ndi_event_hdl = (struct ndi_event_hdl *)handle; 1321 ndi_event_defs = ndi_events->ndi_event_defs; 1322 high_plevels = other_plevels = 0; 1323 1324 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1325 1326 /* check for mixing events at high level with the other types */ 1327 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1328 if (ndi_event_defs[i].ndi_event_plevel == EPL_HIGHLEVEL) { 1329 high_plevels++; 1330 } else { 1331 other_plevels++; 1332 } 1333 } 1334 1335 /* 1336 * bail out if high level events are mixed with other types in this 1337 * event set or the set is incompatible with the set in the handle 1338 */ 1339 if ((high_plevels && other_plevels) || 1340 (other_plevels && ndi_event_hdl->ndi_evthdl_high_plevels) || 1341 (high_plevels && ndi_event_hdl->ndi_evthdl_other_plevels)) { 1342 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1343 1344 return (NDI_FAILURE); 1345 } 1346 1347 /* 1348 * check for duplicate events in both the existing handle 1349 * and the event set, add events if not duplicates 1350 */ 1351 next = ndi_event_hdl->ndi_evthdl_cookie_list; 1352 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1353 while (next != NULL) { 1354 len = strlen(NDI_EVENT_NAME(next)) + 1; 1355 if (strncmp(NDI_EVENT_NAME(next), 1356 ndi_event_defs[i].ndi_event_name, len) == 0) { 1357 dup = 1; 1358 break; 1359 } 1360 1361 prev = next; 1362 next = next->next_cookie; 1363 } 1364 1365 if (dup == 0) { 1366 new_cookie = kmem_zalloc(sizeof (ndi_event_cookie_t), 1367 km_flag); 1368 1369 if (!new_cookie) 1370 return (NDI_FAILURE); 1371 1372 if (ndi_event_hdl->ndi_evthdl_n_events == 0) { 1373 ndi_event_hdl->ndi_evthdl_cookie_list = 1374 new_cookie; 1375 } else { 1376 prev->next_cookie = new_cookie; 1377 } 1378 1379 ndi_event_hdl->ndi_evthdl_n_events++; 1380 1381 /* 1382 * set up new cookie 1383 */ 1384 new_cookie->definition = &ndi_event_defs[i]; 1385 new_cookie->ddip = ndi_event_hdl->ndi_evthdl_dip; 1386 1387 } else { 1388 /* 1389 * event not added, must correct plevel numbers 1390 */ 1391 if (ndi_event_defs[i].ndi_event_plevel == 1392 EPL_HIGHLEVEL) { 1393 high_plevels--; 1394 } else { 1395 other_plevels--; 1396 } 1397 } 1398 1399 dup = 0; 1400 next = ndi_event_hdl->ndi_evthdl_cookie_list; 1401 prev = NULL; 1402 1403 } 1404 1405 ndi_event_hdl->ndi_evthdl_high_plevels += high_plevels; 1406 ndi_event_hdl->ndi_evthdl_other_plevels += other_plevels; 1407 1408 ASSERT((ndi_event_hdl->ndi_evthdl_high_plevels == 0) || 1409 (ndi_event_hdl->ndi_evthdl_other_plevels == 0)); 1410 1411 #ifdef NDI_EVENT_DEBUG 1412 if (ndi_event_debug) { 1413 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_bind_set"); 1414 } 1415 #endif /* NDI_EVENT_DEBUG */ 1416 1417 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1418 1419 return (NDI_SUCCESS); 1420 } 1421 1422 /* 1423 * ndi_event_unbind_set() unbinds a set of events, previously 1424 * bound using ndi_event_bind_set(), from the NDI event 1425 * handle. 1426 * 1427 * This routine will unbind all events in the event set. If an event, 1428 * specified in the event set, is not found in the handle, this 1429 * routine will proceed onto the next member of the set as if the event 1430 * was never specified. 1431 * 1432 * The event set may be a subset of the set of events that 1433 * was previously bound to the handle. For example, events 1434 * can be individually unbound. 1435 * 1436 * An event cannot be unbound if callbacks are still 1437 * registered against the event. 1438 */ 1439 /*ARGSUSED*/ 1440 int 1441 ndi_event_unbind_set(ndi_event_hdl_t handle, ndi_event_set_t *ndi_events, 1442 uint_t flag) 1443 { 1444 ndi_event_definition_t *ndi_event_defs; 1445 int len; 1446 uint_t i; 1447 int rval; 1448 ndi_event_cookie_t *cookie_list; 1449 ndi_event_cookie_t *prev = NULL; 1450 1451 ASSERT(ndi_events); 1452 ASSERT(handle); 1453 1454 /* 1455 * binding must be performed during attach/detac 1456 */ 1457 if (!DEVI_IS_ATTACHING(handle->ndi_evthdl_dip) && 1458 !DEVI_IS_DETACHING(handle->ndi_evthdl_dip)) { 1459 cmn_err(CE_WARN, "ndi_event_bind_set must be called within " 1460 "attach or detach"); 1461 return (NDI_FAILURE); 1462 } 1463 1464 /* bail out if ndi_event_set is outdated */ 1465 if (ndi_events->ndi_events_version != NDI_EVENTS_REV1) { 1466 return (NDI_FAILURE); 1467 } 1468 1469 ASSERT(ndi_events->ndi_event_defs); 1470 1471 ndi_event_defs = ndi_events->ndi_event_defs; 1472 1473 mutex_enter(&handle->ndi_evthdl_mutex); 1474 mutex_enter(&handle->ndi_evthdl_cb_mutex); 1475 1476 /* 1477 * Verify that all events in the event set are eligible 1478 * for unbinding(ie. there are no outstanding callbacks). 1479 * If any one of the events are ineligible, fail entire 1480 * operation. 1481 */ 1482 1483 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1484 cookie_list = handle->ndi_evthdl_cookie_list; 1485 while (cookie_list != NULL) { 1486 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1487 if (strncmp(NDI_EVENT_NAME(cookie_list), 1488 ndi_event_defs[i].ndi_event_name, len) == 0) { 1489 1490 ASSERT(cookie_list->callback_list == NULL); 1491 if (cookie_list->callback_list) { 1492 rval = NDI_FAILURE; 1493 goto done; 1494 } 1495 break; 1496 } else { 1497 cookie_list = cookie_list->next_cookie; 1498 } 1499 } 1500 } 1501 1502 /* 1503 * remove all events found within the handle 1504 * If an event is not found, this function will proceed as if the event 1505 * was never specified. 1506 */ 1507 1508 for (i = 0; i < ndi_events->ndi_n_events; i++) { 1509 cookie_list = handle->ndi_evthdl_cookie_list; 1510 prev = NULL; 1511 while (cookie_list != NULL) { 1512 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1513 if (strncmp(NDI_EVENT_NAME(cookie_list), 1514 ndi_event_defs[i].ndi_event_name, len) == 0) { 1515 1516 /* 1517 * can not unbind an event definition with 1518 * outstanding callbacks 1519 */ 1520 if (cookie_list->callback_list) { 1521 rval = NDI_FAILURE; 1522 goto done; 1523 } 1524 1525 /* remove this cookie from the list */ 1526 if (prev != NULL) { 1527 prev->next_cookie = 1528 cookie_list->next_cookie; 1529 } else { 1530 handle->ndi_evthdl_cookie_list = 1531 cookie_list->next_cookie; 1532 } 1533 1534 /* adjust plevel counts */ 1535 if (NDI_EVENT_PLEVEL(cookie_list) == 1536 EPL_HIGHLEVEL) { 1537 handle->ndi_evthdl_high_plevels--; 1538 } else { 1539 handle->ndi_evthdl_other_plevels--; 1540 } 1541 1542 /* adjust cookie count */ 1543 handle->ndi_evthdl_n_events--; 1544 1545 /* free the cookie */ 1546 kmem_free(cookie_list, 1547 sizeof (ndi_event_cookie_t)); 1548 1549 cookie_list = handle->ndi_evthdl_cookie_list; 1550 break; 1551 1552 } else { 1553 prev = cookie_list; 1554 cookie_list = cookie_list->next_cookie; 1555 } 1556 1557 } 1558 1559 } 1560 1561 #ifdef NDI_EVENT_DEBUG 1562 if (ndi_event_debug) { 1563 ndi_event_dump_hdl(handle, "ndi_event_unbind_set"); 1564 } 1565 #endif /* NDI_EVENT_DEBUG */ 1566 1567 rval = NDI_SUCCESS; 1568 1569 done: 1570 mutex_exit(&handle->ndi_evthdl_cb_mutex); 1571 mutex_exit(&handle->ndi_evthdl_mutex); 1572 1573 return (rval); 1574 } 1575 1576 /* 1577 * ndi_event_retrieve_cookie(): 1578 * Return an event cookie for eventname if this nexus driver 1579 * has defined the named event. The event cookie returned 1580 * by this function is used to register callback handlers 1581 * for the event. 1582 * 1583 * ndi_event_retrieve_cookie() is intended to be used in the 1584 * nexus driver's bus_get_eventcookie busop routine. 1585 * 1586 * If the event is not defined by this bus nexus driver, and flag 1587 * does not include NDI_EVENT_NOPASS, then ndi_event_retrieve_cookie() 1588 * will pass the request up the device tree hierarchy by calling 1589 * ndi_busop_get_eventcookie(9N). 1590 * If the event is not defined by this bus nexus driver, and flag 1591 * does include NDI_EVENT_NOPASS, ndi_event_retrieve_cookie() 1592 * will return NDI_FAILURE. The caller may then determine what further 1593 * action to take, such as using a different handle, passing the 1594 * request up the device tree using ndi_busop_get_eventcookie(9N), 1595 * or returning the failure to the caller, thus blocking the 1596 * progress of the request up the tree. 1597 */ 1598 int 1599 ndi_event_retrieve_cookie(ndi_event_hdl_t handle, 1600 dev_info_t *rdip, 1601 char *eventname, 1602 ddi_eventcookie_t *cookiep, 1603 uint_t flag) 1604 { 1605 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1606 int len; 1607 ndi_event_cookie_t *cookie_list; 1608 1609 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1610 1611 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list; 1612 /* 1613 * search the cookie list for the event name and return 1614 * cookie if found. 1615 */ 1616 while (cookie_list != NULL) { 1617 1618 len = strlen(NDI_EVENT_NAME(cookie_list)) + 1; 1619 if (strncmp(NDI_EVENT_NAME(cookie_list), eventname, 1620 len) == 0) { 1621 *cookiep = (ddi_eventcookie_t)cookie_list; 1622 1623 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1624 return (NDI_SUCCESS); 1625 } 1626 1627 cookie_list = cookie_list->next_cookie; 1628 } 1629 1630 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1631 /* 1632 * event was not found, pass up or return failure 1633 */ 1634 if ((flag & NDI_EVENT_NOPASS) == 0) { 1635 return (ndi_busop_get_eventcookie( 1636 ndi_event_hdl->ndi_evthdl_dip, rdip, eventname, cookiep)); 1637 } else { 1638 return (NDI_FAILURE); 1639 } 1640 } 1641 1642 /* 1643 * check whether this nexus defined this event and look up attributes 1644 */ 1645 static int 1646 ndi_event_is_defined(ndi_event_hdl_t handle, 1647 ddi_eventcookie_t cookie, int *attributes) 1648 { 1649 1650 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1651 ndi_event_cookie_t *cookie_list; 1652 1653 ASSERT(mutex_owned(&handle->ndi_evthdl_mutex)); 1654 1655 cookie_list = ndi_event_hdl->ndi_evthdl_cookie_list; 1656 while (cookie_list != NULL) { 1657 if (cookie_list == NDI_EVENT(cookie)) { 1658 if (attributes) 1659 *attributes = 1660 NDI_EVENT_ATTRIBUTES(cookie_list); 1661 1662 return (NDI_SUCCESS); 1663 } 1664 1665 cookie_list = cookie_list->next_cookie; 1666 } 1667 1668 return (NDI_FAILURE); 1669 } 1670 1671 /* 1672 * ndi_event_add_callback(): adds an event callback registration 1673 * to the event cookie defining this event. 1674 * 1675 * Refer also to bus_add_eventcall(9n) and ndi_busop_add_eventcall(9n). 1676 * 1677 * ndi_event_add_callback(9n) is intended to be used in 1678 * the nexus driver's bus_add_eventcall(9n) busop function. 1679 * 1680 * If the event is not defined by this bus nexus driver, 1681 * ndi_event_add_callback() will return NDI_FAILURE. 1682 */ 1683 int 1684 ndi_event_add_callback(ndi_event_hdl_t handle, dev_info_t *child_dip, 1685 ddi_eventcookie_t cookie, 1686 void (*event_callback)(dev_info_t *, 1687 ddi_eventcookie_t, void *arg, void *impldata), 1688 void *arg, 1689 uint_t flag, 1690 ddi_callback_id_t *cb_id) 1691 { 1692 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1693 int km_flag = ((flag & NDI_NOSLEEP) ? KM_NOSLEEP : KM_SLEEP); 1694 ndi_event_callbacks_t *cb; 1695 1696 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1697 1698 /* 1699 * if the event was not bound to this handle, return failure 1700 */ 1701 if (ndi_event_is_defined(handle, cookie, NULL) != NDI_SUCCESS) { 1702 1703 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1704 return (NDI_FAILURE); 1705 1706 } 1707 1708 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1709 1710 /* 1711 * allocate space for a callback structure 1712 */ 1713 cb = kmem_zalloc(sizeof (ndi_event_callbacks_t), km_flag); 1714 if (cb == NULL) { 1715 return (NDI_FAILURE); 1716 } 1717 1718 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1719 1720 /* initialize callback structure */ 1721 cb->ndi_evtcb_dip = child_dip; 1722 cb->ndi_evtcb_callback = event_callback; 1723 cb->ndi_evtcb_arg = arg; 1724 cb->ndi_evtcb_cookie = cookie; 1725 cb->devname = (char *)ddi_driver_name(child_dip); 1726 1727 *cb_id = (ddi_callback_id_t)cb; 1728 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1729 1730 /* add this callback structure to the list */ 1731 if (NDI_EVENT(cookie)->callback_list) { 1732 cb->ndi_evtcb_next = NDI_EVENT(cookie)->callback_list; 1733 NDI_EVENT(cookie)->callback_list->ndi_evtcb_prev = cb; 1734 NDI_EVENT(cookie)->callback_list = cb; 1735 } else { 1736 NDI_EVENT(cookie)->callback_list = cb; 1737 } 1738 #ifdef NDI_EVENT_DEBUG 1739 if (ndi_event_debug) { 1740 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_add_callback"); 1741 } 1742 #endif /* NDI_EVENT_DEBUG */ 1743 1744 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1745 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1746 1747 return (NDI_SUCCESS); 1748 } 1749 1750 /* 1751 * ndi_event_remove_callback(): 1752 * 1753 * ndi_event_remove_callback() removes a callback that was 1754 * previously registered using ndi_event_add_callback(9N). 1755 * Refer also to bus_remove_eventcall(9n) and 1756 * ndi_busop_remove_eventcall(9n). 1757 * ndi_event_remove_callback(9n) is intended to be used in 1758 * the nexus driver's bus_remove_eventcall (9n) busop function. 1759 * If the event is not defined by this bus nexus driver, 1760 * ndi_event_remove_callback() will return NDI_FAILURE. 1761 */ 1762 static void do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl, 1763 ddi_callback_id_t cb_id); 1764 1765 int 1766 ndi_event_remove_callback(ndi_event_hdl_t handle, ddi_callback_id_t cb_id) 1767 { 1768 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1769 1770 ASSERT(cb_id); 1771 1772 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1773 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1774 1775 do_ndi_event_remove_callback(ndi_event_hdl, cb_id); 1776 1777 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1778 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1779 1780 return (NDI_SUCCESS); 1781 } 1782 1783 /*ARGSUSED*/ 1784 static void 1785 do_ndi_event_remove_callback(struct ndi_event_hdl *ndi_event_hdl, 1786 ddi_callback_id_t cb_id) 1787 { 1788 ndi_event_callbacks_t *cb = (ndi_event_callbacks_t *)cb_id; 1789 ASSERT(cb); 1790 1791 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_mutex)); 1792 ASSERT(mutex_owned(&ndi_event_hdl->ndi_evthdl_cb_mutex)); 1793 1794 /* remove from callback linked list */ 1795 if (cb->ndi_evtcb_prev) { 1796 cb->ndi_evtcb_prev->ndi_evtcb_next = cb->ndi_evtcb_next; 1797 } 1798 1799 if (cb->ndi_evtcb_next) { 1800 cb->ndi_evtcb_next->ndi_evtcb_prev = cb->ndi_evtcb_prev; 1801 } 1802 1803 if (NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list == cb) { 1804 NDI_EVENT(cb->ndi_evtcb_cookie)->callback_list = 1805 cb->ndi_evtcb_next; 1806 } 1807 1808 kmem_free(cb, sizeof (ndi_event_callbacks_t)); 1809 } 1810 1811 /* 1812 * ndi_event_run_callbacks() performs event callbacks for the event 1813 * specified by cookie, if this is among those bound to the 1814 * supplied handle. 1815 * If the event is among those bound to the handle, none, 1816 * some, or all of the handlers registered for the event 1817 * will be called, according to the delivery attributes of 1818 * the event. 1819 * If the event attributes include NDI_EVENT_POST_TO_ALL 1820 * (the default), all the handlers for the event will be 1821 * called in an unspecified order. 1822 * If the event attributes include NDI_EVENT_POST_TO_TGT, only 1823 * the handlers (if any) registered by the driver identified by 1824 * rdip will be called. 1825 * If the event identified by cookie is not bound to the handle, 1826 * NDI_FAILURE will be returned. 1827 */ 1828 int 1829 ndi_event_run_callbacks(ndi_event_hdl_t handle, dev_info_t *child_dip, 1830 ddi_eventcookie_t cookie, void *bus_impldata) 1831 { 1832 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1833 ndi_event_callbacks_t *next, *cb; 1834 int attributes; 1835 1836 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1837 1838 /* if this is not our event, fail */ 1839 if (ndi_event_is_defined(handle, cookie, &attributes) != 1840 NDI_SUCCESS) { 1841 1842 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1843 return (NDI_FAILURE); 1844 } 1845 1846 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1847 1848 #ifdef NDI_EVENT_DEBUG 1849 if (ndi_event_debug) { 1850 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t" 1851 "producer dip=%p (%s%d): cookie = %p, name = %s\n", 1852 (void *)ndi_event_hdl->ndi_evthdl_dip, 1853 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip), 1854 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip), 1855 (void *)cookie, 1856 ndi_event_cookie_to_name(handle, cookie)); 1857 } 1858 #endif /* #ifdef NDI_EVENT_DEBUG */ 1859 1860 1861 /* 1862 * The callback handlers may call conversion functions. The conversion 1863 * functions may hold the ndi_evthdl_mutex during execution. Thus, to 1864 * avoid a recursive mutex problem, only the ndi_evthdl_cb_mutex is 1865 * held. The ndi_evthdl_mutex is not held when running the callbacks. 1866 */ 1867 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1868 1869 /* perform callbacks */ 1870 next = NDI_EVENT(cookie)->callback_list; 1871 while (next != NULL) { 1872 1873 cb = next; 1874 next = next->ndi_evtcb_next; 1875 1876 ASSERT(cb->ndi_evtcb_cookie == cookie); 1877 1878 if (attributes == NDI_EVENT_POST_TO_TGT && 1879 child_dip != cb->ndi_evtcb_dip) { 1880 continue; 1881 } 1882 1883 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip, cb->ndi_evtcb_cookie, 1884 cb->ndi_evtcb_arg, bus_impldata); 1885 1886 #ifdef NDI_EVENT_DEBUG 1887 if (ndi_event_debug) { 1888 cmn_err(CE_CONT, 1889 "\t\tconsumer dip=%p (%s%d)\n", 1890 (void *)cb->ndi_evtcb_dip, 1891 ddi_node_name(cb->ndi_evtcb_dip), 1892 ddi_get_instance(cb->ndi_evtcb_dip)); 1893 } 1894 #endif 1895 1896 } 1897 1898 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1899 1900 #ifdef NDI_EVENT_DEBUG 1901 if (ndi_event_debug) { 1902 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1903 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks"); 1904 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1905 } 1906 #endif /* NDI_EVENT_DEBUG */ 1907 1908 return (NDI_SUCCESS); 1909 } 1910 1911 1912 /* 1913 * perform one callback for a specified cookie and just one target 1914 */ 1915 int 1916 ndi_event_do_callback(ndi_event_hdl_t handle, dev_info_t *child_dip, 1917 ddi_eventcookie_t cookie, void *bus_impldata) 1918 { 1919 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 1920 ndi_event_callbacks_t *next, *cb; 1921 int attributes; 1922 1923 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1924 1925 /* if this is not our event, fail */ 1926 if (ndi_event_is_defined(handle, cookie, &attributes) != 1927 NDI_SUCCESS) { 1928 1929 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1930 1931 return (NDI_FAILURE); 1932 } 1933 1934 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1935 1936 #ifdef NDI_EVENT_DEBUG 1937 if (ndi_event_debug) { 1938 cmn_err(CE_CONT, "ndi_event_run_callbacks:\n\t" 1939 "producer dip=%p (%s%d): cookie = %p, name = %s\n", 1940 (void *)ndi_event_hdl->ndi_evthdl_dip, 1941 ddi_node_name(ndi_event_hdl->ndi_evthdl_dip), 1942 ddi_get_instance(ndi_event_hdl->ndi_evthdl_dip), 1943 (void *)cookie, 1944 ndi_event_cookie_to_name(handle, cookie)); 1945 } 1946 #endif 1947 1948 1949 /* 1950 * we only grab the cb mutex because the callback handlers 1951 * may call the conversion functions which would cause a recursive 1952 * mutex problem 1953 */ 1954 mutex_enter(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1955 1956 /* perform callbacks */ 1957 for (next = NDI_EVENT(cookie)->callback_list; next != NULL; ) { 1958 cb = next; 1959 next = next->ndi_evtcb_next; 1960 1961 if (cb->ndi_evtcb_dip == child_dip) { 1962 cb->ndi_evtcb_callback(cb->ndi_evtcb_dip, 1963 cb->ndi_evtcb_cookie, cb->ndi_evtcb_arg, 1964 bus_impldata); 1965 1966 #ifdef NDI_EVENT_DEBUG 1967 if (ndi_event_debug) { 1968 cmn_err(CE_CONT, 1969 "\t\tconsumer dip=%p (%s%d)\n", 1970 (void *)cb->ndi_evtcb_dip, 1971 ddi_node_name(cb->ndi_evtcb_dip), 1972 ddi_get_instance(cb->ndi_evtcb_dip)); 1973 } 1974 #endif 1975 break; 1976 } 1977 } 1978 1979 mutex_exit(&ndi_event_hdl->ndi_evthdl_cb_mutex); 1980 1981 #ifdef NDI_EVENT_DEBUG 1982 if (ndi_event_debug) { 1983 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 1984 ndi_event_dump_hdl(ndi_event_hdl, "ndi_event_run_callbacks"); 1985 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 1986 } 1987 #endif /* NDI_EVENT_DEBUG */ 1988 1989 return (NDI_SUCCESS); 1990 } 1991 1992 1993 /* 1994 * ndi_event_tag_to_cookie: utility function to find an event cookie 1995 * given an event tag 1996 */ 1997 ddi_eventcookie_t 1998 ndi_event_tag_to_cookie(ndi_event_hdl_t handle, int event_tag) 1999 { 2000 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2001 ndi_event_cookie_t *list; 2002 2003 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2004 2005 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2006 while (list != NULL) { 2007 if (NDI_EVENT_TAG(list) == event_tag) { 2008 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2009 return ((ddi_eventcookie_t)list); 2010 } 2011 2012 list = list->next_cookie; 2013 } 2014 2015 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2016 return (NULL); 2017 } 2018 2019 /* 2020 * ndi_event_cookie_to_tag: utility function to find a event tag 2021 * given an event_cookie 2022 */ 2023 int 2024 ndi_event_cookie_to_tag(ndi_event_hdl_t handle, ddi_eventcookie_t cookie) 2025 { 2026 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2027 ndi_event_cookie_t *list; 2028 2029 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2030 2031 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2032 2033 while (list != NULL) { 2034 if ((ddi_eventcookie_t)list == cookie) { 2035 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2036 return (NDI_EVENT_TAG(list)); 2037 } 2038 2039 list = list->next_cookie; 2040 } 2041 2042 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2043 return (NDI_FAILURE); 2044 2045 } 2046 2047 /* 2048 * ndi_event_cookie_to_name: utility function to find an event name 2049 * given an event_cookie 2050 */ 2051 char * 2052 ndi_event_cookie_to_name(ndi_event_hdl_t handle, ddi_eventcookie_t cookie) 2053 { 2054 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2055 ndi_event_cookie_t *list; 2056 2057 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2058 2059 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2060 2061 while (list != NULL) { 2062 if (list == NDI_EVENT(cookie)) { 2063 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2064 return (NDI_EVENT_NAME(list)); 2065 } 2066 2067 list = list->next_cookie; 2068 } 2069 2070 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2071 return (NULL); 2072 } 2073 2074 /* 2075 * ndi_event_tag_to_name: utility function to find an event name 2076 * given an event tag 2077 */ 2078 char * 2079 ndi_event_tag_to_name(ndi_event_hdl_t handle, int event_tag) 2080 { 2081 struct ndi_event_hdl *ndi_event_hdl = (struct ndi_event_hdl *)handle; 2082 ndi_event_cookie_t *list; 2083 2084 mutex_enter(&ndi_event_hdl->ndi_evthdl_mutex); 2085 2086 list = ndi_event_hdl->ndi_evthdl_cookie_list; 2087 2088 while (list) { 2089 if (NDI_EVENT_TAG(list) == event_tag) { 2090 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2091 return (NDI_EVENT_NAME(list)); 2092 } 2093 2094 list = list->next_cookie; 2095 } 2096 2097 mutex_exit(&ndi_event_hdl->ndi_evthdl_mutex); 2098 2099 return (NULL); 2100 } 2101 2102 #ifdef NDI_EVENT_DEBUG 2103 void 2104 ndi_event_dump_hdl(struct ndi_event_hdl *hdl, char *location) 2105 { 2106 2107 2108 ndi_event_callbacks_t *next; 2109 ndi_event_cookie_t *list; 2110 2111 ASSERT(mutex_owned(&hdl->ndi_evthdl_mutex)); 2112 list = hdl->ndi_evthdl_cookie_list; 2113 2114 cmn_err(CE_CONT, "%s: event handle (%p): dip = %p (%s%d)\n", 2115 location, (void *)hdl, (void *)hdl->ndi_evthdl_dip, 2116 ddi_node_name(hdl->ndi_evthdl_dip), 2117 ddi_get_instance(hdl->ndi_evthdl_dip)); 2118 cmn_err(CE_CONT, "\thigh=%d other=%d n=%d\n", 2119 hdl->ndi_evthdl_high_plevels, hdl->ndi_evthdl_other_plevels, 2120 hdl->ndi_evthdl_n_events); 2121 2122 cmn_err(CE_CONT, "\tevent cookies:\n"); 2123 while (list) { 2124 cmn_err(CE_CONT, "\t\ttag=%d name=%s p=%d a=%x dd=%p\n", 2125 NDI_EVENT_TAG(list), NDI_EVENT_NAME(list), 2126 NDI_EVENT_PLEVEL(list), NDI_EVENT_ATTRIBUTES(list), 2127 (void *)NDI_EVENT_DDIP(list)); 2128 cmn_err(CE_CONT, "\t\tcallbacks:\n"); 2129 for (next = list->callback_list; next != NULL; 2130 next = next->ndi_evtcb_next) { 2131 cmn_err(CE_CONT, 2132 "\t\t dip=%p (%s%d) cookie=%p arg=%p\n", 2133 (void*)next->ndi_evtcb_dip, 2134 ddi_driver_name(next->ndi_evtcb_dip), 2135 ddi_get_instance(next->ndi_evtcb_dip), 2136 (void *)next->ndi_evtcb_cookie, 2137 next->ndi_evtcb_arg); 2138 } 2139 2140 list = list->next_cookie; 2141 } 2142 2143 cmn_err(CE_CONT, "\n"); 2144 } 2145 #endif 2146 2147 int 2148 ndi_dev_is_prom_node(dev_info_t *dip) 2149 { 2150 return (DEVI(dip)->devi_node_class == DDI_NC_PROM); 2151 } 2152 2153 int 2154 ndi_dev_is_pseudo_node(dev_info_t *dip) 2155 { 2156 /* 2157 * NOTE: this does NOT mean the pseudo branch of the device tree, 2158 * it means the node was created by software (DEVI_SID_NODEID | 2159 * DEVI_PSEUDO_NODEID) instead of being generated from a PROM node. 2160 */ 2161 return (DEVI(dip)->devi_node_class == DDI_NC_PSEUDO); 2162 } 2163 2164 int 2165 ndi_dev_is_persistent_node(dev_info_t *dip) 2166 { 2167 return ((DEVI(dip)->devi_node_attributes & DDI_PERSISTENT) != 0); 2168 } 2169 2170 int 2171 i_ndi_dev_is_auto_assigned_node(dev_info_t *dip) 2172 { 2173 return ((DEVI(dip)->devi_node_attributes & 2174 DDI_AUTO_ASSIGNED_NODEID) != 0); 2175 } 2176 2177 void 2178 i_ndi_set_node_class(dev_info_t *dip, ddi_node_class_t c) 2179 { 2180 DEVI(dip)->devi_node_class = c; 2181 } 2182 2183 ddi_node_class_t 2184 i_ndi_get_node_class(dev_info_t *dip) 2185 { 2186 return (DEVI(dip)->devi_node_class); 2187 } 2188 2189 void 2190 i_ndi_set_node_attributes(dev_info_t *dip, int p) 2191 { 2192 DEVI(dip)->devi_node_attributes = p; 2193 } 2194 2195 int 2196 i_ndi_get_node_attributes(dev_info_t *dip) 2197 { 2198 return (DEVI(dip)->devi_node_attributes); 2199 } 2200 2201 void 2202 i_ndi_set_nodeid(dev_info_t *dip, int n) 2203 { 2204 DEVI(dip)->devi_nodeid = n; 2205 } 2206 2207 void 2208 ndi_set_acc_fault(ddi_acc_handle_t ah) 2209 { 2210 i_ddi_acc_set_fault(ah); 2211 } 2212 2213 void 2214 ndi_clr_acc_fault(ddi_acc_handle_t ah) 2215 { 2216 i_ddi_acc_clr_fault(ah); 2217 } 2218 2219 void 2220 ndi_set_dma_fault(ddi_dma_handle_t dh) 2221 { 2222 i_ddi_dma_set_fault(dh); 2223 } 2224 2225 void 2226 ndi_clr_dma_fault(ddi_dma_handle_t dh) 2227 { 2228 i_ddi_dma_clr_fault(dh); 2229 } 2230 2231 /* 2232 * The default fault-handler, called when the event posted by 2233 * ddi_dev_report_fault() reaches rootnex. 2234 */ 2235 static void 2236 i_ddi_fault_handler(dev_info_t *dip, struct ddi_fault_event_data *fedp) 2237 { 2238 ASSERT(fedp); 2239 2240 mutex_enter(&(DEVI(dip)->devi_lock)); 2241 if (!DEVI_IS_DEVICE_OFFLINE(dip)) { 2242 switch (fedp->f_impact) { 2243 case DDI_SERVICE_LOST: 2244 DEVI_SET_DEVICE_DOWN(dip); 2245 break; 2246 2247 case DDI_SERVICE_DEGRADED: 2248 DEVI_SET_DEVICE_DEGRADED(dip); 2249 break; 2250 2251 case DDI_SERVICE_UNAFFECTED: 2252 default: 2253 break; 2254 2255 case DDI_SERVICE_RESTORED: 2256 DEVI_SET_DEVICE_UP(dip); 2257 break; 2258 } 2259 } 2260 mutex_exit(&(DEVI(dip)->devi_lock)); 2261 } 2262 2263 /* 2264 * The default fault-logger, called when the event posted by 2265 * ddi_dev_report_fault() reaches rootnex. 2266 */ 2267 /*ARGSUSED*/ 2268 static void 2269 i_ddi_fault_logger(dev_info_t *rdip, struct ddi_fault_event_data *fedp) 2270 { 2271 ddi_devstate_t newstate; 2272 const char *action; 2273 const char *servstate; 2274 const char *location; 2275 int bad; 2276 int changed; 2277 int level; 2278 int still; 2279 2280 ASSERT(fedp); 2281 2282 bad = 0; 2283 switch (fedp->f_location) { 2284 case DDI_DATAPATH_FAULT: 2285 location = "in datapath to"; 2286 break; 2287 case DDI_DEVICE_FAULT: 2288 location = "in"; 2289 break; 2290 case DDI_EXTERNAL_FAULT: 2291 location = "external to"; 2292 break; 2293 default: 2294 location = "somewhere near"; 2295 bad = 1; 2296 break; 2297 } 2298 2299 newstate = ddi_get_devstate(fedp->f_dip); 2300 switch (newstate) { 2301 case DDI_DEVSTATE_OFFLINE: 2302 servstate = "unavailable"; 2303 break; 2304 case DDI_DEVSTATE_DOWN: 2305 servstate = "unavailable"; 2306 break; 2307 case DDI_DEVSTATE_QUIESCED: 2308 servstate = "suspended"; 2309 break; 2310 case DDI_DEVSTATE_DEGRADED: 2311 servstate = "degraded"; 2312 break; 2313 default: 2314 servstate = "available"; 2315 break; 2316 } 2317 2318 changed = (newstate != fedp->f_oldstate); 2319 level = (newstate < fedp->f_oldstate) ? CE_WARN : CE_NOTE; 2320 switch (fedp->f_impact) { 2321 case DDI_SERVICE_LOST: 2322 case DDI_SERVICE_DEGRADED: 2323 case DDI_SERVICE_UNAFFECTED: 2324 /* fault detected; service [still] <servstate> */ 2325 action = "fault detected"; 2326 still = !changed; 2327 break; 2328 2329 case DDI_SERVICE_RESTORED: 2330 if (newstate != DDI_DEVSTATE_UP) { 2331 /* fault cleared; service still <servstate> */ 2332 action = "fault cleared"; 2333 still = 1; 2334 } else if (changed) { 2335 /* fault cleared; service <servstate> */ 2336 action = "fault cleared"; 2337 still = 0; 2338 } else { 2339 /* no fault; service <servstate> */ 2340 action = "no fault"; 2341 still = 0; 2342 } 2343 break; 2344 2345 default: 2346 bad = 1; 2347 break; 2348 } 2349 2350 cmn_err(level, "!%s%d: %s %s device; service %s%s"+(bad|changed), 2351 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip), 2352 bad ? "invalid report of fault" : action, 2353 location, still ? "still " : "", servstate); 2354 2355 cmn_err(level, "!%s%d: %s"+(bad|changed), 2356 ddi_driver_name(fedp->f_dip), ddi_get_instance(fedp->f_dip), 2357 fedp->f_message); 2358 } 2359 2360 /* 2361 * Platform-settable pointers to fault handler and logger functions. 2362 * These are called by the default rootnex event-posting code when 2363 * a fault event reaches rootnex. 2364 */ 2365 void (*plat_fault_handler)(dev_info_t *, struct ddi_fault_event_data *) = 2366 i_ddi_fault_handler; 2367 void (*plat_fault_logger)(dev_info_t *, struct ddi_fault_event_data *) = 2368 i_ddi_fault_logger; 2369 2370 /* 2371 * Rootnex event definitions ... 2372 */ 2373 enum rootnex_event_tags { 2374 ROOTNEX_FAULT_EVENT 2375 }; 2376 static ndi_event_hdl_t rootnex_event_hdl; 2377 static ndi_event_definition_t rootnex_event_set[] = { 2378 { 2379 ROOTNEX_FAULT_EVENT, 2380 DDI_DEVI_FAULT_EVENT, 2381 EPL_INTERRUPT, 2382 NDI_EVENT_POST_TO_ALL 2383 } 2384 }; 2385 static ndi_event_set_t rootnex_events = { 2386 NDI_EVENTS_REV1, 2387 sizeof (rootnex_event_set) / sizeof (rootnex_event_set[0]), 2388 rootnex_event_set 2389 }; 2390 2391 /* 2392 * Initialize rootnex event handle 2393 */ 2394 void 2395 i_ddi_rootnex_init_events(dev_info_t *dip) 2396 { 2397 if (ndi_event_alloc_hdl(dip, (ddi_iblock_cookie_t)(LOCK_LEVEL-1), 2398 &rootnex_event_hdl, NDI_SLEEP) == NDI_SUCCESS) { 2399 if (ndi_event_bind_set(rootnex_event_hdl, 2400 &rootnex_events, NDI_SLEEP) != NDI_SUCCESS) { 2401 (void) ndi_event_free_hdl(rootnex_event_hdl); 2402 rootnex_event_hdl = NULL; 2403 } 2404 } 2405 } 2406 2407 /* 2408 * Event-handling functions for rootnex 2409 * These provide the standard implementation of fault handling 2410 */ 2411 /*ARGSUSED*/ 2412 int 2413 i_ddi_rootnex_get_eventcookie(dev_info_t *dip, dev_info_t *rdip, 2414 char *eventname, ddi_eventcookie_t *cookiep) 2415 { 2416 if (rootnex_event_hdl == NULL) 2417 return (NDI_FAILURE); 2418 return (ndi_event_retrieve_cookie(rootnex_event_hdl, rdip, eventname, 2419 cookiep, NDI_EVENT_NOPASS)); 2420 } 2421 2422 /*ARGSUSED*/ 2423 int 2424 i_ddi_rootnex_add_eventcall(dev_info_t *dip, dev_info_t *rdip, 2425 ddi_eventcookie_t eventid, void (*handler)(dev_info_t *dip, 2426 ddi_eventcookie_t event, void *arg, void *impl_data), void *arg, 2427 ddi_callback_id_t *cb_id) 2428 { 2429 if (rootnex_event_hdl == NULL) 2430 return (NDI_FAILURE); 2431 return (ndi_event_add_callback(rootnex_event_hdl, rdip, 2432 eventid, handler, arg, NDI_SLEEP, cb_id)); 2433 } 2434 2435 /*ARGSUSED*/ 2436 int 2437 i_ddi_rootnex_remove_eventcall(dev_info_t *dip, ddi_callback_id_t cb_id) 2438 { 2439 if (rootnex_event_hdl == NULL) 2440 return (NDI_FAILURE); 2441 2442 return (ndi_event_remove_callback(rootnex_event_hdl, cb_id)); 2443 } 2444 2445 /*ARGSUSED*/ 2446 int 2447 i_ddi_rootnex_post_event(dev_info_t *dip, dev_info_t *rdip, 2448 ddi_eventcookie_t eventid, void *impl_data) 2449 { 2450 int tag; 2451 2452 if (rootnex_event_hdl == NULL) 2453 return (NDI_FAILURE); 2454 2455 tag = ndi_event_cookie_to_tag(rootnex_event_hdl, eventid); 2456 if (tag == ROOTNEX_FAULT_EVENT) { 2457 (*plat_fault_handler)(rdip, impl_data); 2458 (*plat_fault_logger)(rdip, impl_data); 2459 } 2460 return (ndi_event_run_callbacks(rootnex_event_hdl, rdip, 2461 eventid, impl_data)); 2462 } 2463 2464 /* 2465 * ndi_set_bus_private/ndi_get_bus_private: 2466 * Get/set device bus private data in devinfo. 2467 */ 2468 void 2469 ndi_set_bus_private(dev_info_t *dip, boolean_t up, uint32_t port_type, 2470 void *data) 2471 { 2472 if (up) { 2473 DEVI(dip)->devi_bus.port_up.info.port.type = port_type; 2474 DEVI(dip)->devi_bus.port_up.priv_p = data; 2475 } else { 2476 DEVI(dip)->devi_bus.port_down.info.port.type = port_type; 2477 DEVI(dip)->devi_bus.port_down.priv_p = data; 2478 } 2479 } 2480 2481 void * 2482 ndi_get_bus_private(dev_info_t *dip, boolean_t up) 2483 { 2484 if (up) 2485 return (DEVI(dip)->devi_bus.port_up.priv_p); 2486 else 2487 return (DEVI(dip)->devi_bus.port_down.priv_p); 2488 } 2489 2490 boolean_t 2491 ndi_port_type(dev_info_t *dip, boolean_t up, uint32_t port_type) 2492 { 2493 if (up) { 2494 return ((DEVI(dip)->devi_bus.port_up.info.port.type) == 2495 port_type); 2496 } else { 2497 return ((DEVI(dip)->devi_bus.port_down.info.port.type) == 2498 port_type); 2499 } 2500 } 2501 2502 /* Interfaces for 'self' to set/get a child's flavor */ 2503 void 2504 ndi_flavor_set(dev_info_t *child, ndi_flavor_t child_flavor) 2505 { 2506 DEVI(child)->devi_flavor = child_flavor; 2507 } 2508 2509 ndi_flavor_t 2510 ndi_flavor_get(dev_info_t *child) 2511 { 2512 return (DEVI(child)->devi_flavor); 2513 } 2514 2515 /* 2516 * Interfaces to maintain flavor-specific private data of flavored 2517 * children of self. 2518 * 2519 * The flavor count always includes the default (0) vanilla flavor, 2520 * but storage for the vanilla flavor data pointer is in the same 2521 * place that ddi_[sg]et_driver_private uses, so the flavorv 2522 * storage is just for flavors 1..{nflavors-1}. 2523 */ 2524 void 2525 ndi_flavorv_alloc(dev_info_t *self, int nflavors) 2526 { 2527 ASSERT(nflavors > 0 && (DEVI(self)->devi_flavorv == NULL || 2528 nflavors == DEVI(self)->devi_flavorv_n)); 2529 if (nflavors <= 1 || (DEVI(self)->devi_flavorv)) { 2530 return; 2531 } 2532 DEVI(self)->devi_flavorv = 2533 kmem_zalloc((nflavors - 1) * sizeof (void *), KM_SLEEP); 2534 DEVI(self)->devi_flavorv_n = nflavors; 2535 } 2536 2537 void 2538 ndi_flavorv_set(dev_info_t *self, ndi_flavor_t child_flavor, void *v) 2539 { 2540 ASSERT(child_flavor < DEVI(self)->devi_flavorv_n && 2541 DEVI(self)->devi_flavorv != NULL); 2542 if (child_flavor > DEVI(self)->devi_flavorv_n || 2543 DEVI(self)->devi_flavorv == NULL) { 2544 return; 2545 } 2546 if (child_flavor == NDI_FLAVOR_VANILLA) { 2547 ddi_set_driver_private(self, v); 2548 } else { 2549 DEVI(self)->devi_flavorv[child_flavor - 1] = v; 2550 } 2551 } 2552 2553 void * 2554 ndi_flavorv_get(dev_info_t *self, ndi_flavor_t child_flavor) 2555 { 2556 ASSERT(child_flavor < DEVI(self)->devi_flavorv_n && 2557 DEVI(self)->devi_flavorv != NULL); 2558 if (child_flavor > DEVI(self)->devi_flavorv_n || 2559 DEVI(self)->devi_flavorv == NULL) { 2560 return (NULL); 2561 } 2562 if (child_flavor == NDI_FLAVOR_VANILLA) { 2563 return (ddi_get_driver_private(self)); 2564 } else { 2565 return (DEVI(self)->devi_flavorv[child_flavor - 1]); 2566 } 2567 } 2568