1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright (c) 2006, 2010, Oracle and/or its affiliates. All rights reserved. 23 * Copyright (c) 2018 by Delphix. All rights reserved. 24 * Copyright 2023 Oxide Computer Company 25 */ 26 27 #include <sys/note.h> 28 #include <sys/t_lock.h> 29 #include <sys/cmn_err.h> 30 #include <sys/instance.h> 31 #include <sys/conf.h> 32 #include <sys/stat.h> 33 #include <sys/ddi.h> 34 #include <sys/hwconf.h> 35 #include <sys/sunddi.h> 36 #include <sys/sunndi.h> 37 #include <sys/sunmdi.h> 38 #include <sys/ddi_impldefs.h> 39 #include <sys/ndi_impldefs.h> 40 #include <sys/kobj.h> 41 #include <sys/devcache.h> 42 #include <sys/devid_cache.h> 43 #include <sys/sysmacros.h> 44 45 /* 46 * Discovery refers to the heroic effort made to discover a device which 47 * cannot be accessed at the physical path where it once resided. Discovery 48 * involves walking the entire device tree attaching all possible disk 49 * instances, to search for the device referenced by a devid. Obviously, 50 * full device discovery is something to be avoided where possible. 51 * Note that simply invoking devfsadm(8) is equivalent to running full 52 * discovery at the devid cache level. 53 * 54 * Reasons why a disk may not be accessible: 55 * disk powered off 56 * disk removed or cable disconnected 57 * disk or adapter broken 58 * 59 * Note that discovery is not needed and cannot succeed in any of these 60 * cases. 61 * 62 * When discovery may succeed: 63 * Discovery will result in success when a device has been moved 64 * to a different address. Note that it's recommended that 65 * devfsadm(8) be invoked (no arguments required) whenever a system's 66 * h/w configuration has been updated. Alternatively, a 67 * reconfiguration boot can be used to accomplish the same result. 68 * 69 * Note that discovery is not necessary to be able to correct an access 70 * failure for a device which was powered off. Assuming the cache has an 71 * entry for such a device, simply powering it on should permit the system 72 * to access it. If problems persist after powering it on, invoke 73 * devfsadm(8). 74 * 75 * Discovery prior to mounting root is only of interest when booting 76 * from a filesystem which accesses devices by device id, which of 77 * not all do. 78 * 79 * Tunables 80 * 81 * devid_discovery_boot (default 1) 82 * Number of times discovery will be attempted prior to mounting root. 83 * Must be done at least once to recover from corrupted or missing 84 * devid cache backing store. Probably there's no reason to ever 85 * set this to greater than one as a missing device will remain 86 * unavailable no matter how often the system searches for it. 87 * 88 * devid_discovery_postboot (default 1) 89 * Number of times discovery will be attempted after mounting root. 90 * This must be performed at least once to discover any devices 91 * needed after root is mounted which may have been powered 92 * off and moved before booting. 93 * Setting this to a larger positive number will introduce 94 * some inconsistency in system operation. Searching for a device 95 * will take an indeterminate amount of time, sometimes slower, 96 * sometimes faster. In addition, the system will sometimes 97 * discover a newly powered on device, sometimes it won't. 98 * Use of this option is not therefore recommended. 99 * 100 * devid_discovery_postboot_always (default 0) 101 * Set to 1, the system will always attempt full discovery. 102 * 103 * devid_discovery_secs (default 0) 104 * Set to a positive value, the system will attempt full discovery 105 * but with a minimum delay between attempts. A device search 106 * within the period of time specified will result in failure. 107 * 108 * devid_cache_read_disable (default 0) 109 * Set to 1 to disable reading /etc/devices/devid_cache. 110 * Devid cache will continue to operate normally but 111 * at least one discovery attempt will be required. 112 * 113 * devid_cache_write_disable (default 0) 114 * Set to 1 to disable updates to /etc/devices/devid_cache. 115 * Any updates to the devid cache will not be preserved across a reboot. 116 * 117 * devid_report_error (default 0) 118 * Set to 1 to enable some error messages related to devid 119 * cache failures. 120 * 121 * The devid is packed in the cache file as a byte array. For 122 * portability, this could be done in the encoded string format. 123 */ 124 125 126 int devid_discovery_boot = 1; 127 int devid_discovery_postboot = 1; 128 int devid_discovery_postboot_always = 0; 129 int devid_discovery_secs = 0; 130 131 int devid_cache_read_disable = 0; 132 int devid_cache_write_disable = 0; 133 134 int devid_report_error = 0; 135 136 137 /* 138 * State to manage discovery of devices providing a devid 139 */ 140 static int devid_discovery_busy = 0; 141 static kmutex_t devid_discovery_mutex; 142 static kcondvar_t devid_discovery_cv; 143 static clock_t devid_last_discovery = 0; 144 145 146 #ifdef DEBUG 147 int nvp_devid_debug = 0; 148 int devid_debug = 0; 149 int devid_log_registers = 0; 150 int devid_log_finds = 0; 151 int devid_log_lookups = 0; 152 int devid_log_discovery = 0; 153 int devid_log_matches = 0; 154 int devid_log_paths = 0; 155 int devid_log_failures = 0; 156 int devid_log_hold = 0; 157 int devid_log_unregisters = 0; 158 int devid_log_removes = 0; 159 int devid_register_debug = 0; 160 int devid_log_stale = 0; 161 int devid_log_detaches = 0; 162 #endif /* DEBUG */ 163 164 /* 165 * devid cache file registration for cache reads and updates 166 */ 167 static nvf_ops_t devid_cache_ops = { 168 "/etc/devices/devid_cache", /* path to cache */ 169 devid_cache_unpack_nvlist, /* read: nvlist to nvp */ 170 devid_cache_pack_list, /* write: nvp to nvlist */ 171 devid_list_free, /* free data list */ 172 NULL /* write complete callback */ 173 }; 174 175 /* 176 * handle to registered devid cache handlers 177 */ 178 nvf_handle_t dcfd_handle; 179 180 181 /* 182 * Initialize devid cache file management 183 */ 184 void 185 devid_cache_init(void) 186 { 187 dcfd_handle = nvf_register_file(&devid_cache_ops); 188 ASSERT(dcfd_handle); 189 190 list_create(nvf_list(dcfd_handle), sizeof (nvp_devid_t), 191 offsetof(nvp_devid_t, nvp_link)); 192 193 mutex_init(&devid_discovery_mutex, NULL, MUTEX_DEFAULT, NULL); 194 cv_init(&devid_discovery_cv, NULL, CV_DRIVER, NULL); 195 } 196 197 /* 198 * Read and initialize the devid cache from the persistent store 199 */ 200 void 201 devid_cache_read(void) 202 { 203 if (!devid_cache_read_disable) { 204 rw_enter(nvf_lock(dcfd_handle), RW_WRITER); 205 ASSERT(list_head(nvf_list(dcfd_handle)) == NULL); 206 (void) nvf_read_file(dcfd_handle); 207 rw_exit(nvf_lock(dcfd_handle)); 208 } 209 } 210 211 static void 212 devid_nvp_free(nvp_devid_t *dp) 213 { 214 if (dp->nvp_devpath) 215 kmem_free(dp->nvp_devpath, strlen(dp->nvp_devpath)+1); 216 if (dp->nvp_devid) 217 kmem_free(dp->nvp_devid, ddi_devid_sizeof(dp->nvp_devid)); 218 219 kmem_free(dp, sizeof (nvp_devid_t)); 220 } 221 222 static void 223 devid_list_free(nvf_handle_t fd) 224 { 225 list_t *listp; 226 nvp_devid_t *np; 227 228 ASSERT(RW_WRITE_HELD(nvf_lock(dcfd_handle))); 229 230 listp = nvf_list(fd); 231 while (np = list_head(listp)) { 232 list_remove(listp, np); 233 devid_nvp_free(np); 234 } 235 } 236 237 /* 238 * Free an nvp element in a list 239 */ 240 static void 241 devid_nvp_unlink_and_free(nvf_handle_t fd, nvp_devid_t *np) 242 { 243 list_remove(nvf_list(fd), np); 244 devid_nvp_free(np); 245 } 246 247 /* 248 * Unpack a device path/nvlist pair to the list of devid cache elements. 249 * Used to parse the nvlist format when reading 250 * /etc/devices/devid_cache 251 */ 252 static int 253 devid_cache_unpack_nvlist(nvf_handle_t fd, nvlist_t *nvl, char *name) 254 { 255 nvp_devid_t *np; 256 ddi_devid_t devidp; 257 int rval; 258 uint_t n; 259 260 NVP_DEVID_DEBUG_PATH((name)); 261 ASSERT(RW_WRITE_HELD(nvf_lock(dcfd_handle))); 262 263 /* 264 * check path for a devid 265 */ 266 rval = nvlist_lookup_byte_array(nvl, 267 DP_DEVID_ID, (uchar_t **)&devidp, &n); 268 if (rval == 0) { 269 if (ddi_devid_valid(devidp) == DDI_SUCCESS) { 270 ASSERT(n == ddi_devid_sizeof(devidp)); 271 np = kmem_zalloc(sizeof (nvp_devid_t), KM_SLEEP); 272 np->nvp_devpath = i_ddi_strdup(name, KM_SLEEP); 273 np->nvp_devid = kmem_alloc(n, KM_SLEEP); 274 (void) bcopy(devidp, np->nvp_devid, n); 275 list_insert_tail(nvf_list(fd), np); 276 NVP_DEVID_DEBUG_DEVID((np->nvp_devid)); 277 } else { 278 DEVIDERR((CE_CONT, 279 "%s: invalid devid\n", name)); 280 } 281 } else { 282 DEVIDERR((CE_CONT, 283 "%s: devid not available\n", name)); 284 } 285 286 return (0); 287 } 288 289 /* 290 * Pack the list of devid cache elements into a single nvlist 291 * Used when writing the nvlist file. 292 */ 293 static int 294 devid_cache_pack_list(nvf_handle_t fd, nvlist_t **ret_nvl) 295 { 296 nvlist_t *nvl, *sub_nvl; 297 nvp_devid_t *np; 298 int rval; 299 list_t *listp; 300 301 ASSERT(RW_WRITE_HELD(nvf_lock(dcfd_handle))); 302 303 rval = nvlist_alloc(&nvl, NV_UNIQUE_NAME, KM_SLEEP); 304 if (rval != 0) { 305 nvf_error("%s: nvlist alloc error %d\n", 306 nvf_cache_name(fd), rval); 307 return (DDI_FAILURE); 308 } 309 310 listp = nvf_list(fd); 311 for (np = list_head(listp); np; np = list_next(listp, np)) { 312 if (np->nvp_devid == NULL) 313 continue; 314 NVP_DEVID_DEBUG_PATH(np->nvp_devpath); 315 rval = nvlist_alloc(&sub_nvl, NV_UNIQUE_NAME, KM_SLEEP); 316 if (rval != 0) { 317 nvf_error("%s: nvlist alloc error %d\n", 318 nvf_cache_name(fd), rval); 319 sub_nvl = NULL; 320 goto err; 321 } 322 323 rval = nvlist_add_byte_array(sub_nvl, DP_DEVID_ID, 324 (uchar_t *)np->nvp_devid, 325 ddi_devid_sizeof(np->nvp_devid)); 326 if (rval == 0) { 327 NVP_DEVID_DEBUG_DEVID(np->nvp_devid); 328 } else { 329 nvf_error( 330 "%s: nvlist add error %d (devid)\n", 331 nvf_cache_name(fd), rval); 332 goto err; 333 } 334 335 rval = nvlist_add_nvlist(nvl, np->nvp_devpath, sub_nvl); 336 if (rval != 0) { 337 nvf_error("%s: nvlist add error %d (sublist)\n", 338 nvf_cache_name(fd), rval); 339 goto err; 340 } 341 nvlist_free(sub_nvl); 342 } 343 344 *ret_nvl = nvl; 345 return (DDI_SUCCESS); 346 347 err: 348 nvlist_free(sub_nvl); 349 nvlist_free(nvl); 350 *ret_nvl = NULL; 351 return (DDI_FAILURE); 352 } 353 354 static int 355 e_devid_do_discovery(void) 356 { 357 ASSERT(mutex_owned(&devid_discovery_mutex)); 358 359 if (i_ddi_io_initialized() == 0) { 360 if (devid_discovery_boot > 0) { 361 devid_discovery_boot--; 362 return (1); 363 } 364 } else { 365 if (devid_discovery_postboot_always > 0) 366 return (1); 367 if (devid_discovery_postboot > 0) { 368 devid_discovery_postboot--; 369 return (1); 370 } 371 if (devid_discovery_secs > 0) { 372 if ((ddi_get_lbolt() - devid_last_discovery) > 373 drv_usectohz(devid_discovery_secs * MICROSEC)) { 374 return (1); 375 } 376 } 377 } 378 379 DEVID_LOG_DISC((CE_CONT, "devid_discovery: no discovery\n")); 380 return (0); 381 } 382 383 static void 384 e_ddi_devid_hold_by_major(major_t major) 385 { 386 DEVID_LOG_DISC((CE_CONT, 387 "devid_discovery: ddi_hold_installed_driver %d\n", major)); 388 389 if (ddi_hold_installed_driver(major) == NULL) 390 return; 391 392 ddi_rele_driver(major); 393 } 394 395 /* legacy support - see below */ 396 static char *e_ddi_devid_hold_driver_list[] = { "sd", "ssd" }; 397 398 #define N_DRIVERS_TO_HOLD \ 399 (sizeof (e_ddi_devid_hold_driver_list) / sizeof (char *)) 400 401 static void 402 e_ddi_devid_hold_installed_driver(ddi_devid_t devid) 403 { 404 impl_devid_t *id = (impl_devid_t *)devid; 405 major_t major, hint_major; 406 char hint[DEVID_HINT_SIZE + 1]; 407 struct devnames *dnp; 408 char **drvp; 409 int i; 410 411 /* Count non-null bytes */ 412 for (i = 0; i < DEVID_HINT_SIZE; i++) 413 if (id->did_driver[i] == '\0') 414 break; 415 416 /* Make a copy of the driver hint */ 417 bcopy(id->did_driver, hint, i); 418 hint[i] = '\0'; 419 420 /* search for the devid using the hint driver */ 421 hint_major = ddi_name_to_major(hint); 422 if (hint_major != DDI_MAJOR_T_NONE) { 423 e_ddi_devid_hold_by_major(hint_major); 424 } 425 426 /* 427 * search for the devid with each driver declaring 428 * itself as a devid registrant. 429 */ 430 for (major = 0; major < devcnt; major++) { 431 if (major == hint_major) 432 continue; 433 dnp = &devnamesp[major]; 434 if (dnp->dn_flags & DN_DEVID_REGISTRANT) { 435 e_ddi_devid_hold_by_major(major); 436 } 437 } 438 439 /* 440 * Legacy support: may be removed once an upgrade mechanism 441 * for driver conf files is available. 442 */ 443 drvp = e_ddi_devid_hold_driver_list; 444 for (i = 0; i < N_DRIVERS_TO_HOLD; i++, drvp++) { 445 major = ddi_name_to_major(*drvp); 446 if (major != DDI_MAJOR_T_NONE && major != hint_major) { 447 e_ddi_devid_hold_by_major(major); 448 } 449 } 450 } 451 452 /* 453 * Return success if discovery was attempted, to indicate 454 * that the desired device may now be available. 455 */ 456 int 457 e_ddi_devid_discovery(ddi_devid_t devid) 458 { 459 int flags; 460 int rval = DDI_SUCCESS; 461 462 mutex_enter(&devid_discovery_mutex); 463 464 if (devid_discovery_busy) { 465 DEVID_LOG_DISC((CE_CONT, "devid_discovery: busy\n")); 466 while (devid_discovery_busy) { 467 cv_wait(&devid_discovery_cv, &devid_discovery_mutex); 468 } 469 } else if (e_devid_do_discovery()) { 470 devid_discovery_busy = 1; 471 mutex_exit(&devid_discovery_mutex); 472 473 if (i_ddi_io_initialized() == 0) { 474 e_ddi_devid_hold_installed_driver(devid); 475 } else { 476 DEVID_LOG_DISC((CE_CONT, 477 "devid_discovery: ndi_devi_config\n")); 478 flags = NDI_DEVI_PERSIST | NDI_CONFIG | NDI_NO_EVENT; 479 if (i_ddi_io_initialized()) 480 flags |= NDI_DRV_CONF_REPROBE; 481 (void) ndi_devi_config(ddi_root_node(), flags); 482 } 483 484 mutex_enter(&devid_discovery_mutex); 485 devid_discovery_busy = 0; 486 cv_broadcast(&devid_discovery_cv); 487 if (devid_discovery_secs > 0) 488 devid_last_discovery = ddi_get_lbolt(); 489 DEVID_LOG_DISC((CE_CONT, "devid_discovery: done\n")); 490 } else { 491 rval = DDI_FAILURE; 492 DEVID_LOG_DISC((CE_CONT, "no devid discovery\n")); 493 } 494 495 mutex_exit(&devid_discovery_mutex); 496 497 return (rval); 498 } 499 500 /* 501 * As part of registering a devid for a device, 502 * update the devid cache with this device/devid pair 503 * or note that this combination has registered. 504 * 505 * If a devpath is provided it will be used as the path to register the 506 * devid against, otherwise we use ddi_pathname(dip). In both cases 507 * we duplicate the path string so that it can be cached/freed indepdently 508 * of the original owner. 509 */ 510 static int 511 e_devid_cache_register_cmn(dev_info_t *dip, ddi_devid_t devid, char *devpath) 512 { 513 nvp_devid_t *np; 514 nvp_devid_t *new_nvp; 515 ddi_devid_t new_devid; 516 int new_devid_size; 517 char *path, *fullpath; 518 ddi_devid_t free_devid = NULL; 519 int pathlen; 520 list_t *listp; 521 int is_dirty = 0; 522 523 524 ASSERT(ddi_devid_valid(devid) == DDI_SUCCESS); 525 526 if (devpath) { 527 pathlen = strlen(devpath) + 1; 528 path = kmem_alloc(pathlen, KM_SLEEP); 529 bcopy(devpath, path, pathlen); 530 } else { 531 /* 532 * We are willing to accept DS_BOUND nodes if we can form a full 533 * ddi_pathname (i.e. the node is part way to becomming 534 * DS_INITIALIZED and devi_addr/ddi_get_name_addr are non-NULL). 535 */ 536 if (ddi_get_name_addr(dip) == NULL) 537 return (DDI_FAILURE); 538 539 fullpath = kmem_alloc(MAXPATHLEN, KM_SLEEP); 540 (void) ddi_pathname(dip, fullpath); 541 pathlen = strlen(fullpath) + 1; 542 path = kmem_alloc(pathlen, KM_SLEEP); 543 bcopy(fullpath, path, pathlen); 544 kmem_free(fullpath, MAXPATHLEN); 545 } 546 547 DEVID_LOG_REG(("register", devid, path)); 548 549 new_nvp = kmem_zalloc(sizeof (nvp_devid_t), KM_SLEEP); 550 new_devid_size = ddi_devid_sizeof(devid); 551 new_devid = kmem_alloc(new_devid_size, KM_SLEEP); 552 (void) bcopy(devid, new_devid, new_devid_size); 553 554 rw_enter(nvf_lock(dcfd_handle), RW_WRITER); 555 556 listp = nvf_list(dcfd_handle); 557 for (np = list_head(listp); np; np = list_next(listp, np)) { 558 if (strcmp(path, np->nvp_devpath) == 0) { 559 DEVID_DEBUG2((CE_CONT, 560 "register: %s path match\n", path)); 561 if (np->nvp_devid == NULL) { 562 replace: np->nvp_devid = new_devid; 563 np->nvp_flags |= 564 NVP_DEVID_DIP | NVP_DEVID_REGISTERED; 565 np->nvp_dip = dip; 566 if (!devid_cache_write_disable) { 567 nvf_mark_dirty(dcfd_handle); 568 is_dirty = 1; 569 } 570 rw_exit(nvf_lock(dcfd_handle)); 571 kmem_free(new_nvp, sizeof (nvp_devid_t)); 572 kmem_free(path, pathlen); 573 goto exit; 574 } 575 if (ddi_devid_valid(np->nvp_devid) != DDI_SUCCESS) { 576 /* replace invalid devid */ 577 free_devid = np->nvp_devid; 578 goto replace; 579 } 580 /* 581 * We're registering an already-cached path 582 * Does the device's devid match the cache? 583 */ 584 if (ddi_devid_compare(devid, np->nvp_devid) != 0) { 585 DEVID_DEBUG((CE_CONT, "devid register: " 586 "devid %s does not match\n", path)); 587 /* 588 * We do not expect devids to change, log it. 589 */ 590 char *devid_stored = 591 ddi_devid_str_encode(np->nvp_devid, NULL); 592 char *devid_new = 593 ddi_devid_str_encode(devid, NULL); 594 595 cmn_err(CE_CONT, "devid register: devid for " 596 "%s does not match. stored: %s, new: %s.", 597 path, devid_stored, devid_new); 598 599 ddi_devid_str_free(devid_stored); 600 ddi_devid_str_free(devid_new); 601 602 /* 603 * Replace cached devid for this path 604 * with newly registered devid. A devid 605 * may map to multiple paths but one path 606 * should only map to one devid. 607 */ 608 devid_nvp_unlink_and_free(dcfd_handle, np); 609 np = NULL; 610 break; 611 } else { 612 DEVID_DEBUG2((CE_CONT, 613 "devid register: %s devid match\n", path)); 614 np->nvp_flags |= 615 NVP_DEVID_DIP | NVP_DEVID_REGISTERED; 616 np->nvp_dip = dip; 617 rw_exit(nvf_lock(dcfd_handle)); 618 kmem_free(new_nvp, sizeof (nvp_devid_t)); 619 kmem_free(path, pathlen); 620 kmem_free(new_devid, new_devid_size); 621 return (DDI_SUCCESS); 622 } 623 } 624 } 625 626 /* 627 * Add newly registered devid to the cache 628 */ 629 ASSERT(np == NULL); 630 631 new_nvp->nvp_devpath = path; 632 new_nvp->nvp_flags = NVP_DEVID_DIP | NVP_DEVID_REGISTERED; 633 new_nvp->nvp_dip = dip; 634 new_nvp->nvp_devid = new_devid; 635 636 if (!devid_cache_write_disable) { 637 is_dirty = 1; 638 nvf_mark_dirty(dcfd_handle); 639 } 640 list_insert_tail(nvf_list(dcfd_handle), new_nvp); 641 642 rw_exit(nvf_lock(dcfd_handle)); 643 644 exit: 645 if (free_devid) 646 kmem_free(free_devid, ddi_devid_sizeof(free_devid)); 647 648 if (is_dirty) 649 nvf_wake_daemon(); 650 651 return (DDI_SUCCESS); 652 } 653 654 int 655 e_devid_cache_register(dev_info_t *dip, ddi_devid_t devid) 656 { 657 return (e_devid_cache_register_cmn(dip, devid, NULL)); 658 } 659 660 /* 661 * Unregister a device's devid; the devinfo may hit on multiple entries 662 * arising from both pHCI and vHCI paths. 663 * Called as an instance detachs. 664 * Invalidate the devid's devinfo reference. 665 * Devid-path remains in the cache. 666 */ 667 668 void 669 e_devid_cache_unregister(dev_info_t *dip) 670 { 671 nvp_devid_t *np; 672 list_t *listp; 673 674 rw_enter(nvf_lock(dcfd_handle), RW_WRITER); 675 676 listp = nvf_list(dcfd_handle); 677 for (np = list_head(listp); np; np = list_next(listp, np)) { 678 if (np->nvp_devid == NULL) 679 continue; 680 if ((np->nvp_flags & NVP_DEVID_DIP) && np->nvp_dip == dip) { 681 DEVID_LOG_UNREG((CE_CONT, 682 "unregister: %s\n", np->nvp_devpath)); 683 np->nvp_flags &= ~NVP_DEVID_DIP; 684 np->nvp_dip = NULL; 685 } 686 } 687 688 rw_exit(nvf_lock(dcfd_handle)); 689 } 690 691 int 692 e_devid_cache_pathinfo(mdi_pathinfo_t *pip, ddi_devid_t devid) 693 { 694 char *path = mdi_pi_pathname(pip); 695 696 return (e_devid_cache_register_cmn(mdi_pi_get_client(pip), devid, 697 path)); 698 } 699 700 /* 701 * Purge devid cache of stale devids 702 */ 703 void 704 devid_cache_cleanup(void) 705 { 706 nvp_devid_t *np, *next; 707 list_t *listp; 708 int is_dirty = 0; 709 710 rw_enter(nvf_lock(dcfd_handle), RW_WRITER); 711 712 listp = nvf_list(dcfd_handle); 713 for (np = list_head(listp); np; np = next) { 714 next = list_next(listp, np); 715 if (np->nvp_devid == NULL) 716 continue; 717 if ((np->nvp_flags & NVP_DEVID_REGISTERED) == 0) { 718 DEVID_LOG_REMOVE((CE_CONT, 719 "cleanup: %s\n", np->nvp_devpath)); 720 if (!devid_cache_write_disable) { 721 nvf_mark_dirty(dcfd_handle); 722 is_dirty = 0; 723 } 724 devid_nvp_unlink_and_free(dcfd_handle, np); 725 } 726 } 727 728 rw_exit(nvf_lock(dcfd_handle)); 729 730 if (is_dirty) 731 nvf_wake_daemon(); 732 } 733 734 735 /* 736 * Build a list of dev_t's for a device/devid 737 * 738 * The effect of this function is cumulative, adding dev_t's 739 * for the device to the list of all dev_t's for a given 740 * devid. 741 */ 742 static void 743 e_devid_minor_to_devlist( 744 dev_info_t *dip, 745 char *minor_name, 746 int ndevts_alloced, 747 int *devtcntp, 748 dev_t *devtsp) 749 { 750 struct ddi_minor_data *dmdp; 751 int minor_all = 0; 752 int ndevts = *devtcntp; 753 754 ASSERT(i_ddi_devi_attached(dip)); 755 756 /* are we looking for a set of minor nodes? */ 757 if ((minor_name == DEVID_MINOR_NAME_ALL) || 758 (minor_name == DEVID_MINOR_NAME_ALL_CHR) || 759 (minor_name == DEVID_MINOR_NAME_ALL_BLK)) 760 minor_all = 1; 761 762 /* Find matching minor names */ 763 ndi_devi_enter(dip); 764 for (dmdp = DEVI(dip)->devi_minor; dmdp; dmdp = dmdp->next) { 765 766 /* Skip non-minors, and non matching minor names */ 767 if ((dmdp->type != DDM_MINOR) || ((minor_all == 0) && 768 strcmp(dmdp->ddm_name, minor_name))) 769 continue; 770 771 /* filter out minor_all mismatches */ 772 if (minor_all && 773 (((minor_name == DEVID_MINOR_NAME_ALL_CHR) && 774 (dmdp->ddm_spec_type != S_IFCHR)) || 775 ((minor_name == DEVID_MINOR_NAME_ALL_BLK) && 776 (dmdp->ddm_spec_type != S_IFBLK)))) 777 continue; 778 779 if (ndevts < ndevts_alloced) 780 devtsp[ndevts] = dmdp->ddm_dev; 781 ndevts++; 782 } 783 ndi_devi_exit(dip); 784 785 *devtcntp = ndevts; 786 } 787 788 /* 789 * Search for cached entries matching a devid 790 * Return two lists: 791 * a list of dev_info nodes, for those devices in the attached state 792 * a list of pathnames whose instances registered the given devid 793 * If the lists passed in are not sufficient to return the matching 794 * references, return the size of lists required. 795 * The dev_info nodes are returned with a hold that the caller must release. 796 */ 797 static int 798 e_devid_cache_devi_path_lists(ddi_devid_t devid, int retmax, 799 int *retndevis, dev_info_t **retdevis, int *retnpaths, char **retpaths) 800 { 801 nvp_devid_t *np; 802 int ndevis, npaths; 803 dev_info_t *dip, *pdip; 804 int maxdevis = 0; 805 int maxpaths = 0; 806 list_t *listp; 807 808 ndevis = 0; 809 npaths = 0; 810 listp = nvf_list(dcfd_handle); 811 for (np = list_head(listp); np; np = list_next(listp, np)) { 812 if (np->nvp_devid == NULL) 813 continue; 814 if (ddi_devid_valid(np->nvp_devid) != DDI_SUCCESS) { 815 DEVIDERR((CE_CONT, 816 "find: invalid devid %s\n", 817 np->nvp_devpath)); 818 continue; 819 } 820 if (ddi_devid_compare(devid, np->nvp_devid) == 0) { 821 DEVID_DEBUG2((CE_CONT, 822 "find: devid match: %s 0x%x\n", 823 np->nvp_devpath, np->nvp_flags)); 824 DEVID_LOG_MATCH(("find", devid, np->nvp_devpath)); 825 DEVID_LOG_PATHS((CE_CONT, "%s\n", np->nvp_devpath)); 826 827 /* 828 * Check if we have a cached devinfo reference for this 829 * devid. Place a hold on it to prevent detach 830 * Otherwise, use the path instead. 831 * Note: returns with a hold on each dev_info 832 * node in the list. 833 */ 834 dip = NULL; 835 if (np->nvp_flags & NVP_DEVID_DIP) { 836 pdip = ddi_get_parent(np->nvp_dip); 837 if (ndi_devi_tryenter(pdip)) { 838 dip = np->nvp_dip; 839 ndi_hold_devi(dip); 840 ndi_devi_exit(pdip); 841 ASSERT(!DEVI_IS_ATTACHING(dip)); 842 ASSERT(!DEVI_IS_DETACHING(dip)); 843 } else { 844 DEVID_LOG_DETACH((CE_CONT, 845 "may be detaching: %s\n", 846 np->nvp_devpath)); 847 } 848 } 849 850 if (dip) { 851 if (ndevis < retmax) { 852 retdevis[ndevis++] = dip; 853 } else { 854 ndi_rele_devi(dip); 855 } 856 maxdevis++; 857 } else { 858 if (npaths < retmax) 859 retpaths[npaths++] = np->nvp_devpath; 860 maxpaths++; 861 } 862 } 863 } 864 865 *retndevis = ndevis; 866 *retnpaths = npaths; 867 return (maxdevis > maxpaths ? maxdevis : maxpaths); 868 } 869 870 871 /* 872 * Search the devid cache, returning dev_t list for all 873 * device paths mapping to the device identified by the 874 * given devid. 875 * 876 * Primary interface used by ddi_lyr_devid_to_devlist() 877 */ 878 int 879 e_devid_cache_to_devt_list(ddi_devid_t devid, char *minor_name, 880 int *retndevts, dev_t **retdevts) 881 { 882 char *path, **paths; 883 int i, j, n; 884 dev_t *devts, *udevts; 885 dev_t tdevt; 886 int ndevts, undevts, ndevts_alloced; 887 dev_info_t *devi, **devis; 888 int ndevis, npaths, nalloced; 889 ddi_devid_t match_devid; 890 891 DEVID_LOG_FIND(("find", devid, NULL)); 892 893 ASSERT(ddi_devid_valid(devid) == DDI_SUCCESS); 894 if (ddi_devid_valid(devid) != DDI_SUCCESS) { 895 DEVID_LOG_ERR(("invalid devid", devid, NULL)); 896 return (DDI_FAILURE); 897 } 898 899 nalloced = 128; 900 901 for (;;) { 902 paths = kmem_zalloc(nalloced * sizeof (char *), KM_SLEEP); 903 devis = kmem_zalloc(nalloced * sizeof (dev_info_t *), KM_SLEEP); 904 905 rw_enter(nvf_lock(dcfd_handle), RW_READER); 906 n = e_devid_cache_devi_path_lists(devid, nalloced, 907 &ndevis, devis, &npaths, paths); 908 if (n <= nalloced) 909 break; 910 rw_exit(nvf_lock(dcfd_handle)); 911 for (i = 0; i < ndevis; i++) 912 ndi_rele_devi(devis[i]); 913 kmem_free(paths, nalloced * sizeof (char *)); 914 kmem_free(devis, nalloced * sizeof (dev_info_t *)); 915 nalloced = n + 128; 916 } 917 918 for (i = 0; i < npaths; i++) { 919 path = i_ddi_strdup(paths[i], KM_SLEEP); 920 paths[i] = path; 921 } 922 rw_exit(nvf_lock(dcfd_handle)); 923 924 if (ndevis == 0 && npaths == 0) { 925 DEVID_LOG_ERR(("no devid found", devid, NULL)); 926 kmem_free(paths, nalloced * sizeof (char *)); 927 kmem_free(devis, nalloced * sizeof (dev_info_t *)); 928 return (DDI_FAILURE); 929 } 930 931 ndevts_alloced = 128; 932 restart: 933 ndevts = 0; 934 devts = kmem_alloc(ndevts_alloced * sizeof (dev_t), KM_SLEEP); 935 for (i = 0; i < ndevis; i++) { 936 ASSERT(!DEVI_IS_ATTACHING(devis[i])); 937 ASSERT(!DEVI_IS_DETACHING(devis[i])); 938 e_devid_minor_to_devlist(devis[i], minor_name, 939 ndevts_alloced, &ndevts, devts); 940 if (ndevts > ndevts_alloced) { 941 kmem_free(devts, ndevts_alloced * sizeof (dev_t)); 942 ndevts_alloced += 128; 943 goto restart; 944 } 945 } 946 for (i = 0; i < npaths; i++) { 947 DEVID_LOG_LOOKUP((CE_CONT, "lookup %s\n", paths[i])); 948 devi = e_ddi_hold_devi_by_path(paths[i], 0); 949 if (devi == NULL) { 950 DEVID_LOG_STALE(("stale device reference", 951 devid, paths[i])); 952 continue; 953 } 954 /* 955 * Verify the newly attached device registered a matching devid 956 */ 957 if (i_ddi_devi_get_devid(DDI_DEV_T_ANY, devi, 958 &match_devid) != DDI_SUCCESS) { 959 DEVIDERR((CE_CONT, 960 "%s: no devid registered on attach\n", 961 paths[i])); 962 ddi_release_devi(devi); 963 continue; 964 } 965 966 if (ddi_devid_compare(devid, match_devid) != 0) { 967 DEVID_LOG_STALE(("new devid registered", 968 devid, paths[i])); 969 ddi_release_devi(devi); 970 ddi_devid_free(match_devid); 971 continue; 972 } 973 ddi_devid_free(match_devid); 974 975 e_devid_minor_to_devlist(devi, minor_name, 976 ndevts_alloced, &ndevts, devts); 977 ddi_release_devi(devi); 978 if (ndevts > ndevts_alloced) { 979 kmem_free(devts, 980 ndevts_alloced * sizeof (dev_t)); 981 ndevts_alloced += 128; 982 goto restart; 983 } 984 } 985 986 /* drop hold from e_devid_cache_devi_path_lists */ 987 for (i = 0; i < ndevis; i++) { 988 ndi_rele_devi(devis[i]); 989 } 990 for (i = 0; i < npaths; i++) { 991 kmem_free(paths[i], strlen(paths[i]) + 1); 992 } 993 kmem_free(paths, nalloced * sizeof (char *)); 994 kmem_free(devis, nalloced * sizeof (dev_info_t *)); 995 996 if (ndevts == 0) { 997 DEVID_LOG_ERR(("no devid found", devid, NULL)); 998 kmem_free(devts, ndevts_alloced * sizeof (dev_t)); 999 return (DDI_FAILURE); 1000 } 1001 1002 /* 1003 * Build the final list of sorted dev_t's with duplicates collapsed so 1004 * returned results are consistent. This prevents implementation 1005 * artifacts from causing unnecessary changes in SVM namespace. 1006 */ 1007 /* bubble sort */ 1008 for (i = 0; i < (ndevts - 1); i++) { 1009 for (j = 0; j < ((ndevts - 1) - i); j++) { 1010 if (devts[j + 1] < devts[j]) { 1011 tdevt = devts[j]; 1012 devts[j] = devts[j + 1]; 1013 devts[j + 1] = tdevt; 1014 } 1015 } 1016 } 1017 1018 /* determine number of unique values */ 1019 for (undevts = ndevts, i = 1; i < ndevts; i++) { 1020 if (devts[i - 1] == devts[i]) 1021 undevts--; 1022 } 1023 1024 /* allocate unique */ 1025 udevts = kmem_alloc(undevts * sizeof (dev_t), KM_SLEEP); 1026 1027 /* copy unique */ 1028 udevts[0] = devts[0]; 1029 for (i = 1, j = 1; i < ndevts; i++) { 1030 if (devts[i - 1] != devts[i]) 1031 udevts[j++] = devts[i]; 1032 } 1033 ASSERT(j == undevts); 1034 1035 kmem_free(devts, ndevts_alloced * sizeof (dev_t)); 1036 1037 *retndevts = undevts; 1038 *retdevts = udevts; 1039 1040 return (DDI_SUCCESS); 1041 } 1042 1043 void 1044 e_devid_cache_free_devt_list(int ndevts, dev_t *devt_list) 1045 { 1046 kmem_free(devt_list, ndevts * sizeof (dev_t *)); 1047 } 1048 1049 /* 1050 * If given a full path and NULL ua, search for a cache entry 1051 * whose path matches the full path. On a cache hit duplicate the 1052 * devid of the matched entry into the given devid (caller 1053 * must free); nodenamebuf is not touched for this usage. 1054 * 1055 * Given a path and a non-NULL unit address, search the cache for any entry 1056 * matching "<path>/%@<unit-address>" where '%' is a wildcard meaning 1057 * any node name. The path should not end a '/'. On a cache hit 1058 * duplicate the devid as before (caller must free) and copy into 1059 * the caller-provided nodenamebuf (if not NULL) the nodename of the 1060 * matched entry. 1061 * 1062 * We must not make use of nvp_dip since that may be NULL for cached 1063 * entries that are not present in the current tree. 1064 */ 1065 int 1066 e_devid_cache_path_to_devid(char *path, char *ua, 1067 char *nodenamebuf, ddi_devid_t *devidp) 1068 { 1069 size_t pathlen, ualen; 1070 int rv = DDI_FAILURE; 1071 nvp_devid_t *np; 1072 list_t *listp; 1073 char *cand; 1074 1075 if (path == NULL || *path == '\0' || (ua && *ua == '\0') || 1076 devidp == NULL) 1077 return (DDI_FAILURE); 1078 1079 *devidp = NULL; 1080 1081 pathlen = 0; 1082 ualen = 0; 1083 if (ua) { 1084 pathlen = strlen(path); 1085 ualen = strlen(ua); 1086 } 1087 1088 rw_enter(nvf_lock(dcfd_handle), RW_READER); 1089 1090 listp = nvf_list(dcfd_handle); 1091 for (np = list_head(listp); np; np = list_next(listp, np)) { 1092 size_t nodelen, candlen, n; 1093 ddi_devid_t devid_dup; 1094 char *uasep, *node; 1095 1096 if (np->nvp_devid == NULL) 1097 continue; 1098 1099 if (ddi_devid_valid(np->nvp_devid) != DDI_SUCCESS) { 1100 DEVIDERR((CE_CONT, 1101 "pathsearch: invalid devid %s\n", 1102 np->nvp_devpath)); 1103 continue; 1104 } 1105 1106 cand = np->nvp_devpath; /* candidate path */ 1107 1108 /* If a full pathname was provided the compare is easy */ 1109 if (ua == NULL) { 1110 if (strcmp(cand, path) == 0) 1111 goto match; 1112 else 1113 continue; 1114 } 1115 1116 /* 1117 * The compare for initial path plus ua and unknown nodename 1118 * is trickier. 1119 * 1120 * Does the initial path component match 'path'? 1121 */ 1122 if (strncmp(path, cand, pathlen) != 0) 1123 continue; 1124 1125 candlen = strlen(cand); 1126 1127 /* 1128 * The next character must be a '/' and there must be no 1129 * further '/' thereafter. Begin by checking that the 1130 * candidate is long enough to include at mininum a 1131 * "/<nodename>@<ua>" after the initial portion already 1132 * matched assuming a nodename length of 1. 1133 */ 1134 if (candlen < pathlen + 1 + 1 + 1 + ualen || 1135 cand[pathlen] != '/' || 1136 strchr(cand + pathlen + 1, '/') != NULL) 1137 continue; 1138 1139 node = cand + pathlen + 1; /* <node>@<ua> string */ 1140 1141 /* 1142 * Find the '@' before the unit address. Check for 1143 * unit address match. 1144 */ 1145 if ((uasep = strchr(node, '@')) == NULL) 1146 continue; 1147 1148 /* 1149 * Check we still have enough length and that ua matches 1150 */ 1151 nodelen = (uintptr_t)uasep - (uintptr_t)node; 1152 if (candlen < pathlen + 1 + nodelen + 1 + ualen || 1153 strncmp(ua, uasep + 1, ualen) != 0) 1154 continue; 1155 match: 1156 n = ddi_devid_sizeof(np->nvp_devid); 1157 devid_dup = kmem_alloc(n, KM_SLEEP); /* caller must free */ 1158 (void) bcopy(np->nvp_devid, devid_dup, n); 1159 *devidp = devid_dup; 1160 1161 if (ua && nodenamebuf) { 1162 (void) strncpy(nodenamebuf, node, nodelen); 1163 nodenamebuf[nodelen] = '\0'; 1164 } 1165 1166 rv = DDI_SUCCESS; 1167 break; 1168 } 1169 1170 rw_exit(nvf_lock(dcfd_handle)); 1171 1172 return (rv); 1173 } 1174 1175 #ifdef DEBUG 1176 static void 1177 devid_log(char *fmt, ddi_devid_t devid, char *path) 1178 { 1179 char *devidstr = ddi_devid_str_encode(devid, NULL); 1180 if (path) { 1181 cmn_err(CE_CONT, "%s: %s %s\n", fmt, path, devidstr); 1182 } else { 1183 cmn_err(CE_CONT, "%s: %s\n", fmt, devidstr); 1184 } 1185 ddi_devid_str_free(devidstr); 1186 } 1187 #endif /* DEBUG */ 1188