1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 /* 22 * Copyright 2007 Sun Microsystems, Inc. All rights reserved. 23 * Use is subject to license terms. 24 */ 25 26 #pragma ident "%Z%%M% %I% %E% SMI" 27 28 /* 29 * Pool import support functions. 30 * 31 * To import a pool, we rely on reading the configuration information from the 32 * ZFS label of each device. If we successfully read the label, then we 33 * organize the configuration information in the following hierarchy: 34 * 35 * pool guid -> toplevel vdev guid -> label txg 36 * 37 * Duplicate entries matching this same tuple will be discarded. Once we have 38 * examined every device, we pick the best label txg config for each toplevel 39 * vdev. We then arrange these toplevel vdevs into a complete pool config, and 40 * update any paths that have changed. Finally, we attempt to import the pool 41 * using our derived config, and record the results. 42 */ 43 44 #include <devid.h> 45 #include <dirent.h> 46 #include <errno.h> 47 #include <libintl.h> 48 #include <stdlib.h> 49 #include <string.h> 50 #include <sys/stat.h> 51 #include <unistd.h> 52 #include <fcntl.h> 53 54 #include <sys/vdev_impl.h> 55 56 #include "libzfs.h" 57 #include "libzfs_impl.h" 58 59 /* 60 * Intermediate structures used to gather configuration information. 61 */ 62 typedef struct config_entry { 63 uint64_t ce_txg; 64 nvlist_t *ce_config; 65 struct config_entry *ce_next; 66 } config_entry_t; 67 68 typedef struct vdev_entry { 69 uint64_t ve_guid; 70 config_entry_t *ve_configs; 71 struct vdev_entry *ve_next; 72 } vdev_entry_t; 73 74 typedef struct pool_entry { 75 uint64_t pe_guid; 76 vdev_entry_t *pe_vdevs; 77 struct pool_entry *pe_next; 78 } pool_entry_t; 79 80 typedef struct name_entry { 81 char *ne_name; 82 uint64_t ne_guid; 83 struct name_entry *ne_next; 84 } name_entry_t; 85 86 typedef struct pool_list { 87 pool_entry_t *pools; 88 name_entry_t *names; 89 } pool_list_t; 90 91 static char * 92 get_devid(const char *path) 93 { 94 int fd; 95 ddi_devid_t devid; 96 char *minor, *ret; 97 98 if ((fd = open(path, O_RDONLY)) < 0) 99 return (NULL); 100 101 minor = NULL; 102 ret = NULL; 103 if (devid_get(fd, &devid) == 0) { 104 if (devid_get_minor_name(fd, &minor) == 0) 105 ret = devid_str_encode(devid, minor); 106 if (minor != NULL) 107 devid_str_free(minor); 108 devid_free(devid); 109 } 110 (void) close(fd); 111 112 return (ret); 113 } 114 115 116 /* 117 * Go through and fix up any path and/or devid information for the given vdev 118 * configuration. 119 */ 120 static int 121 fix_paths(nvlist_t *nv, name_entry_t *names) 122 { 123 nvlist_t **child; 124 uint_t c, children; 125 uint64_t guid; 126 name_entry_t *ne, *best; 127 char *path, *devid; 128 int matched; 129 130 if (nvlist_lookup_nvlist_array(nv, ZPOOL_CONFIG_CHILDREN, 131 &child, &children) == 0) { 132 for (c = 0; c < children; c++) 133 if (fix_paths(child[c], names) != 0) 134 return (-1); 135 return (0); 136 } 137 138 /* 139 * This is a leaf (file or disk) vdev. In either case, go through 140 * the name list and see if we find a matching guid. If so, replace 141 * the path and see if we can calculate a new devid. 142 * 143 * There may be multiple names associated with a particular guid, in 144 * which case we have overlapping slices or multiple paths to the same 145 * disk. If this is the case, then we want to pick the path that is 146 * the most similar to the original, where "most similar" is the number 147 * of matching characters starting from the end of the path. This will 148 * preserve slice numbers even if the disks have been reorganized, and 149 * will also catch preferred disk names if multiple paths exist. 150 */ 151 verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_GUID, &guid) == 0); 152 if (nvlist_lookup_string(nv, ZPOOL_CONFIG_PATH, &path) != 0) 153 path = NULL; 154 155 matched = 0; 156 best = NULL; 157 for (ne = names; ne != NULL; ne = ne->ne_next) { 158 if (ne->ne_guid == guid) { 159 const char *src, *dst; 160 int count; 161 162 if (path == NULL) { 163 best = ne; 164 break; 165 } 166 167 src = ne->ne_name + strlen(ne->ne_name) - 1; 168 dst = path + strlen(path) - 1; 169 for (count = 0; src >= ne->ne_name && dst >= path; 170 src--, dst--, count++) 171 if (*src != *dst) 172 break; 173 174 /* 175 * At this point, 'count' is the number of characters 176 * matched from the end. 177 */ 178 if (count > matched || best == NULL) { 179 best = ne; 180 matched = count; 181 } 182 } 183 } 184 185 if (best == NULL) 186 return (0); 187 188 if (nvlist_add_string(nv, ZPOOL_CONFIG_PATH, best->ne_name) != 0) 189 return (-1); 190 191 if ((devid = get_devid(best->ne_name)) == NULL) { 192 (void) nvlist_remove_all(nv, ZPOOL_CONFIG_DEVID); 193 } else { 194 if (nvlist_add_string(nv, ZPOOL_CONFIG_DEVID, devid) != 0) 195 return (-1); 196 devid_str_free(devid); 197 } 198 199 return (0); 200 } 201 202 /* 203 * Add the given configuration to the list of known devices. 204 */ 205 static int 206 add_config(libzfs_handle_t *hdl, pool_list_t *pl, const char *path, 207 nvlist_t *config) 208 { 209 uint64_t pool_guid, vdev_guid, top_guid, txg, state; 210 pool_entry_t *pe; 211 vdev_entry_t *ve; 212 config_entry_t *ce; 213 name_entry_t *ne; 214 215 /* 216 * If this is a hot spare not currently in use or level 2 cache 217 * device, add it to the list of names to translate, but don't do 218 * anything else. 219 */ 220 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE, 221 &state) == 0 && 222 (state == POOL_STATE_SPARE || state == POOL_STATE_L2CACHE) && 223 nvlist_lookup_uint64(config, ZPOOL_CONFIG_GUID, &vdev_guid) == 0) { 224 if ((ne = zfs_alloc(hdl, sizeof (name_entry_t))) == NULL) 225 return (-1); 226 227 if ((ne->ne_name = zfs_strdup(hdl, path)) == NULL) { 228 free(ne); 229 return (-1); 230 } 231 ne->ne_guid = vdev_guid; 232 ne->ne_next = pl->names; 233 pl->names = ne; 234 return (0); 235 } 236 237 /* 238 * If we have a valid config but cannot read any of these fields, then 239 * it means we have a half-initialized label. In vdev_label_init() 240 * we write a label with txg == 0 so that we can identify the device 241 * in case the user refers to the same disk later on. If we fail to 242 * create the pool, we'll be left with a label in this state 243 * which should not be considered part of a valid pool. 244 */ 245 if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_GUID, 246 &pool_guid) != 0 || 247 nvlist_lookup_uint64(config, ZPOOL_CONFIG_GUID, 248 &vdev_guid) != 0 || 249 nvlist_lookup_uint64(config, ZPOOL_CONFIG_TOP_GUID, 250 &top_guid) != 0 || 251 nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_TXG, 252 &txg) != 0 || txg == 0) { 253 nvlist_free(config); 254 return (0); 255 } 256 257 /* 258 * First, see if we know about this pool. If not, then add it to the 259 * list of known pools. 260 */ 261 for (pe = pl->pools; pe != NULL; pe = pe->pe_next) { 262 if (pe->pe_guid == pool_guid) 263 break; 264 } 265 266 if (pe == NULL) { 267 if ((pe = zfs_alloc(hdl, sizeof (pool_entry_t))) == NULL) { 268 nvlist_free(config); 269 return (-1); 270 } 271 pe->pe_guid = pool_guid; 272 pe->pe_next = pl->pools; 273 pl->pools = pe; 274 } 275 276 /* 277 * Second, see if we know about this toplevel vdev. Add it if its 278 * missing. 279 */ 280 for (ve = pe->pe_vdevs; ve != NULL; ve = ve->ve_next) { 281 if (ve->ve_guid == top_guid) 282 break; 283 } 284 285 if (ve == NULL) { 286 if ((ve = zfs_alloc(hdl, sizeof (vdev_entry_t))) == NULL) { 287 nvlist_free(config); 288 return (-1); 289 } 290 ve->ve_guid = top_guid; 291 ve->ve_next = pe->pe_vdevs; 292 pe->pe_vdevs = ve; 293 } 294 295 /* 296 * Third, see if we have a config with a matching transaction group. If 297 * so, then we do nothing. Otherwise, add it to the list of known 298 * configs. 299 */ 300 for (ce = ve->ve_configs; ce != NULL; ce = ce->ce_next) { 301 if (ce->ce_txg == txg) 302 break; 303 } 304 305 if (ce == NULL) { 306 if ((ce = zfs_alloc(hdl, sizeof (config_entry_t))) == NULL) { 307 nvlist_free(config); 308 return (-1); 309 } 310 ce->ce_txg = txg; 311 ce->ce_config = config; 312 ce->ce_next = ve->ve_configs; 313 ve->ve_configs = ce; 314 } else { 315 nvlist_free(config); 316 } 317 318 /* 319 * At this point we've successfully added our config to the list of 320 * known configs. The last thing to do is add the vdev guid -> path 321 * mappings so that we can fix up the configuration as necessary before 322 * doing the import. 323 */ 324 if ((ne = zfs_alloc(hdl, sizeof (name_entry_t))) == NULL) 325 return (-1); 326 327 if ((ne->ne_name = zfs_strdup(hdl, path)) == NULL) { 328 free(ne); 329 return (-1); 330 } 331 332 ne->ne_guid = vdev_guid; 333 ne->ne_next = pl->names; 334 pl->names = ne; 335 336 return (0); 337 } 338 339 /* 340 * Returns true if the named pool matches the given GUID. 341 */ 342 static int 343 pool_active(libzfs_handle_t *hdl, const char *name, uint64_t guid, 344 boolean_t *isactive) 345 { 346 zpool_handle_t *zhp; 347 uint64_t theguid; 348 349 if (zpool_open_silent(hdl, name, &zhp) != 0) 350 return (-1); 351 352 if (zhp == NULL) { 353 *isactive = B_FALSE; 354 return (0); 355 } 356 357 verify(nvlist_lookup_uint64(zhp->zpool_config, ZPOOL_CONFIG_POOL_GUID, 358 &theguid) == 0); 359 360 zpool_close(zhp); 361 362 *isactive = (theguid == guid); 363 return (0); 364 } 365 366 static nvlist_t * 367 refresh_config(libzfs_handle_t *hdl, nvlist_t *config) 368 { 369 nvlist_t *nvl; 370 zfs_cmd_t zc = { 0 }; 371 int err; 372 373 if (zcmd_write_conf_nvlist(hdl, &zc, config) != 0) 374 return (NULL); 375 376 if (zcmd_alloc_dst_nvlist(hdl, &zc, 377 zc.zc_nvlist_conf_size * 2) != 0) { 378 zcmd_free_nvlists(&zc); 379 return (NULL); 380 } 381 382 while ((err = ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_TRYIMPORT, 383 &zc)) != 0 && errno == ENOMEM) { 384 if (zcmd_expand_dst_nvlist(hdl, &zc) != 0) { 385 zcmd_free_nvlists(&zc); 386 return (NULL); 387 } 388 } 389 390 if (err) { 391 (void) zpool_standard_error(hdl, errno, 392 dgettext(TEXT_DOMAIN, "cannot discover pools")); 393 zcmd_free_nvlists(&zc); 394 return (NULL); 395 } 396 397 if (zcmd_read_dst_nvlist(hdl, &zc, &nvl) != 0) { 398 zcmd_free_nvlists(&zc); 399 return (NULL); 400 } 401 402 zcmd_free_nvlists(&zc); 403 return (nvl); 404 } 405 406 /* 407 * Convert our list of pools into the definitive set of configurations. We 408 * start by picking the best config for each toplevel vdev. Once that's done, 409 * we assemble the toplevel vdevs into a full config for the pool. We make a 410 * pass to fix up any incorrect paths, and then add it to the main list to 411 * return to the user. 412 */ 413 static nvlist_t * 414 get_configs(libzfs_handle_t *hdl, pool_list_t *pl) 415 { 416 pool_entry_t *pe; 417 vdev_entry_t *ve; 418 config_entry_t *ce; 419 nvlist_t *ret = NULL, *config = NULL, *tmp, *nvtop, *nvroot; 420 nvlist_t **spares, **l2cache; 421 uint_t i, nspares, nl2cache; 422 boolean_t config_seen; 423 uint64_t best_txg; 424 char *name, *hostname; 425 uint64_t version, guid; 426 uint_t children = 0; 427 nvlist_t **child = NULL; 428 uint_t c; 429 boolean_t isactive; 430 uint64_t hostid; 431 nvlist_t *nvl; 432 433 if (nvlist_alloc(&ret, 0, 0) != 0) 434 goto nomem; 435 436 for (pe = pl->pools; pe != NULL; pe = pe->pe_next) { 437 uint64_t id; 438 439 if (nvlist_alloc(&config, NV_UNIQUE_NAME, 0) != 0) 440 goto nomem; 441 config_seen = B_FALSE; 442 443 /* 444 * Iterate over all toplevel vdevs. Grab the pool configuration 445 * from the first one we find, and then go through the rest and 446 * add them as necessary to the 'vdevs' member of the config. 447 */ 448 for (ve = pe->pe_vdevs; ve != NULL; ve = ve->ve_next) { 449 450 /* 451 * Determine the best configuration for this vdev by 452 * selecting the config with the latest transaction 453 * group. 454 */ 455 best_txg = 0; 456 for (ce = ve->ve_configs; ce != NULL; 457 ce = ce->ce_next) { 458 459 if (ce->ce_txg > best_txg) { 460 tmp = ce->ce_config; 461 best_txg = ce->ce_txg; 462 } 463 } 464 465 if (!config_seen) { 466 /* 467 * Copy the relevant pieces of data to the pool 468 * configuration: 469 * 470 * version 471 * pool guid 472 * name 473 * pool state 474 * hostid (if available) 475 * hostname (if available) 476 */ 477 uint64_t state; 478 479 verify(nvlist_lookup_uint64(tmp, 480 ZPOOL_CONFIG_VERSION, &version) == 0); 481 if (nvlist_add_uint64(config, 482 ZPOOL_CONFIG_VERSION, version) != 0) 483 goto nomem; 484 verify(nvlist_lookup_uint64(tmp, 485 ZPOOL_CONFIG_POOL_GUID, &guid) == 0); 486 if (nvlist_add_uint64(config, 487 ZPOOL_CONFIG_POOL_GUID, guid) != 0) 488 goto nomem; 489 verify(nvlist_lookup_string(tmp, 490 ZPOOL_CONFIG_POOL_NAME, &name) == 0); 491 if (nvlist_add_string(config, 492 ZPOOL_CONFIG_POOL_NAME, name) != 0) 493 goto nomem; 494 verify(nvlist_lookup_uint64(tmp, 495 ZPOOL_CONFIG_POOL_STATE, &state) == 0); 496 if (nvlist_add_uint64(config, 497 ZPOOL_CONFIG_POOL_STATE, state) != 0) 498 goto nomem; 499 hostid = 0; 500 if (nvlist_lookup_uint64(tmp, 501 ZPOOL_CONFIG_HOSTID, &hostid) == 0) { 502 if (nvlist_add_uint64(config, 503 ZPOOL_CONFIG_HOSTID, hostid) != 0) 504 goto nomem; 505 verify(nvlist_lookup_string(tmp, 506 ZPOOL_CONFIG_HOSTNAME, 507 &hostname) == 0); 508 if (nvlist_add_string(config, 509 ZPOOL_CONFIG_HOSTNAME, 510 hostname) != 0) 511 goto nomem; 512 } 513 514 config_seen = B_TRUE; 515 } 516 517 /* 518 * Add this top-level vdev to the child array. 519 */ 520 verify(nvlist_lookup_nvlist(tmp, 521 ZPOOL_CONFIG_VDEV_TREE, &nvtop) == 0); 522 verify(nvlist_lookup_uint64(nvtop, ZPOOL_CONFIG_ID, 523 &id) == 0); 524 if (id >= children) { 525 nvlist_t **newchild; 526 527 newchild = zfs_alloc(hdl, (id + 1) * 528 sizeof (nvlist_t *)); 529 if (newchild == NULL) 530 goto nomem; 531 532 for (c = 0; c < children; c++) 533 newchild[c] = child[c]; 534 535 free(child); 536 child = newchild; 537 children = id + 1; 538 } 539 if (nvlist_dup(nvtop, &child[id], 0) != 0) 540 goto nomem; 541 542 } 543 544 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_GUID, 545 &guid) == 0); 546 547 /* 548 * Look for any missing top-level vdevs. If this is the case, 549 * create a faked up 'missing' vdev as a placeholder. We cannot 550 * simply compress the child array, because the kernel performs 551 * certain checks to make sure the vdev IDs match their location 552 * in the configuration. 553 */ 554 for (c = 0; c < children; c++) 555 if (child[c] == NULL) { 556 nvlist_t *missing; 557 if (nvlist_alloc(&missing, NV_UNIQUE_NAME, 558 0) != 0) 559 goto nomem; 560 if (nvlist_add_string(missing, 561 ZPOOL_CONFIG_TYPE, 562 VDEV_TYPE_MISSING) != 0 || 563 nvlist_add_uint64(missing, 564 ZPOOL_CONFIG_ID, c) != 0 || 565 nvlist_add_uint64(missing, 566 ZPOOL_CONFIG_GUID, 0ULL) != 0) { 567 nvlist_free(missing); 568 goto nomem; 569 } 570 child[c] = missing; 571 } 572 573 /* 574 * Put all of this pool's top-level vdevs into a root vdev. 575 */ 576 if (nvlist_alloc(&nvroot, NV_UNIQUE_NAME, 0) != 0) 577 goto nomem; 578 if (nvlist_add_string(nvroot, ZPOOL_CONFIG_TYPE, 579 VDEV_TYPE_ROOT) != 0 || 580 nvlist_add_uint64(nvroot, ZPOOL_CONFIG_ID, 0ULL) != 0 || 581 nvlist_add_uint64(nvroot, ZPOOL_CONFIG_GUID, guid) != 0 || 582 nvlist_add_nvlist_array(nvroot, ZPOOL_CONFIG_CHILDREN, 583 child, children) != 0) { 584 nvlist_free(nvroot); 585 goto nomem; 586 } 587 588 for (c = 0; c < children; c++) 589 nvlist_free(child[c]); 590 free(child); 591 children = 0; 592 child = NULL; 593 594 /* 595 * Go through and fix up any paths and/or devids based on our 596 * known list of vdev GUID -> path mappings. 597 */ 598 if (fix_paths(nvroot, pl->names) != 0) { 599 nvlist_free(nvroot); 600 goto nomem; 601 } 602 603 /* 604 * Add the root vdev to this pool's configuration. 605 */ 606 if (nvlist_add_nvlist(config, ZPOOL_CONFIG_VDEV_TREE, 607 nvroot) != 0) { 608 nvlist_free(nvroot); 609 goto nomem; 610 } 611 nvlist_free(nvroot); 612 613 /* 614 * Determine if this pool is currently active, in which case we 615 * can't actually import it. 616 */ 617 verify(nvlist_lookup_string(config, ZPOOL_CONFIG_POOL_NAME, 618 &name) == 0); 619 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_GUID, 620 &guid) == 0); 621 622 if (pool_active(hdl, name, guid, &isactive) != 0) 623 goto error; 624 625 if (isactive) { 626 nvlist_free(config); 627 config = NULL; 628 continue; 629 } 630 631 if ((nvl = refresh_config(hdl, config)) == NULL) 632 goto error; 633 634 nvlist_free(config); 635 config = nvl; 636 637 /* 638 * Go through and update the paths for spares, now that we have 639 * them. 640 */ 641 verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_VDEV_TREE, 642 &nvroot) == 0); 643 if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_SPARES, 644 &spares, &nspares) == 0) { 645 for (i = 0; i < nspares; i++) { 646 if (fix_paths(spares[i], pl->names) != 0) 647 goto nomem; 648 } 649 } 650 651 /* 652 * Update the paths for l2cache devices. 653 */ 654 if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_L2CACHE, 655 &l2cache, &nl2cache) == 0) { 656 for (i = 0; i < nl2cache; i++) { 657 if (fix_paths(l2cache[i], pl->names) != 0) 658 goto nomem; 659 } 660 } 661 662 /* 663 * Restore the original information read from the actual label. 664 */ 665 (void) nvlist_remove(config, ZPOOL_CONFIG_HOSTID, 666 DATA_TYPE_UINT64); 667 (void) nvlist_remove(config, ZPOOL_CONFIG_HOSTNAME, 668 DATA_TYPE_STRING); 669 if (hostid != 0) { 670 verify(nvlist_add_uint64(config, ZPOOL_CONFIG_HOSTID, 671 hostid) == 0); 672 verify(nvlist_add_string(config, ZPOOL_CONFIG_HOSTNAME, 673 hostname) == 0); 674 } 675 676 /* 677 * Add this pool to the list of configs. 678 */ 679 verify(nvlist_lookup_string(config, ZPOOL_CONFIG_POOL_NAME, 680 &name) == 0); 681 if (nvlist_add_nvlist(ret, name, config) != 0) 682 goto nomem; 683 684 nvlist_free(config); 685 config = NULL; 686 } 687 688 return (ret); 689 690 nomem: 691 (void) no_memory(hdl); 692 error: 693 nvlist_free(config); 694 nvlist_free(ret); 695 for (c = 0; c < children; c++) 696 nvlist_free(child[c]); 697 free(child); 698 699 return (NULL); 700 } 701 702 /* 703 * Return the offset of the given label. 704 */ 705 static uint64_t 706 label_offset(uint64_t size, int l) 707 { 708 ASSERT(P2PHASE_TYPED(size, sizeof (vdev_label_t), uint64_t) == 0); 709 return (l * sizeof (vdev_label_t) + (l < VDEV_LABELS / 2 ? 710 0 : size - VDEV_LABELS * sizeof (vdev_label_t))); 711 } 712 713 /* 714 * Given a file descriptor, read the label information and return an nvlist 715 * describing the configuration, if there is one. 716 */ 717 int 718 zpool_read_label(int fd, nvlist_t **config) 719 { 720 struct stat64 statbuf; 721 int l; 722 vdev_label_t *label; 723 uint64_t state, txg, size; 724 725 *config = NULL; 726 727 if (fstat64(fd, &statbuf) == -1) 728 return (0); 729 size = P2ALIGN_TYPED(statbuf.st_size, sizeof (vdev_label_t), uint64_t); 730 731 if ((label = malloc(sizeof (vdev_label_t))) == NULL) 732 return (-1); 733 734 for (l = 0; l < VDEV_LABELS; l++) { 735 if (pread(fd, label, sizeof (vdev_label_t), 736 label_offset(size, l)) != sizeof (vdev_label_t)) 737 continue; 738 739 if (nvlist_unpack(label->vl_vdev_phys.vp_nvlist, 740 sizeof (label->vl_vdev_phys.vp_nvlist), config, 0) != 0) 741 continue; 742 743 if (nvlist_lookup_uint64(*config, ZPOOL_CONFIG_POOL_STATE, 744 &state) != 0 || state > POOL_STATE_L2CACHE) { 745 nvlist_free(*config); 746 continue; 747 } 748 749 if (state != POOL_STATE_SPARE && state != POOL_STATE_L2CACHE && 750 (nvlist_lookup_uint64(*config, ZPOOL_CONFIG_POOL_TXG, 751 &txg) != 0 || txg == 0)) { 752 nvlist_free(*config); 753 continue; 754 } 755 756 free(label); 757 return (0); 758 } 759 760 free(label); 761 *config = NULL; 762 return (0); 763 } 764 765 /* 766 * Given a list of directories to search, find all pools stored on disk. This 767 * includes partial pools which are not available to import. If no args are 768 * given (argc is 0), then the default directory (/dev/dsk) is searched. 769 */ 770 nvlist_t * 771 zpool_find_import(libzfs_handle_t *hdl, int argc, char **argv) 772 { 773 int i; 774 DIR *dirp = NULL; 775 struct dirent64 *dp; 776 char path[MAXPATHLEN]; 777 struct stat64 statbuf; 778 nvlist_t *ret = NULL, *config; 779 static char *default_dir = "/dev/dsk"; 780 int fd; 781 pool_list_t pools = { 0 }; 782 pool_entry_t *pe, *penext; 783 vdev_entry_t *ve, *venext; 784 config_entry_t *ce, *cenext; 785 name_entry_t *ne, *nenext; 786 787 788 if (argc == 0) { 789 argc = 1; 790 argv = &default_dir; 791 } 792 793 /* 794 * Go through and read the label configuration information from every 795 * possible device, organizing the information according to pool GUID 796 * and toplevel GUID. 797 */ 798 for (i = 0; i < argc; i++) { 799 if (argv[i][0] != '/') { 800 (void) zfs_error_fmt(hdl, EZFS_BADPATH, 801 dgettext(TEXT_DOMAIN, "cannot open '%s'"), 802 argv[i]); 803 goto error; 804 } 805 806 if ((dirp = opendir(argv[i])) == NULL) { 807 zfs_error_aux(hdl, strerror(errno)); 808 (void) zfs_error_fmt(hdl, EZFS_BADPATH, 809 dgettext(TEXT_DOMAIN, "cannot open '%s'"), 810 argv[i]); 811 goto error; 812 } 813 814 /* 815 * This is not MT-safe, but we have no MT consumers of libzfs 816 */ 817 while ((dp = readdir64(dirp)) != NULL) { 818 819 (void) snprintf(path, sizeof (path), "%s/%s", 820 argv[i], dp->d_name); 821 822 if (stat64(path, &statbuf) != 0) 823 continue; 824 825 /* 826 * Ignore directories (which includes "." and ".."). 827 */ 828 if (S_ISDIR(statbuf.st_mode)) 829 continue; 830 831 /* 832 * Ignore special (non-character or non-block) files. 833 */ 834 if (!S_ISREG(statbuf.st_mode) && 835 !S_ISBLK(statbuf.st_mode)) 836 continue; 837 838 if ((fd = open64(path, O_RDONLY)) < 0) 839 continue; 840 841 if ((zpool_read_label(fd, &config)) != 0) { 842 (void) close(fd); 843 (void) no_memory(hdl); 844 goto error; 845 } 846 847 (void) close(fd); 848 849 if (config != NULL) 850 if (add_config(hdl, &pools, path, config) != 0) 851 goto error; 852 } 853 854 (void) closedir(dirp); 855 dirp = NULL; 856 } 857 858 ret = get_configs(hdl, &pools); 859 860 error: 861 for (pe = pools.pools; pe != NULL; pe = penext) { 862 penext = pe->pe_next; 863 for (ve = pe->pe_vdevs; ve != NULL; ve = venext) { 864 venext = ve->ve_next; 865 for (ce = ve->ve_configs; ce != NULL; ce = cenext) { 866 cenext = ce->ce_next; 867 if (ce->ce_config) 868 nvlist_free(ce->ce_config); 869 free(ce); 870 } 871 free(ve); 872 } 873 free(pe); 874 } 875 876 for (ne = pools.names; ne != NULL; ne = nenext) { 877 nenext = ne->ne_next; 878 if (ne->ne_name) 879 free(ne->ne_name); 880 free(ne); 881 } 882 883 if (dirp) 884 (void) closedir(dirp); 885 886 return (ret); 887 } 888 889 /* 890 * Given a cache file, return the contents as a list of importable pools. 891 */ 892 nvlist_t * 893 zpool_find_import_cached(libzfs_handle_t *hdl, const char *cachefile) 894 { 895 char *buf; 896 int fd; 897 struct stat64 statbuf; 898 nvlist_t *raw, *src, *dst; 899 nvlist_t *pools; 900 nvpair_t *elem; 901 char *name; 902 uint64_t guid; 903 boolean_t active; 904 905 if ((fd = open(cachefile, O_RDONLY)) < 0) { 906 zfs_error_aux(hdl, "%s", strerror(errno)); 907 (void) zfs_error(hdl, EZFS_BADCACHE, 908 dgettext(TEXT_DOMAIN, "failed to open cache file")); 909 return (NULL); 910 } 911 912 if (fstat64(fd, &statbuf) != 0) { 913 zfs_error_aux(hdl, "%s", strerror(errno)); 914 (void) close(fd); 915 (void) zfs_error(hdl, EZFS_BADCACHE, 916 dgettext(TEXT_DOMAIN, "failed to get size of cache file")); 917 return (NULL); 918 } 919 920 if ((buf = zfs_alloc(hdl, statbuf.st_size)) == NULL) { 921 (void) close(fd); 922 return (NULL); 923 } 924 925 if (read(fd, buf, statbuf.st_size) != statbuf.st_size) { 926 (void) close(fd); 927 free(buf); 928 (void) zfs_error(hdl, EZFS_BADCACHE, 929 dgettext(TEXT_DOMAIN, 930 "failed to read cache file contents")); 931 return (NULL); 932 } 933 934 (void) close(fd); 935 936 if (nvlist_unpack(buf, statbuf.st_size, &raw, 0) != 0) { 937 free(buf); 938 (void) zfs_error(hdl, EZFS_BADCACHE, 939 dgettext(TEXT_DOMAIN, 940 "invalid or corrupt cache file contents")); 941 return (NULL); 942 } 943 944 free(buf); 945 946 /* 947 * Go through and get the current state of the pools and refresh their 948 * state. 949 */ 950 if (nvlist_alloc(&pools, 0, 0) != 0) { 951 (void) no_memory(hdl); 952 nvlist_free(raw); 953 return (NULL); 954 } 955 956 elem = NULL; 957 while ((elem = nvlist_next_nvpair(raw, elem)) != NULL) { 958 verify(nvpair_value_nvlist(elem, &src) == 0); 959 960 verify(nvlist_lookup_string(src, ZPOOL_CONFIG_POOL_NAME, 961 &name) == 0); 962 verify(nvlist_lookup_uint64(src, ZPOOL_CONFIG_POOL_GUID, 963 &guid) == 0); 964 965 if (pool_active(hdl, name, guid, &active) != 0) { 966 nvlist_free(raw); 967 nvlist_free(pools); 968 return (NULL); 969 } 970 971 if (active) 972 continue; 973 974 if ((dst = refresh_config(hdl, src)) == NULL) { 975 nvlist_free(raw); 976 nvlist_free(pools); 977 return (NULL); 978 } 979 980 if (nvlist_add_nvlist(pools, nvpair_name(elem), dst) != 0) { 981 (void) no_memory(hdl); 982 nvlist_free(dst); 983 nvlist_free(raw); 984 nvlist_free(pools); 985 return (NULL); 986 } 987 988 nvlist_free(dst); 989 } 990 991 nvlist_free(raw); 992 return (pools); 993 } 994 995 996 boolean_t 997 find_guid(nvlist_t *nv, uint64_t guid) 998 { 999 uint64_t tmp; 1000 nvlist_t **child; 1001 uint_t c, children; 1002 1003 verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_GUID, &tmp) == 0); 1004 if (tmp == guid) 1005 return (B_TRUE); 1006 1007 if (nvlist_lookup_nvlist_array(nv, ZPOOL_CONFIG_CHILDREN, 1008 &child, &children) == 0) { 1009 for (c = 0; c < children; c++) 1010 if (find_guid(child[c], guid)) 1011 return (B_TRUE); 1012 } 1013 1014 return (B_FALSE); 1015 } 1016 1017 typedef struct aux_cbdata { 1018 const char *cb_type; 1019 uint64_t cb_guid; 1020 zpool_handle_t *cb_zhp; 1021 } aux_cbdata_t; 1022 1023 static int 1024 find_aux(zpool_handle_t *zhp, void *data) 1025 { 1026 aux_cbdata_t *cbp = data; 1027 nvlist_t **list; 1028 uint_t i, count; 1029 uint64_t guid; 1030 nvlist_t *nvroot; 1031 1032 verify(nvlist_lookup_nvlist(zhp->zpool_config, ZPOOL_CONFIG_VDEV_TREE, 1033 &nvroot) == 0); 1034 1035 if (nvlist_lookup_nvlist_array(nvroot, cbp->cb_type, 1036 &list, &count) == 0) { 1037 for (i = 0; i < count; i++) { 1038 verify(nvlist_lookup_uint64(list[i], 1039 ZPOOL_CONFIG_GUID, &guid) == 0); 1040 if (guid == cbp->cb_guid) { 1041 cbp->cb_zhp = zhp; 1042 return (1); 1043 } 1044 } 1045 } 1046 1047 zpool_close(zhp); 1048 return (0); 1049 } 1050 1051 /* 1052 * Determines if the pool is in use. If so, it returns true and the state of 1053 * the pool as well as the name of the pool. Both strings are allocated and 1054 * must be freed by the caller. 1055 */ 1056 int 1057 zpool_in_use(libzfs_handle_t *hdl, int fd, pool_state_t *state, char **namestr, 1058 boolean_t *inuse) 1059 { 1060 nvlist_t *config; 1061 char *name; 1062 boolean_t ret; 1063 uint64_t guid, vdev_guid; 1064 zpool_handle_t *zhp; 1065 nvlist_t *pool_config; 1066 uint64_t stateval, isspare; 1067 aux_cbdata_t cb = { 0 }; 1068 boolean_t isactive; 1069 1070 *inuse = B_FALSE; 1071 1072 if (zpool_read_label(fd, &config) != 0) { 1073 (void) no_memory(hdl); 1074 return (-1); 1075 } 1076 1077 if (config == NULL) 1078 return (0); 1079 1080 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE, 1081 &stateval) == 0); 1082 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_GUID, 1083 &vdev_guid) == 0); 1084 1085 if (stateval != POOL_STATE_SPARE && stateval != POOL_STATE_L2CACHE) { 1086 verify(nvlist_lookup_string(config, ZPOOL_CONFIG_POOL_NAME, 1087 &name) == 0); 1088 verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_GUID, 1089 &guid) == 0); 1090 } 1091 1092 switch (stateval) { 1093 case POOL_STATE_EXPORTED: 1094 ret = B_TRUE; 1095 break; 1096 1097 case POOL_STATE_ACTIVE: 1098 /* 1099 * For an active pool, we have to determine if it's really part 1100 * of a currently active pool (in which case the pool will exist 1101 * and the guid will be the same), or whether it's part of an 1102 * active pool that was disconnected without being explicitly 1103 * exported. 1104 */ 1105 if (pool_active(hdl, name, guid, &isactive) != 0) { 1106 nvlist_free(config); 1107 return (-1); 1108 } 1109 1110 if (isactive) { 1111 /* 1112 * Because the device may have been removed while 1113 * offlined, we only report it as active if the vdev is 1114 * still present in the config. Otherwise, pretend like 1115 * it's not in use. 1116 */ 1117 if ((zhp = zpool_open_canfail(hdl, name)) != NULL && 1118 (pool_config = zpool_get_config(zhp, NULL)) 1119 != NULL) { 1120 nvlist_t *nvroot; 1121 1122 verify(nvlist_lookup_nvlist(pool_config, 1123 ZPOOL_CONFIG_VDEV_TREE, &nvroot) == 0); 1124 ret = find_guid(nvroot, vdev_guid); 1125 } else { 1126 ret = B_FALSE; 1127 } 1128 1129 /* 1130 * If this is an active spare within another pool, we 1131 * treat it like an unused hot spare. This allows the 1132 * user to create a pool with a hot spare that currently 1133 * in use within another pool. Since we return B_TRUE, 1134 * libdiskmgt will continue to prevent generic consumers 1135 * from using the device. 1136 */ 1137 if (ret && nvlist_lookup_uint64(config, 1138 ZPOOL_CONFIG_IS_SPARE, &isspare) == 0 && isspare) 1139 stateval = POOL_STATE_SPARE; 1140 1141 if (zhp != NULL) 1142 zpool_close(zhp); 1143 } else { 1144 stateval = POOL_STATE_POTENTIALLY_ACTIVE; 1145 ret = B_TRUE; 1146 } 1147 break; 1148 1149 case POOL_STATE_SPARE: 1150 /* 1151 * For a hot spare, it can be either definitively in use, or 1152 * potentially active. To determine if it's in use, we iterate 1153 * over all pools in the system and search for one with a spare 1154 * with a matching guid. 1155 * 1156 * Due to the shared nature of spares, we don't actually report 1157 * the potentially active case as in use. This means the user 1158 * can freely create pools on the hot spares of exported pools, 1159 * but to do otherwise makes the resulting code complicated, and 1160 * we end up having to deal with this case anyway. 1161 */ 1162 cb.cb_zhp = NULL; 1163 cb.cb_guid = vdev_guid; 1164 cb.cb_type = ZPOOL_CONFIG_SPARES; 1165 if (zpool_iter(hdl, find_aux, &cb) == 1) { 1166 name = (char *)zpool_get_name(cb.cb_zhp); 1167 ret = TRUE; 1168 } else { 1169 ret = FALSE; 1170 } 1171 break; 1172 1173 case POOL_STATE_L2CACHE: 1174 1175 /* 1176 * Check if any pool is currently using this l2cache device. 1177 */ 1178 cb.cb_zhp = NULL; 1179 cb.cb_guid = vdev_guid; 1180 cb.cb_type = ZPOOL_CONFIG_L2CACHE; 1181 if (zpool_iter(hdl, find_aux, &cb) == 1) { 1182 name = (char *)zpool_get_name(cb.cb_zhp); 1183 ret = TRUE; 1184 } else { 1185 ret = FALSE; 1186 } 1187 break; 1188 1189 default: 1190 ret = B_FALSE; 1191 } 1192 1193 1194 if (ret) { 1195 if ((*namestr = zfs_strdup(hdl, name)) == NULL) { 1196 if (cb.cb_zhp) 1197 zpool_close(cb.cb_zhp); 1198 nvlist_free(config); 1199 return (-1); 1200 } 1201 *state = (pool_state_t)stateval; 1202 } 1203 1204 if (cb.cb_zhp) 1205 zpool_close(cb.cb_zhp); 1206 1207 nvlist_free(config); 1208 *inuse = ret; 1209 return (0); 1210 } 1211