xref: /titanic_51/usr/src/lib/libzfs/common/libzfs_pool.c (revision c81d47afd05baeb768e2f032636019b717899efd)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
24  * Use is subject to license terms.
25  */
26 
27 #pragma ident	"%Z%%M%	%I%	%E% SMI"
28 
29 #include <alloca.h>
30 #include <assert.h>
31 #include <ctype.h>
32 #include <errno.h>
33 #include <devid.h>
34 #include <dirent.h>
35 #include <fcntl.h>
36 #include <libintl.h>
37 #include <stdio.h>
38 #include <stdlib.h>
39 #include <strings.h>
40 #include <unistd.h>
41 #include <sys/zfs_ioctl.h>
42 #include <sys/zio.h>
43 #include <strings.h>
44 
45 #include "zfs_namecheck.h"
46 #include "libzfs_impl.h"
47 
48 /*
49  * Validate the given pool name, optionally putting an extended error message in
50  * 'buf'.
51  */
52 static boolean_t
53 zpool_name_valid(libzfs_handle_t *hdl, boolean_t isopen, const char *pool)
54 {
55 	namecheck_err_t why;
56 	char what;
57 	int ret;
58 
59 	ret = pool_namecheck(pool, &why, &what);
60 
61 	/*
62 	 * The rules for reserved pool names were extended at a later point.
63 	 * But we need to support users with existing pools that may now be
64 	 * invalid.  So we only check for this expanded set of names during a
65 	 * create (or import), and only in userland.
66 	 */
67 	if (ret == 0 && !isopen &&
68 	    (strncmp(pool, "mirror", 6) == 0 ||
69 	    strncmp(pool, "raidz", 5) == 0 ||
70 	    strncmp(pool, "spare", 5) == 0)) {
71 		zfs_error_aux(hdl,
72 		    dgettext(TEXT_DOMAIN, "name is reserved"));
73 		return (B_FALSE);
74 	}
75 
76 
77 	if (ret != 0) {
78 		if (hdl != NULL) {
79 			switch (why) {
80 			case NAME_ERR_TOOLONG:
81 				zfs_error_aux(hdl,
82 				    dgettext(TEXT_DOMAIN, "name is too long"));
83 				break;
84 
85 			case NAME_ERR_INVALCHAR:
86 				zfs_error_aux(hdl,
87 				    dgettext(TEXT_DOMAIN, "invalid character "
88 				    "'%c' in pool name"), what);
89 				break;
90 
91 			case NAME_ERR_NOLETTER:
92 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
93 				    "name must begin with a letter"));
94 				break;
95 
96 			case NAME_ERR_RESERVED:
97 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
98 				    "name is reserved"));
99 				break;
100 
101 			case NAME_ERR_DISKLIKE:
102 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
103 				    "pool name is reserved"));
104 				break;
105 
106 			case NAME_ERR_LEADING_SLASH:
107 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
108 				    "leading slash in name"));
109 				break;
110 
111 			case NAME_ERR_EMPTY_COMPONENT:
112 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
113 				    "empty component in name"));
114 				break;
115 
116 			case NAME_ERR_TRAILING_SLASH:
117 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
118 				    "trailing slash in name"));
119 				break;
120 
121 			case NAME_ERR_MULTIPLE_AT:
122 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
123 				    "multiple '@' delimiters in name"));
124 				break;
125 
126 			}
127 		}
128 		return (B_FALSE);
129 	}
130 
131 	return (B_TRUE);
132 }
133 
134 /*
135  * Open a handle to the given pool, even if the pool is currently in the FAULTED
136  * state.
137  */
138 zpool_handle_t *
139 zpool_open_canfail(libzfs_handle_t *hdl, const char *pool)
140 {
141 	zpool_handle_t *zhp;
142 	boolean_t missing;
143 
144 	/*
145 	 * Make sure the pool name is valid.
146 	 */
147 	if (!zpool_name_valid(hdl, B_TRUE, pool)) {
148 		(void) zfs_error_fmt(hdl, EZFS_INVALIDNAME,
149 		    dgettext(TEXT_DOMAIN, "cannot open '%s'"),
150 		    pool);
151 		return (NULL);
152 	}
153 
154 	if ((zhp = zfs_alloc(hdl, sizeof (zpool_handle_t))) == NULL)
155 		return (NULL);
156 
157 	zhp->zpool_hdl = hdl;
158 	(void) strlcpy(zhp->zpool_name, pool, sizeof (zhp->zpool_name));
159 
160 	if (zpool_refresh_stats(zhp, &missing) != 0) {
161 		zpool_close(zhp);
162 		return (NULL);
163 	}
164 
165 	if (missing) {
166 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
167 		    "no such pool"));
168 		(void) zfs_error_fmt(hdl, EZFS_NOENT,
169 		    dgettext(TEXT_DOMAIN, "cannot open '%s'"),
170 		    pool);
171 		zpool_close(zhp);
172 		return (NULL);
173 	}
174 
175 	return (zhp);
176 }
177 
178 /*
179  * Like the above, but silent on error.  Used when iterating over pools (because
180  * the configuration cache may be out of date).
181  */
182 int
183 zpool_open_silent(libzfs_handle_t *hdl, const char *pool, zpool_handle_t **ret)
184 {
185 	zpool_handle_t *zhp;
186 	boolean_t missing;
187 
188 	if ((zhp = zfs_alloc(hdl, sizeof (zpool_handle_t))) == NULL)
189 		return (-1);
190 
191 	zhp->zpool_hdl = hdl;
192 	(void) strlcpy(zhp->zpool_name, pool, sizeof (zhp->zpool_name));
193 
194 	if (zpool_refresh_stats(zhp, &missing) != 0) {
195 		zpool_close(zhp);
196 		return (-1);
197 	}
198 
199 	if (missing) {
200 		zpool_close(zhp);
201 		*ret = NULL;
202 		return (0);
203 	}
204 
205 	*ret = zhp;
206 	return (0);
207 }
208 
209 /*
210  * Similar to zpool_open_canfail(), but refuses to open pools in the faulted
211  * state.
212  */
213 zpool_handle_t *
214 zpool_open(libzfs_handle_t *hdl, const char *pool)
215 {
216 	zpool_handle_t *zhp;
217 
218 	if ((zhp = zpool_open_canfail(hdl, pool)) == NULL)
219 		return (NULL);
220 
221 	if (zhp->zpool_state == POOL_STATE_UNAVAIL) {
222 		(void) zfs_error_fmt(hdl, EZFS_POOLUNAVAIL,
223 		    dgettext(TEXT_DOMAIN, "cannot open '%s'"), zhp->zpool_name);
224 		zpool_close(zhp);
225 		return (NULL);
226 	}
227 
228 	return (zhp);
229 }
230 
231 /*
232  * Close the handle.  Simply frees the memory associated with the handle.
233  */
234 void
235 zpool_close(zpool_handle_t *zhp)
236 {
237 	if (zhp->zpool_config)
238 		nvlist_free(zhp->zpool_config);
239 	if (zhp->zpool_old_config)
240 		nvlist_free(zhp->zpool_old_config);
241 	free(zhp);
242 }
243 
244 /*
245  * Return the name of the pool.
246  */
247 const char *
248 zpool_get_name(zpool_handle_t *zhp)
249 {
250 	return (zhp->zpool_name);
251 }
252 
253 /*
254  * Return the GUID of the pool.
255  */
256 uint64_t
257 zpool_get_guid(zpool_handle_t *zhp)
258 {
259 	uint64_t guid;
260 
261 	verify(nvlist_lookup_uint64(zhp->zpool_config, ZPOOL_CONFIG_POOL_GUID,
262 	    &guid) == 0);
263 	return (guid);
264 }
265 
266 /*
267  * Return the version of the pool.
268  */
269 uint64_t
270 zpool_get_version(zpool_handle_t *zhp)
271 {
272 	uint64_t version;
273 
274 	verify(nvlist_lookup_uint64(zhp->zpool_config, ZPOOL_CONFIG_VERSION,
275 	    &version) == 0);
276 
277 	return (version);
278 }
279 
280 /*
281  * Return the amount of space currently consumed by the pool.
282  */
283 uint64_t
284 zpool_get_space_used(zpool_handle_t *zhp)
285 {
286 	nvlist_t *nvroot;
287 	vdev_stat_t *vs;
288 	uint_t vsc;
289 
290 	verify(nvlist_lookup_nvlist(zhp->zpool_config, ZPOOL_CONFIG_VDEV_TREE,
291 	    &nvroot) == 0);
292 	verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_STATS,
293 	    (uint64_t **)&vs, &vsc) == 0);
294 
295 	return (vs->vs_alloc);
296 }
297 
298 /*
299  * Return the total space in the pool.
300  */
301 uint64_t
302 zpool_get_space_total(zpool_handle_t *zhp)
303 {
304 	nvlist_t *nvroot;
305 	vdev_stat_t *vs;
306 	uint_t vsc;
307 
308 	verify(nvlist_lookup_nvlist(zhp->zpool_config, ZPOOL_CONFIG_VDEV_TREE,
309 	    &nvroot) == 0);
310 	verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_STATS,
311 	    (uint64_t **)&vs, &vsc) == 0);
312 
313 	return (vs->vs_space);
314 }
315 
316 /*
317  * Return the alternate root for this pool, if any.
318  */
319 int
320 zpool_get_root(zpool_handle_t *zhp, char *buf, size_t buflen)
321 {
322 	zfs_cmd_t zc = { 0 };
323 
324 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
325 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_OBJSET_STATS, &zc) != 0 ||
326 	    zc.zc_value[0] == '\0')
327 		return (-1);
328 
329 	(void) strlcpy(buf, zc.zc_value, buflen);
330 
331 	return (0);
332 }
333 
334 /*
335  * Return the state of the pool (ACTIVE or UNAVAILABLE)
336  */
337 int
338 zpool_get_state(zpool_handle_t *zhp)
339 {
340 	return (zhp->zpool_state);
341 }
342 
343 /*
344  * Create the named pool, using the provided vdev list.  It is assumed
345  * that the consumer has already validated the contents of the nvlist, so we
346  * don't have to worry about error semantics.
347  */
348 int
349 zpool_create(libzfs_handle_t *hdl, const char *pool, nvlist_t *nvroot,
350     const char *altroot)
351 {
352 	zfs_cmd_t zc = { 0 };
353 	char msg[1024];
354 
355 	(void) snprintf(msg, sizeof (msg), dgettext(TEXT_DOMAIN,
356 	    "cannot create '%s'"), pool);
357 
358 	if (!zpool_name_valid(hdl, B_FALSE, pool))
359 		return (zfs_error(hdl, EZFS_INVALIDNAME, msg));
360 
361 	if (altroot != NULL && altroot[0] != '/')
362 		return (zfs_error_fmt(hdl, EZFS_BADPATH,
363 		    dgettext(TEXT_DOMAIN, "bad alternate root '%s'"), altroot));
364 
365 	if (zcmd_write_src_nvlist(hdl, &zc, nvroot, NULL) != 0)
366 		return (-1);
367 
368 	(void) strlcpy(zc.zc_name, pool, sizeof (zc.zc_name));
369 
370 	if (altroot != NULL)
371 		(void) strlcpy(zc.zc_value, altroot, sizeof (zc.zc_value));
372 
373 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_CREATE, &zc) != 0) {
374 		zcmd_free_nvlists(&zc);
375 
376 		switch (errno) {
377 		case EBUSY:
378 			/*
379 			 * This can happen if the user has specified the same
380 			 * device multiple times.  We can't reliably detect this
381 			 * until we try to add it and see we already have a
382 			 * label.
383 			 */
384 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
385 			    "one or more vdevs refer to the same device"));
386 			return (zfs_error(hdl, EZFS_BADDEV, msg));
387 
388 		case EOVERFLOW:
389 			/*
390 			 * This occurs when one of the devices is below
391 			 * SPA_MINDEVSIZE.  Unfortunately, we can't detect which
392 			 * device was the problem device since there's no
393 			 * reliable way to determine device size from userland.
394 			 */
395 			{
396 				char buf[64];
397 
398 				zfs_nicenum(SPA_MINDEVSIZE, buf, sizeof (buf));
399 
400 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
401 				    "one or more devices is less than the "
402 				    "minimum size (%s)"), buf);
403 			}
404 			return (zfs_error(hdl, EZFS_BADDEV, msg));
405 
406 		case ENOSPC:
407 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
408 			    "one or more devices is out of space"));
409 			return (zfs_error(hdl, EZFS_BADDEV, msg));
410 
411 		default:
412 			return (zpool_standard_error(hdl, errno, msg));
413 		}
414 	}
415 
416 	zcmd_free_nvlists(&zc);
417 
418 	/*
419 	 * If this is an alternate root pool, then we automatically set the
420 	 * mountpoint of the root dataset to be '/'.
421 	 */
422 	if (altroot != NULL) {
423 		zfs_handle_t *zhp;
424 
425 		verify((zhp = zfs_open(hdl, pool, ZFS_TYPE_ANY)) != NULL);
426 		verify(zfs_prop_set(zhp, zfs_prop_to_name(ZFS_PROP_MOUNTPOINT),
427 		    "/") == 0);
428 
429 		zfs_close(zhp);
430 	}
431 
432 	return (0);
433 }
434 
435 /*
436  * Destroy the given pool.  It is up to the caller to ensure that there are no
437  * datasets left in the pool.
438  */
439 int
440 zpool_destroy(zpool_handle_t *zhp)
441 {
442 	zfs_cmd_t zc = { 0 };
443 	zfs_handle_t *zfp = NULL;
444 	libzfs_handle_t *hdl = zhp->zpool_hdl;
445 	char msg[1024];
446 
447 	if (zhp->zpool_state == POOL_STATE_ACTIVE &&
448 	    (zfp = zfs_open(zhp->zpool_hdl, zhp->zpool_name,
449 	    ZFS_TYPE_FILESYSTEM)) == NULL)
450 		return (-1);
451 
452 	if (zpool_remove_zvol_links(zhp) != 0)
453 		return (-1);
454 
455 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
456 
457 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_POOL_DESTROY, &zc) != 0) {
458 		(void) snprintf(msg, sizeof (msg), dgettext(TEXT_DOMAIN,
459 		    "cannot destroy '%s'"), zhp->zpool_name);
460 
461 		if (errno == EROFS) {
462 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
463 			    "one or more devices is read only"));
464 			(void) zfs_error(hdl, EZFS_BADDEV, msg);
465 		} else {
466 			(void) zpool_standard_error(hdl, errno, msg);
467 		}
468 
469 		if (zfp)
470 			zfs_close(zfp);
471 		return (-1);
472 	}
473 
474 	if (zfp) {
475 		remove_mountpoint(zfp);
476 		zfs_close(zfp);
477 	}
478 
479 	return (0);
480 }
481 
482 /*
483  * Add the given vdevs to the pool.  The caller must have already performed the
484  * necessary verification to ensure that the vdev specification is well-formed.
485  */
486 int
487 zpool_add(zpool_handle_t *zhp, nvlist_t *nvroot)
488 {
489 	zfs_cmd_t zc = { 0 };
490 	int ret;
491 	libzfs_handle_t *hdl = zhp->zpool_hdl;
492 	char msg[1024];
493 	nvlist_t **spares;
494 	uint_t nspares;
495 
496 	(void) snprintf(msg, sizeof (msg), dgettext(TEXT_DOMAIN,
497 	    "cannot add to '%s'"), zhp->zpool_name);
498 
499 	if (zpool_get_version(zhp) < ZFS_VERSION_SPARES &&
500 	    nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_SPARES,
501 	    &spares, &nspares) == 0) {
502 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN, "pool must be "
503 		    "upgraded to add hot spares"));
504 		return (zfs_error(hdl, EZFS_BADVERSION, msg));
505 	}
506 
507 	if (zcmd_write_src_nvlist(hdl, &zc, nvroot, NULL) != 0)
508 		return (-1);
509 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
510 
511 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_VDEV_ADD, &zc) != 0) {
512 		switch (errno) {
513 		case EBUSY:
514 			/*
515 			 * This can happen if the user has specified the same
516 			 * device multiple times.  We can't reliably detect this
517 			 * until we try to add it and see we already have a
518 			 * label.
519 			 */
520 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
521 			    "one or more vdevs refer to the same device"));
522 			(void) zfs_error(hdl, EZFS_BADDEV, msg);
523 			break;
524 
525 		case EOVERFLOW:
526 			/*
527 			 * This occurrs when one of the devices is below
528 			 * SPA_MINDEVSIZE.  Unfortunately, we can't detect which
529 			 * device was the problem device since there's no
530 			 * reliable way to determine device size from userland.
531 			 */
532 			{
533 				char buf[64];
534 
535 				zfs_nicenum(SPA_MINDEVSIZE, buf, sizeof (buf));
536 
537 				zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
538 				    "device is less than the minimum "
539 				    "size (%s)"), buf);
540 			}
541 			(void) zfs_error(hdl, EZFS_BADDEV, msg);
542 			break;
543 
544 		case ENOTSUP:
545 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
546 			    "pool must be upgraded to add raidz2 vdevs"));
547 			(void) zfs_error(hdl, EZFS_BADVERSION, msg);
548 			break;
549 
550 		default:
551 			(void) zpool_standard_error(hdl, errno, msg);
552 		}
553 
554 		ret = -1;
555 	} else {
556 		ret = 0;
557 	}
558 
559 	zcmd_free_nvlists(&zc);
560 
561 	return (ret);
562 }
563 
564 /*
565  * Exports the pool from the system.  The caller must ensure that there are no
566  * mounted datasets in the pool.
567  */
568 int
569 zpool_export(zpool_handle_t *zhp)
570 {
571 	zfs_cmd_t zc = { 0 };
572 
573 	if (zpool_remove_zvol_links(zhp) != 0)
574 		return (-1);
575 
576 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
577 
578 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_POOL_EXPORT, &zc) != 0)
579 		return (zpool_standard_error_fmt(zhp->zpool_hdl, errno,
580 		    dgettext(TEXT_DOMAIN, "cannot export '%s'"),
581 		    zhp->zpool_name));
582 
583 	return (0);
584 }
585 
586 /*
587  * Import the given pool using the known configuration.  The configuration
588  * should have come from zpool_find_import().  The 'newname' and 'altroot'
589  * parameters control whether the pool is imported with a different name or with
590  * an alternate root, respectively.
591  */
592 int
593 zpool_import(libzfs_handle_t *hdl, nvlist_t *config, const char *newname,
594     const char *altroot)
595 {
596 	zfs_cmd_t zc = { 0 };
597 	char *thename;
598 	char *origname;
599 	int ret;
600 
601 	verify(nvlist_lookup_string(config, ZPOOL_CONFIG_POOL_NAME,
602 	    &origname) == 0);
603 
604 	if (newname != NULL) {
605 		if (!zpool_name_valid(hdl, B_FALSE, newname))
606 			return (zfs_error_fmt(hdl, EZFS_INVALIDNAME,
607 			    dgettext(TEXT_DOMAIN, "cannot import '%s'"),
608 			    newname));
609 		thename = (char *)newname;
610 	} else {
611 		thename = origname;
612 	}
613 
614 	if (altroot != NULL && altroot[0] != '/')
615 		return (zfs_error_fmt(hdl, EZFS_BADPATH,
616 		    dgettext(TEXT_DOMAIN, "bad alternate root '%s'"),
617 		    altroot));
618 
619 	(void) strlcpy(zc.zc_name, thename, sizeof (zc.zc_name));
620 
621 	if (altroot != NULL)
622 		(void) strlcpy(zc.zc_value, altroot, sizeof (zc.zc_value));
623 	else
624 		zc.zc_value[0] = '\0';
625 
626 	verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_GUID,
627 	    &zc.zc_guid) == 0);
628 
629 	if (zcmd_write_src_nvlist(hdl, &zc, config, NULL) != 0)
630 		return (-1);
631 
632 	ret = 0;
633 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_IMPORT, &zc) != 0) {
634 		char desc[1024];
635 		if (newname == NULL)
636 			(void) snprintf(desc, sizeof (desc),
637 			    dgettext(TEXT_DOMAIN, "cannot import '%s'"),
638 			    thename);
639 		else
640 			(void) snprintf(desc, sizeof (desc),
641 			    dgettext(TEXT_DOMAIN, "cannot import '%s' as '%s'"),
642 			    origname, thename);
643 
644 		switch (errno) {
645 		case ENOTSUP:
646 			/*
647 			 * Unsupported version.
648 			 */
649 			(void) zfs_error(hdl, EZFS_BADVERSION, desc);
650 			break;
651 
652 		case EINVAL:
653 			(void) zfs_error(hdl, EZFS_INVALCONFIG, desc);
654 			break;
655 
656 		default:
657 			(void) zpool_standard_error(hdl, errno, desc);
658 		}
659 
660 		ret = -1;
661 	} else {
662 		zpool_handle_t *zhp;
663 		/*
664 		 * This should never fail, but play it safe anyway.
665 		 */
666 		if (zpool_open_silent(hdl, thename, &zhp) != 0) {
667 			ret = -1;
668 		} else if (zhp != NULL) {
669 			ret = zpool_create_zvol_links(zhp);
670 			zpool_close(zhp);
671 		}
672 	}
673 
674 	zcmd_free_nvlists(&zc);
675 	return (ret);
676 }
677 
678 /*
679  * Scrub the pool.
680  */
681 int
682 zpool_scrub(zpool_handle_t *zhp, pool_scrub_type_t type)
683 {
684 	zfs_cmd_t zc = { 0 };
685 	char msg[1024];
686 	libzfs_handle_t *hdl = zhp->zpool_hdl;
687 
688 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
689 	zc.zc_cookie = type;
690 
691 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_POOL_SCRUB, &zc) == 0)
692 		return (0);
693 
694 	(void) snprintf(msg, sizeof (msg),
695 	    dgettext(TEXT_DOMAIN, "cannot scrub %s"), zc.zc_name);
696 
697 	if (errno == EBUSY)
698 		return (zfs_error(hdl, EZFS_RESILVERING, msg));
699 	else
700 		return (zpool_standard_error(hdl, errno, msg));
701 }
702 
703 /*
704  * 'avail_spare' is set to TRUE if the provided guid refers to an AVAIL
705  * spare; but FALSE if its an INUSE spare.
706  */
707 static nvlist_t *
708 vdev_to_nvlist_iter(nvlist_t *nv, const char *search, uint64_t guid,
709     boolean_t *avail_spare)
710 {
711 	uint_t c, children;
712 	nvlist_t **child;
713 	uint64_t theguid, present;
714 	char *path;
715 	uint64_t wholedisk = 0;
716 	nvlist_t *ret;
717 
718 	verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_GUID, &theguid) == 0);
719 
720 	if (search == NULL &&
721 	    nvlist_lookup_uint64(nv, ZPOOL_CONFIG_NOT_PRESENT, &present) == 0) {
722 		/*
723 		 * If the device has never been present since import, the only
724 		 * reliable way to match the vdev is by GUID.
725 		 */
726 		if (theguid == guid)
727 			return (nv);
728 	} else if (search != NULL &&
729 	    nvlist_lookup_string(nv, ZPOOL_CONFIG_PATH, &path) == 0) {
730 		(void) nvlist_lookup_uint64(nv, ZPOOL_CONFIG_WHOLE_DISK,
731 		    &wholedisk);
732 		if (wholedisk) {
733 			/*
734 			 * For whole disks, the internal path has 's0', but the
735 			 * path passed in by the user doesn't.
736 			 */
737 			if (strlen(search) == strlen(path) - 2 &&
738 			    strncmp(search, path, strlen(search)) == 0)
739 				return (nv);
740 		} else if (strcmp(search, path) == 0) {
741 			return (nv);
742 		}
743 	}
744 
745 	if (nvlist_lookup_nvlist_array(nv, ZPOOL_CONFIG_CHILDREN,
746 	    &child, &children) != 0)
747 		return (NULL);
748 
749 	for (c = 0; c < children; c++)
750 		if ((ret = vdev_to_nvlist_iter(child[c], search, guid,
751 		    avail_spare)) != NULL)
752 			return (ret);
753 
754 	if (nvlist_lookup_nvlist_array(nv, ZPOOL_CONFIG_SPARES,
755 	    &child, &children) == 0) {
756 		for (c = 0; c < children; c++) {
757 			if ((ret = vdev_to_nvlist_iter(child[c], search, guid,
758 			    avail_spare)) != NULL) {
759 				*avail_spare = B_TRUE;
760 				return (ret);
761 			}
762 		}
763 	}
764 
765 	return (NULL);
766 }
767 
768 nvlist_t *
769 zpool_find_vdev(zpool_handle_t *zhp, const char *path, boolean_t *avail_spare)
770 {
771 	char buf[MAXPATHLEN];
772 	const char *search;
773 	char *end;
774 	nvlist_t *nvroot;
775 	uint64_t guid;
776 
777 	guid = strtoull(path, &end, 10);
778 	if (guid != 0 && *end == '\0') {
779 		search = NULL;
780 	} else if (path[0] != '/') {
781 		(void) snprintf(buf, sizeof (buf), "%s%s", "/dev/dsk/", path);
782 		search = buf;
783 	} else {
784 		search = path;
785 	}
786 
787 	verify(nvlist_lookup_nvlist(zhp->zpool_config, ZPOOL_CONFIG_VDEV_TREE,
788 	    &nvroot) == 0);
789 
790 	*avail_spare = B_FALSE;
791 	return (vdev_to_nvlist_iter(nvroot, search, guid, avail_spare));
792 }
793 
794 /*
795  * Returns TRUE if the given guid corresponds to a spare (INUSE or not).
796  */
797 static boolean_t
798 is_spare(zpool_handle_t *zhp, uint64_t guid)
799 {
800 	uint64_t spare_guid;
801 	nvlist_t *nvroot;
802 	nvlist_t **spares;
803 	uint_t nspares;
804 	int i;
805 
806 	verify(nvlist_lookup_nvlist(zhp->zpool_config, ZPOOL_CONFIG_VDEV_TREE,
807 	    &nvroot) == 0);
808 	if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_SPARES,
809 	    &spares, &nspares) == 0) {
810 		for (i = 0; i < nspares; i++) {
811 			verify(nvlist_lookup_uint64(spares[i],
812 			    ZPOOL_CONFIG_GUID, &spare_guid) == 0);
813 			if (guid == spare_guid)
814 				return (B_TRUE);
815 		}
816 	}
817 
818 	return (B_FALSE);
819 }
820 
821 /*
822  * Bring the specified vdev online
823  */
824 int
825 zpool_vdev_online(zpool_handle_t *zhp, const char *path)
826 {
827 	zfs_cmd_t zc = { 0 };
828 	char msg[1024];
829 	nvlist_t *tgt;
830 	boolean_t avail_spare;
831 	libzfs_handle_t *hdl = zhp->zpool_hdl;
832 
833 	(void) snprintf(msg, sizeof (msg),
834 	    dgettext(TEXT_DOMAIN, "cannot online %s"), path);
835 
836 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
837 	if ((tgt = zpool_find_vdev(zhp, path, &avail_spare)) == NULL)
838 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
839 
840 	verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID, &zc.zc_guid) == 0);
841 
842 	if (avail_spare || is_spare(zhp, zc.zc_guid) == B_TRUE)
843 		return (zfs_error(hdl, EZFS_ISSPARE, msg));
844 
845 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_VDEV_ONLINE, &zc) == 0)
846 		return (0);
847 
848 	return (zpool_standard_error(hdl, errno, msg));
849 }
850 
851 /*
852  * Take the specified vdev offline
853  */
854 int
855 zpool_vdev_offline(zpool_handle_t *zhp, const char *path, int istmp)
856 {
857 	zfs_cmd_t zc = { 0 };
858 	char msg[1024];
859 	nvlist_t *tgt;
860 	boolean_t avail_spare;
861 	libzfs_handle_t *hdl = zhp->zpool_hdl;
862 
863 	(void) snprintf(msg, sizeof (msg),
864 	    dgettext(TEXT_DOMAIN, "cannot offline %s"), path);
865 
866 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
867 	if ((tgt = zpool_find_vdev(zhp, path, &avail_spare)) == NULL)
868 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
869 
870 	verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID, &zc.zc_guid) == 0);
871 
872 	if (avail_spare || is_spare(zhp, zc.zc_guid) == B_TRUE)
873 		return (zfs_error(hdl, EZFS_ISSPARE, msg));
874 
875 	zc.zc_cookie = istmp;
876 
877 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_VDEV_OFFLINE, &zc) == 0)
878 		return (0);
879 
880 	switch (errno) {
881 	case EBUSY:
882 
883 		/*
884 		 * There are no other replicas of this device.
885 		 */
886 		return (zfs_error(hdl, EZFS_NOREPLICAS, msg));
887 
888 	default:
889 		return (zpool_standard_error(hdl, errno, msg));
890 	}
891 }
892 
893 /*
894  * Returns TRUE if the given nvlist is a vdev that was originally swapped in as
895  * a hot spare.
896  */
897 static boolean_t
898 is_replacing_spare(nvlist_t *search, nvlist_t *tgt, int which)
899 {
900 	nvlist_t **child;
901 	uint_t c, children;
902 	char *type;
903 
904 	if (nvlist_lookup_nvlist_array(search, ZPOOL_CONFIG_CHILDREN, &child,
905 	    &children) == 0) {
906 		verify(nvlist_lookup_string(search, ZPOOL_CONFIG_TYPE,
907 		    &type) == 0);
908 
909 		if (strcmp(type, VDEV_TYPE_SPARE) == 0 &&
910 		    children == 2 && child[which] == tgt)
911 			return (B_TRUE);
912 
913 		for (c = 0; c < children; c++)
914 			if (is_replacing_spare(child[c], tgt, which))
915 				return (B_TRUE);
916 	}
917 
918 	return (B_FALSE);
919 }
920 
921 /*
922  * Attach new_disk (fully described by nvroot) to old_disk.
923  * If 'replacing' is specified, tne new disk will replace the old one.
924  */
925 int
926 zpool_vdev_attach(zpool_handle_t *zhp,
927     const char *old_disk, const char *new_disk, nvlist_t *nvroot, int replacing)
928 {
929 	zfs_cmd_t zc = { 0 };
930 	char msg[1024];
931 	int ret;
932 	nvlist_t *tgt;
933 	boolean_t avail_spare;
934 	uint64_t val;
935 	char *path;
936 	nvlist_t **child;
937 	uint_t children;
938 	nvlist_t *config_root;
939 	libzfs_handle_t *hdl = zhp->zpool_hdl;
940 
941 	if (replacing)
942 		(void) snprintf(msg, sizeof (msg), dgettext(TEXT_DOMAIN,
943 		    "cannot replace %s with %s"), old_disk, new_disk);
944 	else
945 		(void) snprintf(msg, sizeof (msg), dgettext(TEXT_DOMAIN,
946 		    "cannot attach %s to %s"), new_disk, old_disk);
947 
948 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
949 	if ((tgt = zpool_find_vdev(zhp, old_disk, &avail_spare)) == 0)
950 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
951 
952 	if (avail_spare)
953 		return (zfs_error(hdl, EZFS_ISSPARE, msg));
954 
955 	verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID, &zc.zc_guid) == 0);
956 	zc.zc_cookie = replacing;
957 
958 	if (nvlist_lookup_nvlist_array(nvroot, ZPOOL_CONFIG_CHILDREN,
959 	    &child, &children) != 0 || children != 1) {
960 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
961 		    "new device must be a single disk"));
962 		return (zfs_error(hdl, EZFS_INVALCONFIG, msg));
963 	}
964 
965 	verify(nvlist_lookup_nvlist(zpool_get_config(zhp, NULL),
966 	    ZPOOL_CONFIG_VDEV_TREE, &config_root) == 0);
967 
968 	/*
969 	 * If the target is a hot spare that has been swapped in, we can only
970 	 * replace it with another hot spare.
971 	 */
972 	if (replacing &&
973 	    nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_IS_SPARE, &val) == 0 &&
974 	    nvlist_lookup_string(child[0], ZPOOL_CONFIG_PATH, &path) == 0 &&
975 	    (zpool_find_vdev(zhp, path, &avail_spare) == NULL ||
976 	    !avail_spare) && is_replacing_spare(config_root, tgt, 1)) {
977 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
978 		    "can only be replaced by another hot spare"));
979 		return (zfs_error(hdl, EZFS_BADTARGET, msg));
980 	}
981 
982 	/*
983 	 * If we are attempting to replace a spare, it canot be applied to an
984 	 * already spared device.
985 	 */
986 	if (replacing &&
987 	    nvlist_lookup_string(child[0], ZPOOL_CONFIG_PATH, &path) == 0 &&
988 	    zpool_find_vdev(zhp, path, &avail_spare) != NULL && avail_spare &&
989 	    is_replacing_spare(config_root, tgt, 0)) {
990 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
991 		    "device has already been replaced with a spare"));
992 		return (zfs_error(hdl, EZFS_BADTARGET, msg));
993 	}
994 
995 	if (zcmd_write_src_nvlist(hdl, &zc, nvroot, NULL) != 0)
996 		return (-1);
997 
998 	ret = ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_VDEV_ATTACH, &zc);
999 
1000 	zcmd_free_nvlists(&zc);
1001 
1002 	if (ret == 0)
1003 		return (0);
1004 
1005 	switch (errno) {
1006 	case ENOTSUP:
1007 		/*
1008 		 * Can't attach to or replace this type of vdev.
1009 		 */
1010 		if (replacing)
1011 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1012 			    "cannot replace a replacing device"));
1013 		else
1014 			zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1015 			    "can only attach to mirrors and top-level "
1016 			    "disks"));
1017 		(void) zfs_error(hdl, EZFS_BADTARGET, msg);
1018 		break;
1019 
1020 	case EINVAL:
1021 		/*
1022 		 * The new device must be a single disk.
1023 		 */
1024 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1025 		    "new device must be a single disk"));
1026 		(void) zfs_error(hdl, EZFS_INVALCONFIG, msg);
1027 		break;
1028 
1029 	case EBUSY:
1030 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN, "%s is busy"),
1031 		    new_disk);
1032 		(void) zfs_error(hdl, EZFS_BADDEV, msg);
1033 		break;
1034 
1035 	case EOVERFLOW:
1036 		/*
1037 		 * The new device is too small.
1038 		 */
1039 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1040 		    "device is too small"));
1041 		(void) zfs_error(hdl, EZFS_BADDEV, msg);
1042 		break;
1043 
1044 	case EDOM:
1045 		/*
1046 		 * The new device has a different alignment requirement.
1047 		 */
1048 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1049 		    "devices have different sector alignment"));
1050 		(void) zfs_error(hdl, EZFS_BADDEV, msg);
1051 		break;
1052 
1053 	case ENAMETOOLONG:
1054 		/*
1055 		 * The resulting top-level vdev spec won't fit in the label.
1056 		 */
1057 		(void) zfs_error(hdl, EZFS_DEVOVERFLOW, msg);
1058 		break;
1059 
1060 	default:
1061 		(void) zpool_standard_error(hdl, errno, msg);
1062 	}
1063 
1064 	return (-1);
1065 }
1066 
1067 /*
1068  * Detach the specified device.
1069  */
1070 int
1071 zpool_vdev_detach(zpool_handle_t *zhp, const char *path)
1072 {
1073 	zfs_cmd_t zc = { 0 };
1074 	char msg[1024];
1075 	nvlist_t *tgt;
1076 	boolean_t avail_spare;
1077 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1078 
1079 	(void) snprintf(msg, sizeof (msg),
1080 	    dgettext(TEXT_DOMAIN, "cannot detach %s"), path);
1081 
1082 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1083 	if ((tgt = zpool_find_vdev(zhp, path, &avail_spare)) == 0)
1084 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
1085 
1086 	if (avail_spare)
1087 		return (zfs_error(hdl, EZFS_ISSPARE, msg));
1088 
1089 	verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID, &zc.zc_guid) == 0);
1090 
1091 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_VDEV_DETACH, &zc) == 0)
1092 		return (0);
1093 
1094 	switch (errno) {
1095 
1096 	case ENOTSUP:
1097 		/*
1098 		 * Can't detach from this type of vdev.
1099 		 */
1100 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN, "only "
1101 		    "applicable to mirror and replacing vdevs"));
1102 		(void) zfs_error(zhp->zpool_hdl, EZFS_BADTARGET, msg);
1103 		break;
1104 
1105 	case EBUSY:
1106 		/*
1107 		 * There are no other replicas of this device.
1108 		 */
1109 		(void) zfs_error(hdl, EZFS_NOREPLICAS, msg);
1110 		break;
1111 
1112 	default:
1113 		(void) zpool_standard_error(hdl, errno, msg);
1114 	}
1115 
1116 	return (-1);
1117 }
1118 
1119 /*
1120  * Remove the given device.  Currently, this is supported only for hot spares.
1121  */
1122 int
1123 zpool_vdev_remove(zpool_handle_t *zhp, const char *path)
1124 {
1125 	zfs_cmd_t zc = { 0 };
1126 	char msg[1024];
1127 	nvlist_t *tgt;
1128 	boolean_t avail_spare;
1129 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1130 
1131 	(void) snprintf(msg, sizeof (msg),
1132 	    dgettext(TEXT_DOMAIN, "cannot remove %s"), path);
1133 
1134 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1135 	if ((tgt = zpool_find_vdev(zhp, path, &avail_spare)) == 0)
1136 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
1137 
1138 	if (!avail_spare) {
1139 		zfs_error_aux(hdl, dgettext(TEXT_DOMAIN,
1140 		    "only inactive hot spares can be removed"));
1141 		return (zfs_error(hdl, EZFS_NODEVICE, msg));
1142 	}
1143 
1144 	verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID, &zc.zc_guid) == 0);
1145 
1146 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_VDEV_REMOVE, &zc) == 0)
1147 		return (0);
1148 
1149 	return (zpool_standard_error(hdl, errno, msg));
1150 }
1151 
1152 /*
1153  * Clear the errors for the pool, or the particular device if specified.
1154  */
1155 int
1156 zpool_clear(zpool_handle_t *zhp, const char *path)
1157 {
1158 	zfs_cmd_t zc = { 0 };
1159 	char msg[1024];
1160 	nvlist_t *tgt;
1161 	boolean_t avail_spare;
1162 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1163 
1164 	if (path)
1165 		(void) snprintf(msg, sizeof (msg),
1166 		    dgettext(TEXT_DOMAIN, "cannot clear errors for %s"),
1167 		    path);
1168 	else
1169 		(void) snprintf(msg, sizeof (msg),
1170 		    dgettext(TEXT_DOMAIN, "cannot clear errors for %s"),
1171 		    zhp->zpool_name);
1172 
1173 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1174 	if (path) {
1175 		if ((tgt = zpool_find_vdev(zhp, path, &avail_spare)) == 0)
1176 			return (zfs_error(hdl, EZFS_NODEVICE, msg));
1177 
1178 		if (avail_spare)
1179 			return (zfs_error(hdl, EZFS_ISSPARE, msg));
1180 
1181 		verify(nvlist_lookup_uint64(tgt, ZPOOL_CONFIG_GUID,
1182 		    &zc.zc_guid) == 0);
1183 	}
1184 
1185 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_CLEAR, &zc) == 0)
1186 		return (0);
1187 
1188 	return (zpool_standard_error(hdl, errno, msg));
1189 }
1190 
1191 /*
1192  * Iterate over all zvols in a given pool by walking the /dev/zvol/dsk/<pool>
1193  * hierarchy.
1194  */
1195 int
1196 zpool_iter_zvol(zpool_handle_t *zhp, int (*cb)(const char *, void *),
1197     void *data)
1198 {
1199 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1200 	char (*paths)[MAXPATHLEN];
1201 	size_t size = 4;
1202 	int curr, fd, base, ret = 0;
1203 	DIR *dirp;
1204 	struct dirent *dp;
1205 	struct stat st;
1206 
1207 	if ((base = open("/dev/zvol/dsk", O_RDONLY)) < 0)
1208 		return (errno == ENOENT ? 0 : -1);
1209 
1210 	if (fstatat(base, zhp->zpool_name, &st, 0) != 0) {
1211 		int err = errno;
1212 		(void) close(base);
1213 		return (err == ENOENT ? 0 : -1);
1214 	}
1215 
1216 	/*
1217 	 * Oddly this wasn't a directory -- ignore that failure since we
1218 	 * know there are no links lower in the (non-existant) hierarchy.
1219 	 */
1220 	if (!S_ISDIR(st.st_mode)) {
1221 		(void) close(base);
1222 		return (0);
1223 	}
1224 
1225 	if ((paths = zfs_alloc(hdl, size * sizeof (paths[0]))) == NULL) {
1226 		(void) close(base);
1227 		return (-1);
1228 	}
1229 
1230 	(void) strlcpy(paths[0], zhp->zpool_name, sizeof (paths[0]));
1231 	curr = 0;
1232 
1233 	while (curr >= 0) {
1234 		if (fstatat(base, paths[curr], &st, AT_SYMLINK_NOFOLLOW) != 0)
1235 			goto err;
1236 
1237 		if (S_ISDIR(st.st_mode)) {
1238 			if ((fd = openat(base, paths[curr], O_RDONLY)) < 0)
1239 				goto err;
1240 
1241 			if ((dirp = fdopendir(fd)) == NULL) {
1242 				(void) close(fd);
1243 				goto err;
1244 			}
1245 
1246 			while ((dp = readdir(dirp)) != NULL) {
1247 				if (dp->d_name[0] == '.')
1248 					continue;
1249 
1250 				if (curr + 1 == size) {
1251 					paths = zfs_realloc(hdl, paths,
1252 					    size * sizeof (paths[0]),
1253 					    size * 2 * sizeof (paths[0]));
1254 					if (paths == NULL) {
1255 						(void) closedir(dirp);
1256 						(void) close(fd);
1257 						goto err;
1258 					}
1259 
1260 					size *= 2;
1261 				}
1262 
1263 				(void) strlcpy(paths[curr + 1], paths[curr],
1264 				    sizeof (paths[curr + 1]));
1265 				(void) strlcat(paths[curr], "/",
1266 				    sizeof (paths[curr]));
1267 				(void) strlcat(paths[curr], dp->d_name,
1268 				    sizeof (paths[curr]));
1269 				curr++;
1270 			}
1271 
1272 			(void) closedir(dirp);
1273 
1274 		} else {
1275 			if ((ret = cb(paths[curr], data)) != 0)
1276 				break;
1277 		}
1278 
1279 		curr--;
1280 	}
1281 
1282 	free(paths);
1283 	(void) close(base);
1284 
1285 	return (ret);
1286 
1287 err:
1288 	free(paths);
1289 	(void) close(base);
1290 	return (-1);
1291 }
1292 
1293 typedef struct zvol_cb {
1294 	zpool_handle_t *zcb_pool;
1295 	boolean_t zcb_create;
1296 } zvol_cb_t;
1297 
1298 /*ARGSUSED*/
1299 static int
1300 do_zvol_create(zfs_handle_t *zhp, void *data)
1301 {
1302 	int ret;
1303 
1304 	if (ZFS_IS_VOLUME(zhp))
1305 		(void) zvol_create_link(zhp->zfs_hdl, zhp->zfs_name);
1306 
1307 	ret = zfs_iter_children(zhp, do_zvol_create, NULL);
1308 
1309 	zfs_close(zhp);
1310 
1311 	return (ret);
1312 }
1313 
1314 /*
1315  * Iterate over all zvols in the pool and make any necessary minor nodes.
1316  */
1317 int
1318 zpool_create_zvol_links(zpool_handle_t *zhp)
1319 {
1320 	zfs_handle_t *zfp;
1321 	int ret;
1322 
1323 	/*
1324 	 * If the pool is unavailable, just return success.
1325 	 */
1326 	if ((zfp = make_dataset_handle(zhp->zpool_hdl,
1327 	    zhp->zpool_name)) == NULL)
1328 		return (0);
1329 
1330 	ret = zfs_iter_children(zfp, do_zvol_create, NULL);
1331 
1332 	zfs_close(zfp);
1333 	return (ret);
1334 }
1335 
1336 static int
1337 do_zvol_remove(const char *dataset, void *data)
1338 {
1339 	zpool_handle_t *zhp = data;
1340 
1341 	return (zvol_remove_link(zhp->zpool_hdl, dataset));
1342 }
1343 
1344 /*
1345  * Iterate over all zvols in the pool and remove any minor nodes.  We iterate
1346  * by examining the /dev links so that a corrupted pool doesn't impede this
1347  * operation.
1348  */
1349 int
1350 zpool_remove_zvol_links(zpool_handle_t *zhp)
1351 {
1352 	return (zpool_iter_zvol(zhp, do_zvol_remove, zhp));
1353 }
1354 
1355 /*
1356  * Convert from a devid string to a path.
1357  */
1358 static char *
1359 devid_to_path(char *devid_str)
1360 {
1361 	ddi_devid_t devid;
1362 	char *minor;
1363 	char *path;
1364 	devid_nmlist_t *list = NULL;
1365 	int ret;
1366 
1367 	if (devid_str_decode(devid_str, &devid, &minor) != 0)
1368 		return (NULL);
1369 
1370 	ret = devid_deviceid_to_nmlist("/dev", devid, minor, &list);
1371 
1372 	devid_str_free(minor);
1373 	devid_free(devid);
1374 
1375 	if (ret != 0)
1376 		return (NULL);
1377 
1378 	if ((path = strdup(list[0].devname)) == NULL)
1379 		return (NULL);
1380 
1381 	devid_free_nmlist(list);
1382 
1383 	return (path);
1384 }
1385 
1386 /*
1387  * Convert from a path to a devid string.
1388  */
1389 static char *
1390 path_to_devid(const char *path)
1391 {
1392 	int fd;
1393 	ddi_devid_t devid;
1394 	char *minor, *ret;
1395 
1396 	if ((fd = open(path, O_RDONLY)) < 0)
1397 		return (NULL);
1398 
1399 	minor = NULL;
1400 	ret = NULL;
1401 	if (devid_get(fd, &devid) == 0) {
1402 		if (devid_get_minor_name(fd, &minor) == 0)
1403 			ret = devid_str_encode(devid, minor);
1404 		if (minor != NULL)
1405 			devid_str_free(minor);
1406 		devid_free(devid);
1407 	}
1408 	(void) close(fd);
1409 
1410 	return (ret);
1411 }
1412 
1413 /*
1414  * Issue the necessary ioctl() to update the stored path value for the vdev.  We
1415  * ignore any failure here, since a common case is for an unprivileged user to
1416  * type 'zpool status', and we'll display the correct information anyway.
1417  */
1418 static void
1419 set_path(zpool_handle_t *zhp, nvlist_t *nv, const char *path)
1420 {
1421 	zfs_cmd_t zc = { 0 };
1422 
1423 	(void) strncpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1424 	(void) strncpy(zc.zc_value, path, sizeof (zc.zc_value));
1425 	verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_GUID,
1426 	    &zc.zc_guid) == 0);
1427 
1428 	(void) ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_VDEV_SETPATH, &zc);
1429 }
1430 
1431 /*
1432  * Given a vdev, return the name to display in iostat.  If the vdev has a path,
1433  * we use that, stripping off any leading "/dev/dsk/"; if not, we use the type.
1434  * We also check if this is a whole disk, in which case we strip off the
1435  * trailing 's0' slice name.
1436  *
1437  * This routine is also responsible for identifying when disks have been
1438  * reconfigured in a new location.  The kernel will have opened the device by
1439  * devid, but the path will still refer to the old location.  To catch this, we
1440  * first do a path -> devid translation (which is fast for the common case).  If
1441  * the devid matches, we're done.  If not, we do a reverse devid -> path
1442  * translation and issue the appropriate ioctl() to update the path of the vdev.
1443  * If 'zhp' is NULL, then this is an exported pool, and we don't need to do any
1444  * of these checks.
1445  */
1446 char *
1447 zpool_vdev_name(libzfs_handle_t *hdl, zpool_handle_t *zhp, nvlist_t *nv)
1448 {
1449 	char *path, *devid;
1450 	uint64_t value;
1451 	char buf[64];
1452 
1453 	if (nvlist_lookup_uint64(nv, ZPOOL_CONFIG_NOT_PRESENT,
1454 	    &value) == 0) {
1455 		verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_GUID,
1456 		    &value) == 0);
1457 		(void) snprintf(buf, sizeof (buf), "%llu",
1458 		    (u_longlong_t)value);
1459 		path = buf;
1460 	} else if (nvlist_lookup_string(nv, ZPOOL_CONFIG_PATH, &path) == 0) {
1461 
1462 		if (zhp != NULL &&
1463 		    nvlist_lookup_string(nv, ZPOOL_CONFIG_DEVID, &devid) == 0) {
1464 			/*
1465 			 * Determine if the current path is correct.
1466 			 */
1467 			char *newdevid = path_to_devid(path);
1468 
1469 			if (newdevid == NULL ||
1470 			    strcmp(devid, newdevid) != 0) {
1471 				char *newpath;
1472 
1473 				if ((newpath = devid_to_path(devid)) != NULL) {
1474 					/*
1475 					 * Update the path appropriately.
1476 					 */
1477 					set_path(zhp, nv, newpath);
1478 					if (nvlist_add_string(nv,
1479 					    ZPOOL_CONFIG_PATH, newpath) == 0)
1480 						verify(nvlist_lookup_string(nv,
1481 						    ZPOOL_CONFIG_PATH,
1482 						    &path) == 0);
1483 					free(newpath);
1484 				}
1485 			}
1486 
1487 			if (newdevid)
1488 				devid_str_free(newdevid);
1489 		}
1490 
1491 		if (strncmp(path, "/dev/dsk/", 9) == 0)
1492 			path += 9;
1493 
1494 		if (nvlist_lookup_uint64(nv, ZPOOL_CONFIG_WHOLE_DISK,
1495 		    &value) == 0 && value) {
1496 			char *tmp = zfs_strdup(hdl, path);
1497 			if (tmp == NULL)
1498 				return (NULL);
1499 			tmp[strlen(path) - 2] = '\0';
1500 			return (tmp);
1501 		}
1502 	} else {
1503 		verify(nvlist_lookup_string(nv, ZPOOL_CONFIG_TYPE, &path) == 0);
1504 
1505 		/*
1506 		 * If it's a raidz device, we need to stick in the parity level.
1507 		 */
1508 		if (strcmp(path, VDEV_TYPE_RAIDZ) == 0) {
1509 			verify(nvlist_lookup_uint64(nv, ZPOOL_CONFIG_NPARITY,
1510 			    &value) == 0);
1511 			(void) snprintf(buf, sizeof (buf), "%s%llu", path,
1512 			    (u_longlong_t)value);
1513 			path = buf;
1514 		}
1515 	}
1516 
1517 	return (zfs_strdup(hdl, path));
1518 }
1519 
1520 static int
1521 zbookmark_compare(const void *a, const void *b)
1522 {
1523 	return (memcmp(a, b, sizeof (zbookmark_t)));
1524 }
1525 
1526 /*
1527  * Retrieve the persistent error log, uniquify the members, and return to the
1528  * caller.
1529  */
1530 int
1531 zpool_get_errlog(zpool_handle_t *zhp, nvlist_t **nverrlistp)
1532 {
1533 	zfs_cmd_t zc = { 0 };
1534 	uint64_t count;
1535 	zbookmark_t *zb = NULL;
1536 	int i;
1537 
1538 	/*
1539 	 * Retrieve the raw error list from the kernel.  If the number of errors
1540 	 * has increased, allocate more space and continue until we get the
1541 	 * entire list.
1542 	 */
1543 	verify(nvlist_lookup_uint64(zhp->zpool_config, ZPOOL_CONFIG_ERRCOUNT,
1544 	    &count) == 0);
1545 	if ((zc.zc_nvlist_dst = (uintptr_t)zfs_alloc(zhp->zpool_hdl,
1546 	    count * sizeof (zbookmark_t))) == (uintptr_t)NULL)
1547 		return (-1);
1548 	zc.zc_nvlist_dst_size = count;
1549 	(void) strcpy(zc.zc_name, zhp->zpool_name);
1550 	for (;;) {
1551 		if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_ERROR_LOG,
1552 		    &zc) != 0) {
1553 			free((void *)(uintptr_t)zc.zc_nvlist_dst);
1554 			if (errno == ENOMEM) {
1555 				if ((zc.zc_nvlist_dst = (uintptr_t)
1556 				    zfs_alloc(zhp->zpool_hdl,
1557 				    zc.zc_nvlist_dst_size)) == (uintptr_t)NULL)
1558 					return (-1);
1559 			} else {
1560 				return (-1);
1561 			}
1562 		} else {
1563 			break;
1564 		}
1565 	}
1566 
1567 	/*
1568 	 * Sort the resulting bookmarks.  This is a little confusing due to the
1569 	 * implementation of ZFS_IOC_ERROR_LOG.  The bookmarks are copied last
1570 	 * to first, and 'zc_nvlist_dst_size' indicates the number of boomarks
1571 	 * _not_ copied as part of the process.  So we point the start of our
1572 	 * array appropriate and decrement the total number of elements.
1573 	 */
1574 	zb = ((zbookmark_t *)(uintptr_t)zc.zc_nvlist_dst) +
1575 	    zc.zc_nvlist_dst_size;
1576 	count -= zc.zc_nvlist_dst_size;
1577 
1578 	qsort(zb, count, sizeof (zbookmark_t), zbookmark_compare);
1579 
1580 	verify(nvlist_alloc(nverrlistp, 0, KM_SLEEP) == 0);
1581 
1582 	/*
1583 	 * Fill in the nverrlistp with nvlist's of dataset and object numbers.
1584 	 */
1585 	for (i = 0; i < count; i++) {
1586 		nvlist_t *nv;
1587 
1588 		/* ignoring zb_blkid and zb_level for now */
1589 		if (i > 0 && zb[i-1].zb_objset == zb[i].zb_objset &&
1590 		    zb[i-1].zb_object == zb[i].zb_object)
1591 			continue;
1592 
1593 		if (nvlist_alloc(&nv, NV_UNIQUE_NAME, KM_SLEEP) != 0)
1594 			goto nomem;
1595 		if (nvlist_add_uint64(nv, ZPOOL_ERR_DATASET,
1596 		    zb[i].zb_objset) != 0) {
1597 			nvlist_free(nv);
1598 			goto nomem;
1599 		}
1600 		if (nvlist_add_uint64(nv, ZPOOL_ERR_OBJECT,
1601 		    zb[i].zb_object) != 0) {
1602 			nvlist_free(nv);
1603 			goto nomem;
1604 		}
1605 		if (nvlist_add_nvlist(*nverrlistp, "ejk", nv) != 0) {
1606 			nvlist_free(nv);
1607 			goto nomem;
1608 		}
1609 		nvlist_free(nv);
1610 	}
1611 
1612 	free((void *)(uintptr_t)zc.zc_nvlist_dst);
1613 	return (0);
1614 
1615 nomem:
1616 	free((void *)(uintptr_t)zc.zc_nvlist_dst);
1617 	return (no_memory(zhp->zpool_hdl));
1618 }
1619 
1620 /*
1621  * Upgrade a ZFS pool to the latest on-disk version.
1622  */
1623 int
1624 zpool_upgrade(zpool_handle_t *zhp)
1625 {
1626 	zfs_cmd_t zc = { 0 };
1627 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1628 
1629 	(void) strcpy(zc.zc_name, zhp->zpool_name);
1630 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_UPGRADE, &zc) != 0)
1631 		return (zpool_standard_error_fmt(hdl, errno,
1632 		    dgettext(TEXT_DOMAIN, "cannot upgrade '%s'"),
1633 		    zhp->zpool_name));
1634 
1635 	return (0);
1636 }
1637 
1638 /*
1639  * Log command history.
1640  *
1641  * 'pool' is B_TRUE if we are logging a command for 'zpool'; B_FALSE
1642  * otherwise ('zfs').  'pool_create' is B_TRUE if we are logging the creation
1643  * of the pool; B_FALSE otherwise.  'path' is the pathanme containing the
1644  * poolname.  'argc' and 'argv' are used to construct the command string.
1645  */
1646 void
1647 zpool_log_history(libzfs_handle_t *hdl, int argc, char **argv, const char *path,
1648     boolean_t pool, boolean_t pool_create)
1649 {
1650 	char cmd_buf[HIS_MAX_RECORD_LEN];
1651 	char *dspath;
1652 	zfs_cmd_t zc = { 0 };
1653 	int i;
1654 
1655 	/* construct the command string */
1656 	(void) strcpy(cmd_buf, pool ? "zpool" : "zfs");
1657 	for (i = 0; i < argc; i++) {
1658 		if (strlen(cmd_buf) + 1 + strlen(argv[i]) > HIS_MAX_RECORD_LEN)
1659 			break;
1660 		(void) strcat(cmd_buf, " ");
1661 		(void) strcat(cmd_buf, argv[i]);
1662 	}
1663 
1664 	/* figure out the poolname */
1665 	dspath = strpbrk(path, "/@");
1666 	if (dspath == NULL) {
1667 		(void) strcpy(zc.zc_name, path);
1668 	} else {
1669 		(void) strncpy(zc.zc_name, path, dspath - path);
1670 		zc.zc_name[dspath-path] = '\0';
1671 	}
1672 
1673 	zc.zc_history = (uint64_t)(uintptr_t)cmd_buf;
1674 	zc.zc_history_len = strlen(cmd_buf);
1675 
1676 	/* overloading zc_history_offset */
1677 	zc.zc_history_offset = pool_create;
1678 
1679 	(void) ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_LOG_HISTORY, &zc);
1680 }
1681 
1682 /*
1683  * Perform ioctl to get some command history of a pool.
1684  *
1685  * 'buf' is the buffer to fill up to 'len' bytes.  'off' is the
1686  * logical offset of the history buffer to start reading from.
1687  *
1688  * Upon return, 'off' is the next logical offset to read from and
1689  * 'len' is the actual amount of bytes read into 'buf'.
1690  */
1691 static int
1692 get_history(zpool_handle_t *zhp, char *buf, uint64_t *off, uint64_t *len)
1693 {
1694 	zfs_cmd_t zc = { 0 };
1695 	libzfs_handle_t *hdl = zhp->zpool_hdl;
1696 
1697 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1698 
1699 	zc.zc_history = (uint64_t)(uintptr_t)buf;
1700 	zc.zc_history_len = *len;
1701 	zc.zc_history_offset = *off;
1702 
1703 	if (ioctl(hdl->libzfs_fd, ZFS_IOC_POOL_GET_HISTORY, &zc) != 0) {
1704 		switch (errno) {
1705 		case EPERM:
1706 			return (zfs_error_fmt(hdl, EZFS_PERM,
1707 			    dgettext(TEXT_DOMAIN,
1708 			    "cannot show history for pool '%s'"),
1709 			    zhp->zpool_name));
1710 		case ENOENT:
1711 			return (zfs_error_fmt(hdl, EZFS_NOHISTORY,
1712 			    dgettext(TEXT_DOMAIN, "cannot get history for pool "
1713 			    "'%s'"), zhp->zpool_name));
1714 		default:
1715 			return (zpool_standard_error_fmt(hdl, errno,
1716 			    dgettext(TEXT_DOMAIN,
1717 			    "cannot get history for '%s'"), zhp->zpool_name));
1718 		}
1719 	}
1720 
1721 	*len = zc.zc_history_len;
1722 	*off = zc.zc_history_offset;
1723 
1724 	return (0);
1725 }
1726 
1727 /*
1728  * Process the buffer of nvlists, unpacking and storing each nvlist record
1729  * into 'records'.  'leftover' is set to the number of bytes that weren't
1730  * processed as there wasn't a complete record.
1731  */
1732 static int
1733 zpool_history_unpack(char *buf, uint64_t bytes_read, uint64_t *leftover,
1734     nvlist_t ***records, uint_t *numrecords)
1735 {
1736 	uint64_t reclen;
1737 	nvlist_t *nv;
1738 	int i;
1739 
1740 	while (bytes_read > sizeof (reclen)) {
1741 
1742 		/* get length of packed record (stored as little endian) */
1743 		for (i = 0, reclen = 0; i < sizeof (reclen); i++)
1744 			reclen += (uint64_t)(((uchar_t *)buf)[i]) << (8*i);
1745 
1746 		if (bytes_read < sizeof (reclen) + reclen)
1747 			break;
1748 
1749 		/* unpack record */
1750 		if (nvlist_unpack(buf + sizeof (reclen), reclen, &nv, 0) != 0)
1751 			return (ENOMEM);
1752 		bytes_read -= sizeof (reclen) + reclen;
1753 		buf += sizeof (reclen) + reclen;
1754 
1755 		/* add record to nvlist array */
1756 		(*numrecords)++;
1757 		if (ISP2(*numrecords + 1)) {
1758 			*records = realloc(*records,
1759 			    *numrecords * 2 * sizeof (nvlist_t *));
1760 		}
1761 		(*records)[*numrecords - 1] = nv;
1762 	}
1763 
1764 	*leftover = bytes_read;
1765 	return (0);
1766 }
1767 
1768 #define	HIS_BUF_LEN	(128*1024)
1769 
1770 /*
1771  * Retrieve the command history of a pool.
1772  */
1773 int
1774 zpool_get_history(zpool_handle_t *zhp, nvlist_t **nvhisp)
1775 {
1776 	char buf[HIS_BUF_LEN];
1777 	uint64_t off = 0;
1778 	nvlist_t **records = NULL;
1779 	uint_t numrecords = 0;
1780 	int err, i;
1781 
1782 	do {
1783 		uint64_t bytes_read = sizeof (buf);
1784 		uint64_t leftover;
1785 
1786 		if ((err = get_history(zhp, buf, &off, &bytes_read)) != 0)
1787 			break;
1788 
1789 		/* if nothing else was read in, we're at EOF, just return */
1790 		if (!bytes_read)
1791 			break;
1792 
1793 		if ((err = zpool_history_unpack(buf, bytes_read,
1794 		    &leftover, &records, &numrecords)) != 0)
1795 			break;
1796 		off -= leftover;
1797 
1798 		/* CONSTCOND */
1799 	} while (1);
1800 
1801 	if (!err) {
1802 		verify(nvlist_alloc(nvhisp, NV_UNIQUE_NAME, 0) == 0);
1803 		verify(nvlist_add_nvlist_array(*nvhisp, ZPOOL_HIST_RECORD,
1804 		    records, numrecords) == 0);
1805 	}
1806 	for (i = 0; i < numrecords; i++)
1807 		nvlist_free(records[i]);
1808 	free(records);
1809 
1810 	return (err);
1811 }
1812 
1813 void
1814 zpool_obj_to_path(zpool_handle_t *zhp, uint64_t dsobj, uint64_t obj,
1815     char *pathname, size_t len)
1816 {
1817 	zfs_cmd_t zc = { 0 };
1818 	boolean_t mounted = B_FALSE;
1819 	char *mntpnt = NULL;
1820 	char dsname[MAXNAMELEN];
1821 
1822 	if (dsobj == 0) {
1823 		/* special case for the MOS */
1824 		(void) snprintf(pathname, len, "<metadata>:<0x%llx>", obj);
1825 		return;
1826 	}
1827 
1828 	/* get the dataset's name */
1829 	(void) strlcpy(zc.zc_name, zhp->zpool_name, sizeof (zc.zc_name));
1830 	zc.zc_obj = dsobj;
1831 	if (ioctl(zhp->zpool_hdl->libzfs_fd,
1832 	    ZFS_IOC_DSOBJ_TO_DSNAME, &zc) != 0) {
1833 		/* just write out a path of two object numbers */
1834 		(void) snprintf(pathname, len, "<0x%llx>:<0x%llx>",
1835 		    dsobj, obj);
1836 		return;
1837 	}
1838 	(void) strlcpy(dsname, zc.zc_value, sizeof (dsname));
1839 
1840 	/* find out if the dataset is mounted */
1841 	mounted = is_mounted(zhp->zpool_hdl, dsname, &mntpnt);
1842 
1843 	/* get the corrupted object's path */
1844 	(void) strlcpy(zc.zc_name, dsname, sizeof (zc.zc_name));
1845 	zc.zc_obj = obj;
1846 	if (ioctl(zhp->zpool_hdl->libzfs_fd, ZFS_IOC_OBJ_TO_PATH,
1847 	    &zc) == 0) {
1848 		if (mounted) {
1849 			(void) snprintf(pathname, len, "%s%s", mntpnt,
1850 			    zc.zc_value);
1851 		} else {
1852 			(void) snprintf(pathname, len, "%s:%s",
1853 			    dsname, zc.zc_value);
1854 		}
1855 	} else {
1856 		(void) snprintf(pathname, len, "%s:<0x%llx>", dsname, obj);
1857 	}
1858 	free(mntpnt);
1859 }
1860