xref: /titanic_44/usr/src/lib/libzfs/common/libzfs_status.c (revision 87308b749adb04ab3f564c765b089962af3edb4d)
1 /*
2  * CDDL HEADER START
3  *
4  * The contents of this file are subject to the terms of the
5  * Common Development and Distribution License (the "License").
6  * You may not use this file except in compliance with the License.
7  *
8  * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
9  * or http://www.opensolaris.org/os/licensing.
10  * See the License for the specific language governing permissions
11  * and limitations under the License.
12  *
13  * When distributing Covered Code, include this CDDL HEADER in each
14  * file and include the License file at usr/src/OPENSOLARIS.LICENSE.
15  * If applicable, add the following below this CDDL HEADER, with the
16  * fields enclosed by brackets "[]" replaced with your own identifying
17  * information: Portions Copyright [yyyy] [name of copyright owner]
18  *
19  * CDDL HEADER END
20  */
21 
22 /*
23  * Copyright (c) 2005, 2010, Oracle and/or its affiliates. All rights reserved.
24  * Copyright (c) 2012 by Delphix. All rights reserved.
25  * Copyright (c) 2013 Steven Hartland. All rights reserved.
26  */
27 
28 /*
29  * This file contains the functions which analyze the status of a pool.  This
30  * include both the status of an active pool, as well as the status exported
31  * pools.  Returns one of the ZPOOL_STATUS_* defines describing the status of
32  * the pool.  This status is independent (to a certain degree) from the state of
33  * the pool.  A pool's state describes only whether or not it is capable of
34  * providing the necessary fault tolerance for data.  The status describes the
35  * overall status of devices.  A pool that is online can still have a device
36  * that is experiencing errors.
37  *
38  * Only a subset of the possible faults can be detected using 'zpool status',
39  * and not all possible errors correspond to a FMA message ID.  The explanation
40  * is left up to the caller, depending on whether it is a live pool or an
41  * import.
42  */
43 
44 #include <libzfs.h>
45 #include <string.h>
46 #include <unistd.h>
47 #include "libzfs_impl.h"
48 #include "zfeature_common.h"
49 
50 /*
51  * Message ID table.  This must be kept in sync with the ZPOOL_STATUS_* defines
52  * in libzfs.h.  Note that there are some status results which go past the end
53  * of this table, and hence have no associated message ID.
54  */
55 static char *zfs_msgid_table[] = {
56 	"ZFS-8000-14",
57 	"ZFS-8000-2Q",
58 	"ZFS-8000-3C",
59 	"ZFS-8000-4J",
60 	"ZFS-8000-5E",
61 	"ZFS-8000-6X",
62 	"ZFS-8000-72",
63 	"ZFS-8000-8A",
64 	"ZFS-8000-9P",
65 	"ZFS-8000-A5",
66 	"ZFS-8000-EY",
67 	"ZFS-8000-HC",
68 	"ZFS-8000-JQ",
69 	"ZFS-8000-K4",
70 };
71 
72 #define	NMSGID	(sizeof (zfs_msgid_table) / sizeof (zfs_msgid_table[0]))
73 
74 /* ARGSUSED */
75 static int
76 vdev_missing(uint64_t state, uint64_t aux, uint64_t errs)
77 {
78 	return (state == VDEV_STATE_CANT_OPEN &&
79 	    aux == VDEV_AUX_OPEN_FAILED);
80 }
81 
82 /* ARGSUSED */
83 static int
84 vdev_faulted(uint64_t state, uint64_t aux, uint64_t errs)
85 {
86 	return (state == VDEV_STATE_FAULTED);
87 }
88 
89 /* ARGSUSED */
90 static int
91 vdev_errors(uint64_t state, uint64_t aux, uint64_t errs)
92 {
93 	return (state == VDEV_STATE_DEGRADED || errs != 0);
94 }
95 
96 /* ARGSUSED */
97 static int
98 vdev_broken(uint64_t state, uint64_t aux, uint64_t errs)
99 {
100 	return (state == VDEV_STATE_CANT_OPEN);
101 }
102 
103 /* ARGSUSED */
104 static int
105 vdev_offlined(uint64_t state, uint64_t aux, uint64_t errs)
106 {
107 	return (state == VDEV_STATE_OFFLINE);
108 }
109 
110 /* ARGSUSED */
111 static int
112 vdev_removed(uint64_t state, uint64_t aux, uint64_t errs)
113 {
114 	return (state == VDEV_STATE_REMOVED);
115 }
116 
117 /*
118  * Detect if any leaf devices that have seen errors or could not be opened.
119  */
120 static boolean_t
121 find_vdev_problem(nvlist_t *vdev, int (*func)(uint64_t, uint64_t, uint64_t))
122 {
123 	nvlist_t **child;
124 	vdev_stat_t *vs;
125 	uint_t c, children;
126 	char *type;
127 
128 	/*
129 	 * Ignore problems within a 'replacing' vdev, since we're presumably in
130 	 * the process of repairing any such errors, and don't want to call them
131 	 * out again.  We'll pick up the fact that a resilver is happening
132 	 * later.
133 	 */
134 	verify(nvlist_lookup_string(vdev, ZPOOL_CONFIG_TYPE, &type) == 0);
135 	if (strcmp(type, VDEV_TYPE_REPLACING) == 0)
136 		return (B_FALSE);
137 
138 	if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_CHILDREN, &child,
139 	    &children) == 0) {
140 		for (c = 0; c < children; c++)
141 			if (find_vdev_problem(child[c], func))
142 				return (B_TRUE);
143 	} else {
144 		verify(nvlist_lookup_uint64_array(vdev, ZPOOL_CONFIG_VDEV_STATS,
145 		    (uint64_t **)&vs, &c) == 0);
146 
147 		if (func(vs->vs_state, vs->vs_aux,
148 		    vs->vs_read_errors +
149 		    vs->vs_write_errors +
150 		    vs->vs_checksum_errors))
151 			return (B_TRUE);
152 	}
153 
154 	/*
155 	 * Check any L2 cache devs
156 	 */
157 	if (nvlist_lookup_nvlist_array(vdev, ZPOOL_CONFIG_L2CACHE, &child,
158 	    &children) == 0) {
159 		for (c = 0; c < children; c++)
160 			if (find_vdev_problem(child[c], func))
161 				return (B_TRUE);
162 	}
163 
164 	return (B_FALSE);
165 }
166 
167 /*
168  * Active pool health status.
169  *
170  * To determine the status for a pool, we make several passes over the config,
171  * picking the most egregious error we find.  In order of importance, we do the
172  * following:
173  *
174  *	- Check for a complete and valid configuration
175  *	- Look for any faulted or missing devices in a non-replicated config
176  *	- Check for any data errors
177  *	- Check for any faulted or missing devices in a replicated config
178  *	- Look for any devices showing errors
179  *	- Check for any resilvering devices
180  *
181  * There can obviously be multiple errors within a single pool, so this routine
182  * only picks the most damaging of all the current errors to report.
183  */
184 static zpool_status_t
185 check_status(nvlist_t *config, boolean_t isimport)
186 {
187 	nvlist_t *nvroot;
188 	vdev_stat_t *vs;
189 	pool_scan_stat_t *ps = NULL;
190 	uint_t vsc, psc;
191 	uint64_t nerr;
192 	uint64_t version;
193 	uint64_t stateval;
194 	uint64_t suspended;
195 	uint64_t hostid = 0;
196 
197 	verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_VERSION,
198 	    &version) == 0);
199 	verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_VDEV_TREE,
200 	    &nvroot) == 0);
201 	verify(nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_VDEV_STATS,
202 	    (uint64_t **)&vs, &vsc) == 0);
203 	verify(nvlist_lookup_uint64(config, ZPOOL_CONFIG_POOL_STATE,
204 	    &stateval) == 0);
205 
206 	/*
207 	 * Currently resilvering a vdev
208 	 */
209 	(void) nvlist_lookup_uint64_array(nvroot, ZPOOL_CONFIG_SCAN_STATS,
210 	    (uint64_t **)&ps, &psc);
211 	if (ps && ps->pss_func == POOL_SCAN_RESILVER &&
212 	    ps->pss_state == DSS_SCANNING)
213 		return (ZPOOL_STATUS_RESILVERING);
214 
215 	/*
216 	 * Pool last accessed by another system.
217 	 */
218 	(void) nvlist_lookup_uint64(config, ZPOOL_CONFIG_HOSTID, &hostid);
219 	if (hostid != 0 && (unsigned long)hostid != gethostid() &&
220 	    stateval == POOL_STATE_ACTIVE)
221 		return (ZPOOL_STATUS_HOSTID_MISMATCH);
222 
223 	/*
224 	 * Newer on-disk version.
225 	 */
226 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
227 	    vs->vs_aux == VDEV_AUX_VERSION_NEWER)
228 		return (ZPOOL_STATUS_VERSION_NEWER);
229 
230 	/*
231 	 * Unsupported feature(s).
232 	 */
233 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
234 	    vs->vs_aux == VDEV_AUX_UNSUP_FEAT) {
235 		nvlist_t *nvinfo;
236 
237 		verify(nvlist_lookup_nvlist(config, ZPOOL_CONFIG_LOAD_INFO,
238 		    &nvinfo) == 0);
239 		if (nvlist_exists(nvinfo, ZPOOL_CONFIG_CAN_RDONLY))
240 			return (ZPOOL_STATUS_UNSUP_FEAT_WRITE);
241 		return (ZPOOL_STATUS_UNSUP_FEAT_READ);
242 	}
243 
244 	/*
245 	 * Check that the config is complete.
246 	 */
247 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
248 	    vs->vs_aux == VDEV_AUX_BAD_GUID_SUM)
249 		return (ZPOOL_STATUS_BAD_GUID_SUM);
250 
251 	/*
252 	 * Check whether the pool has suspended due to failed I/O.
253 	 */
254 	if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_SUSPENDED,
255 	    &suspended) == 0) {
256 		if (suspended == ZIO_FAILURE_MODE_CONTINUE)
257 			return (ZPOOL_STATUS_IO_FAILURE_CONTINUE);
258 		return (ZPOOL_STATUS_IO_FAILURE_WAIT);
259 	}
260 
261 	/*
262 	 * Could not read a log.
263 	 */
264 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
265 	    vs->vs_aux == VDEV_AUX_BAD_LOG) {
266 		return (ZPOOL_STATUS_BAD_LOG);
267 	}
268 
269 	/*
270 	 * Bad devices in non-replicated config.
271 	 */
272 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
273 	    find_vdev_problem(nvroot, vdev_faulted))
274 		return (ZPOOL_STATUS_FAULTED_DEV_NR);
275 
276 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
277 	    find_vdev_problem(nvroot, vdev_missing))
278 		return (ZPOOL_STATUS_MISSING_DEV_NR);
279 
280 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
281 	    find_vdev_problem(nvroot, vdev_broken))
282 		return (ZPOOL_STATUS_CORRUPT_LABEL_NR);
283 
284 	/*
285 	 * Corrupted pool metadata
286 	 */
287 	if (vs->vs_state == VDEV_STATE_CANT_OPEN &&
288 	    vs->vs_aux == VDEV_AUX_CORRUPT_DATA)
289 		return (ZPOOL_STATUS_CORRUPT_POOL);
290 
291 	/*
292 	 * Persistent data errors.
293 	 */
294 	if (!isimport) {
295 		if (nvlist_lookup_uint64(config, ZPOOL_CONFIG_ERRCOUNT,
296 		    &nerr) == 0 && nerr != 0)
297 			return (ZPOOL_STATUS_CORRUPT_DATA);
298 	}
299 
300 	/*
301 	 * Missing devices in a replicated config.
302 	 */
303 	if (find_vdev_problem(nvroot, vdev_faulted))
304 		return (ZPOOL_STATUS_FAULTED_DEV_R);
305 	if (find_vdev_problem(nvroot, vdev_missing))
306 		return (ZPOOL_STATUS_MISSING_DEV_R);
307 	if (find_vdev_problem(nvroot, vdev_broken))
308 		return (ZPOOL_STATUS_CORRUPT_LABEL_R);
309 
310 	/*
311 	 * Devices with errors
312 	 */
313 	if (!isimport && find_vdev_problem(nvroot, vdev_errors))
314 		return (ZPOOL_STATUS_FAILING_DEV);
315 
316 	/*
317 	 * Offlined devices
318 	 */
319 	if (find_vdev_problem(nvroot, vdev_offlined))
320 		return (ZPOOL_STATUS_OFFLINE_DEV);
321 
322 	/*
323 	 * Removed device
324 	 */
325 	if (find_vdev_problem(nvroot, vdev_removed))
326 		return (ZPOOL_STATUS_REMOVED_DEV);
327 
328 	/*
329 	 * Outdated, but usable, version
330 	 */
331 	if (SPA_VERSION_IS_SUPPORTED(version) && version != SPA_VERSION)
332 		return (ZPOOL_STATUS_VERSION_OLDER);
333 
334 	/*
335 	 * Usable pool with disabled features
336 	 */
337 	if (version >= SPA_VERSION_FEATURES) {
338 		int i;
339 		nvlist_t *feat;
340 
341 		if (isimport) {
342 			feat = fnvlist_lookup_nvlist(config,
343 			    ZPOOL_CONFIG_LOAD_INFO);
344 			feat = fnvlist_lookup_nvlist(feat,
345 			    ZPOOL_CONFIG_ENABLED_FEAT);
346 		} else {
347 			feat = fnvlist_lookup_nvlist(config,
348 			    ZPOOL_CONFIG_FEATURE_STATS);
349 		}
350 
351 		for (i = 0; i < SPA_FEATURES; i++) {
352 			zfeature_info_t *fi = &spa_feature_table[i];
353 			if (!nvlist_exists(feat, fi->fi_guid))
354 				return (ZPOOL_STATUS_FEAT_DISABLED);
355 		}
356 	}
357 
358 	return (ZPOOL_STATUS_OK);
359 }
360 
361 zpool_status_t
362 zpool_get_status(zpool_handle_t *zhp, char **msgid)
363 {
364 	zpool_status_t ret = check_status(zhp->zpool_config, B_FALSE);
365 
366 	if (ret >= NMSGID)
367 		*msgid = NULL;
368 	else
369 		*msgid = zfs_msgid_table[ret];
370 
371 	return (ret);
372 }
373 
374 zpool_status_t
375 zpool_import_status(nvlist_t *config, char **msgid)
376 {
377 	zpool_status_t ret = check_status(config, B_TRUE);
378 
379 	if (ret >= NMSGID)
380 		*msgid = NULL;
381 	else
382 		*msgid = zfs_msgid_table[ret];
383 
384 	return (ret);
385 }
386 
387 static void
388 dump_ddt_stat(const ddt_stat_t *dds, int h)
389 {
390 	char refcnt[6];
391 	char blocks[6], lsize[6], psize[6], dsize[6];
392 	char ref_blocks[6], ref_lsize[6], ref_psize[6], ref_dsize[6];
393 
394 	if (dds == NULL || dds->dds_blocks == 0)
395 		return;
396 
397 	if (h == -1)
398 		(void) strcpy(refcnt, "Total");
399 	else
400 		zfs_nicenum(1ULL << h, refcnt, sizeof (refcnt));
401 
402 	zfs_nicenum(dds->dds_blocks, blocks, sizeof (blocks));
403 	zfs_nicenum(dds->dds_lsize, lsize, sizeof (lsize));
404 	zfs_nicenum(dds->dds_psize, psize, sizeof (psize));
405 	zfs_nicenum(dds->dds_dsize, dsize, sizeof (dsize));
406 	zfs_nicenum(dds->dds_ref_blocks, ref_blocks, sizeof (ref_blocks));
407 	zfs_nicenum(dds->dds_ref_lsize, ref_lsize, sizeof (ref_lsize));
408 	zfs_nicenum(dds->dds_ref_psize, ref_psize, sizeof (ref_psize));
409 	zfs_nicenum(dds->dds_ref_dsize, ref_dsize, sizeof (ref_dsize));
410 
411 	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
412 	    refcnt,
413 	    blocks, lsize, psize, dsize,
414 	    ref_blocks, ref_lsize, ref_psize, ref_dsize);
415 }
416 
417 /*
418  * Print the DDT histogram and the column totals.
419  */
420 void
421 zpool_dump_ddt(const ddt_stat_t *dds_total, const ddt_histogram_t *ddh)
422 {
423 	int h;
424 
425 	(void) printf("\n");
426 
427 	(void) printf("bucket   "
428 	    "           allocated             "
429 	    "          referenced          \n");
430 	(void) printf("______   "
431 	    "______________________________   "
432 	    "______________________________\n");
433 
434 	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
435 	    "refcnt",
436 	    "blocks", "LSIZE", "PSIZE", "DSIZE",
437 	    "blocks", "LSIZE", "PSIZE", "DSIZE");
438 
439 	(void) printf("%6s   %6s   %5s   %5s   %5s   %6s   %5s   %5s   %5s\n",
440 	    "------",
441 	    "------", "-----", "-----", "-----",
442 	    "------", "-----", "-----", "-----");
443 
444 	for (h = 0; h < 64; h++)
445 		dump_ddt_stat(&ddh->ddh_stat[h], h);
446 
447 	dump_ddt_stat(dds_total, -1);
448 
449 	(void) printf("\n");
450 }
451