xref: /titanic_50/usr/src/cmd/fm/dicts/ZFS.po (revision 63602c90af2802f2ab666ab39bb4aea7d02500c3)
1#
2# Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
3# Use is subject to license terms.
4#
5# CDDL HEADER START
6#
7# The contents of this file are subject to the terms of the
8# Common Development and Distribution License (the "License").
9# You may not use this file except in compliance with the License.
10#
11# You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
12# or http://www.opensolaris.org/os/licensing.
13# See the License for the specific language governing permissions
14# and limitations under the License.
15#
16# When distributing Covered Code, include this CDDL HEADER in each
17# file and include the License file at usr/src/OPENSOLARIS.LICENSE.
18# If applicable, add the following below this CDDL HEADER, with the
19# fields enclosed by brackets "[]" replaced with your own identifying
20# information: Portions Copyright [yyyy] [name of copyright owner]
21#
22# CDDL HEADER END
23#
24#ident	"%Z%%M%	%I%	%E% SMI"
25#
26# DO NOT EDIT -- this file is generated by the Event Registry.
27#
28#
29# code: ZFS-8000-14
30# keys: ereport.fs.zfs.pool.corrupt_cache
31#
32msgid "ZFS-8000-14.type"
33msgstr "Error"
34msgid "ZFS-8000-14.severity"
35msgstr "Critical"
36msgid "ZFS-8000-14.description"
37msgstr "The ZFS cache file is corrupted  Refer to %s for more information."
38msgid "ZFS-8000-14.response"
39msgstr "No automated response will be taken."
40msgid "ZFS-8000-14.impact"
41msgstr "ZFS filesystems are not available"
42msgid "ZFS-8000-14.action"
43msgstr "\nZFS keeps a list of active pools on the filesystem to avoid having to\nscan all devices when the system is booted.  If this file is corrupted, then\nnormally active pools will not be automatically opened.  The pools can be\nrecovered using the 'zpool import' command:\n\n\n# zpool import\n  pool: test\n    id: 12743384782310107047\n state: ONLINE\naction: The pool can be imported using its name or numeric identifier.\nconfig:\n\n        test              ONLINE\n          c0t0d0          ONLINE\n\n\nThis will automatically scan /dev/dsk for any\ndevices part of a pool.  If devices have been made available in an alternate\nlocation, use the '-d' option to 'zpool import' to search for devices in a\ndifferent directory.\n\nOnce you have determined which pools are available for import, you can\nimport the pool explicitly by specifying the name or numeric\nidentifier:\n\n\n# zpool import test\n\n\nAlternately, you can import all available pools by specifying the\n'-a' option.  Once a pool has been imported, the ZFS cache will be repaired so\nthat the pool will appear normally in the future.\n	"
44#
45# code: ZFS-8000-2Q
46# keys: ereport.fs.zfs.device.missing_r
47#
48msgid "ZFS-8000-2Q.type"
49msgstr "Error"
50msgid "ZFS-8000-2Q.severity"
51msgstr "Major"
52msgid "ZFS-8000-2Q.description"
53msgstr "A device in a replicated configuration could not be\n	    opened.  Refer to %s for more information."
54msgid "ZFS-8000-2Q.response"
55msgstr "A hot spare will be activated if available."
56msgid "ZFS-8000-2Q.impact"
57msgstr "The pool is no longer providing the configured level of\n	    replication."
58msgid "ZFS-8000-2Q.action"
59msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool has\nexperienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices could not be opened.  Sufficient replicas exist for\n        the pool to continue functioning in a degraded state.\naction: Attach the missing device and online it using 'zpool online'.\n   see: http://www.sun.com/msg/ZFS-8000-2Q\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  DEGRADED     0     0     0\n          mirror              DEGRADED     0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            FAULTED      0     0     0  cannot open\n\nerrors: No known data errors\n\n\nDetermine which device failed to open by looking for a FAULTED device\nwith an additional 'cannot open' message.  If this device has been inadvertently\nremoved from the system, attach the device and bring it online with 'zpool\nonline':\n\n\n# zpool online test c0t0d1\n\n\nIf the device is no longer available, the device can be replaced using\nthe 'zpool replace' command:\n\n\n# zpool replace test c0t0d1 c0t0d2\n\n\nIf the device has been replaced by another disk in the same physical\nslot, then the device can be replaced using a single argument to the 'zpool\nreplace' command:\n\n\n# zpool replace test c0t0d1\n\n\nExisting data will be resilvered to the new device.  Once the\nresilvering completes, the device will be removed from the pool.\n\nFor an exported pool\n\nIf this error is encountered during a 'zpool import', it means that one\nof the devices is not attached to the system:\n\n\n# zpool import\n  pool: test\n    id: 10121266328238932306\n state: DEGRADED\nstatus: One or more devices are missing from the system.\naction: The pool can be imported despite missing or damaged devices.  The\n        fault tolerance of the pool may be compromised if imported.\n   see: http://www.sun.com/msg/ZFS-8000-2Q\nconfig:\n\n        test              DEGRADED\n          mirror          DEGRADED\n            c0t0d0        ONLINE\n            c0t0d1        FAULTED   cannot open\n\n\nUnlike when the pool is active on the system, the device cannot be\nreplaced while the pool is exported.  If the device can be attached to the\nsystem, attach the device and run 'zpool import' again.\n\nAlternatively, the pool can be imported as-is, though it will be placed\nin the DEGRADED state due to a missing device.  The device will be marked as\nUNAVAIL.  Once the pool has been imported, the missing device can be replaced as\ndescribed above.\n	"
60#
61# code: ZFS-8000-3C
62# keys: ereport.fs.zfs.device.missing_nr
63#
64msgid "ZFS-8000-3C.type"
65msgstr "Error"
66msgid "ZFS-8000-3C.severity"
67msgstr "Critical"
68msgid "ZFS-8000-3C.description"
69msgstr "A device could not be opened and no replicas are available.  Refer to %s for more information."
70msgid "ZFS-8000-3C.response"
71msgstr "No automated response will be taken."
72msgid "ZFS-8000-3C.impact"
73msgstr "The pool is no longer available"
74msgid "ZFS-8000-3C.action"
75msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: One or more devices could not be opened.  There are insufficient\n	replicas for the pool to continue functioning.\naction: Attach the missing device and online it using 'zpool online'.\n   see: http://www.sun.com/msg/ZFS-8000-3C\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  insufficient replicas\n          c0t0d0              ONLINE       0     0     0\n          c0t0d1              FAULTED      0     0     0  cannot open\n\nerrors: No known data errors\n\n\nIf the device has been temporarily detached from the system, attach the\ndevice to the system and run 'zpool status' again.  The pool should\nautomatically detect the newly attached device and resume functioning.  You may\nhave to mount the filesystems in the pool explicitly using 'zfs\nmount -a'.\n\nIf the device is no longer available and cannot be reattached to the\nsystem, then the pool must be destroyed and re-created from a backup\nsource.\n\nFor an exported pool\n\nIf this error is encountered during a 'zpool import', it means that one\nof the devices is not attached to the system:\n\n\n# zpool import\n  pool: test\n    id: 10121266328238932306\n state: FAULTED\nstatus: One or more devices are missing from the system.\naction: The pool cannot be imported.  Attach the missing devices and\n	try again.\n   see: http://www.sun.com/msg/ZFS-8000-3C\nconfig:\n\n        test              FAULTED   insufficient replicas\n          c0t0d0          ONLINE\n          c0t0d1          FAULTED   cannot open\n\n\nThe pool cannot be imported until the missing device is attached to the\nsystem.  If the device has been made available in an alternate location, use the\n'-d' option to 'zpool import' to search for devices in a different directory.\nIf the missing device is unavailable, then the pool cannot be imported.\n	"
76#
77# code: ZFS-8000-4J
78# keys: ereport.fs.zfs.device.corrupt_label_r
79#
80msgid "ZFS-8000-4J.type"
81msgstr "Error"
82msgid "ZFS-8000-4J.severity"
83msgstr "Major"
84msgid "ZFS-8000-4J.description"
85msgstr "A device could not be opened due to a missing or invalid\n	    device label.  Refer to %s for more information."
86msgid "ZFS-8000-4J.response"
87msgstr "A hot spare will be activated if available."
88msgid "ZFS-8000-4J.impact"
89msgstr "The pool is no longer providing the configured level of\n	    replication."
90msgid "ZFS-8000-4J.action"
91msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices could not be used because the label is missing or\n        invalid.  Sufficient replicas exist for the pool to continue\n        functioning in a degraded state.\naction: Replace the device using 'zpool replace'.\n   see: http://www.sun.com/msg/ZFS-8000-4J\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  DEGRADED     0     0     0\n          mirror              DEGRADED     0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            FAULTED      0     0     0  corrupted data\n\nerrors: No known data errors\n\n\nIf the device has been temporarily detached from the system, attach the\ndevice to the system and run 'zpool status' again.  The pool should\nautomatically detect the newly attached device and resume functioning.\n\nIf the device is no longer available, it can be replaced using 'zpool\nreplace':\n\n\n# zpool replace test c0t0d1 c0t0d2\n\n\nIf the device has been replaced by another disk in the same physical\nslot, then the device can be replaced using a single argument to the 'zpool\nreplace' command:\n\n\n# zpool replace test c0t0d1\n\n\nZFS will begin migrating data to the new device as soon as the replace\nis issued.  Once the resilvering completes, the original device (if different\nfrom the replacement) will be removed, and the pool will be restored to the\nONLINE state.\n\nFor an exported pool\n\nIf this error is encountered while running 'zpool import', the pool can\nbe still be imported despite the failure:\n\n\n# zpool import\n  pool: test\n    id: 5187963178597328409\n state: DEGRADED\nstatus: One or more devices contains corrupted data.  The fault tolerance of\n	the pool may be compromised if imported.\naction: The pool can be imported using its name or numeric identifier.\n   see: http://www.sun.com/msg/ZFS-8000-4J\nconfig:\n\n        test              DEGRADED\n          mirror          DEGRADED\n            c0t0d0        ONLINE\n            c0t0d1        FAULTED   corrupted data\n\n\nTo import the pool, run 'zpool import':\n\n\n# zpool import test\n\n\nOnce the pool has been imported, the damaged device can be replaced\naccording to the above procedure.\n	"
92#
93# code: ZFS-8000-5E
94# keys: ereport.fs.zfs.device.corrupt_label_nr
95#
96msgid "ZFS-8000-5E.type"
97msgstr "Error"
98msgid "ZFS-8000-5E.severity"
99msgstr "Critical"
100msgid "ZFS-8000-5E.description"
101msgstr "A device could not be opened due to a missing or invalid\n	    device label and no replicas are available.  Refer to %s for more information."
102msgid "ZFS-8000-5E.response"
103msgstr "No automated response will be taken."
104msgid "ZFS-8000-5E.impact"
105msgstr "The pool is no longer available"
106msgid "ZFS-8000-5E.action"
107msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: One or more devices could not be used because the the label is missing \n        or invalid.  There are insufficient replicas for the pool to continue\n        functioning.\naction: Destroy and re-create the pool from a backup source.\n   see: http://www.sun.com/msg/ZFS-8000-5E\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  insufficient replicas\n          c0t0d0              FAULTED      0     0     0  corrupted data\n          c0t0d1              ONLINE       0     0     0\n\nerrors: No known data errors\n\n\nThe device listed as FAULTED with 'corrupted data' cannot be opened due\nto a corrupt label.  ZFS will be unable to use the pool, and all data within the\npool is irrevocably lost.  The pool must be destroyed and recreated from an\nappropriate backup source.  Using replicated configurations will prevent this\nfrom happening in the future.\n\nFor an exported pool\n\nIf this error is encountered during 'zpool import', the action is the\nsame.  The pool cannot be imported - all data is lost and must be restored from\nan appropriate backup source.\n	"
108#
109# code: ZFS-8000-6X
110# keys: ereport.fs.zfs.pool.bad_guid_sum
111#
112msgid "ZFS-8000-6X.type"
113msgstr "Error"
114msgid "ZFS-8000-6X.severity"
115msgstr "Critical"
116msgid "ZFS-8000-6X.description"
117msgstr "One or more top level devices are missing.  Refer to %s for more information."
118msgid "ZFS-8000-6X.response"
119msgstr "No automated response will be taken."
120msgid "ZFS-8000-6X.impact"
121msgstr "The pool cannot be imported"
122msgid "ZFS-8000-6X.action"
123msgstr "\nRun 'zpool import' to list which pool cannot be imported:\n\n\n# zpool import\n  pool: test\n    id: 13783646421373024673\n state: FAULTED\nstatus: One or more devices are missing from the system.\naction: The pool cannot be imported.  Attach the missing\n	devices and try again.\n   see: http://www.sun.com/msg/ZFS-8000-6X\nconfig:\n\n        test              FAULTED   missing device\n          c0t0d0          ONLINE\n\n        Additional devices are known to be part of this pool, though their\n        exact configuration cannot be determined.\n\n\nZFS attempts to store enough configuration data on the devices such\nthat the configuration is recoverable from any subset of devices.  In some\ncases, particularly when an entire toplevel virtual device is not attached to\nthe system, ZFS will be unable to determine the complete configuration.  It will\nalways detect that these devices are missing, even if it cannot identify all of\nthe devices.\n\nThe pool cannot be imported until the unknown missing device is\nattached to the system.  If the device has been made available in an alternate\nlocation, use the '-d' option to 'zpool import' to search for devices in a\ndifferent directory.  If the missing device is unavailable, then the pool cannot\nbe imported.\n	"
124#
125# code: ZFS-8000-72
126# keys: ereport.fs.zfs.pool.corrupt_pool
127#
128msgid "ZFS-8000-72.type"
129msgstr "Error"
130msgid "ZFS-8000-72.severity"
131msgstr "Critical"
132msgid "ZFS-8000-72.description"
133msgstr "The metadata required to open the pool is corrupt.  Refer to %s for more information."
134msgid "ZFS-8000-72.response"
135msgstr "No automated response will be taken."
136msgid "ZFS-8000-72.impact"
137msgstr "The pool is no longer available"
138msgid "ZFS-8000-72.action"
139msgstr "\nIf this error is encountered during 'zpool import', see the section\nbelow.  Otherwise, run 'zpool status -x' to determine which pool is\nfaulted:\n\n\n# zpool status -x\n# zpool import\n  pool: test\n    id: 13783646421373024673\n state: FAULTED\nstatus: The pool metadata is corrupted and cannot be opened.\naction: Destroy the pool and restore from backup.\n   see: http://www.sun.com/msg/ZFS-8000-72\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     2  corrupted data\n          mirror              DEGRADED     0     0     2\n            c0t0d0            ONLINE       0     0     2\n            c0t0d1            ONLINE       0     0     2\n\nerrors: No known errors\n\n\nEven though all the devices are available, the on-disk data has been\ncorrupted such that the pool cannot be opened.  All data within the pool is\nlost, and the pool must be destroyed and restored from an appropriate backup\nsource.  ZFS includes built-in metadata replication to prevent this from\nhappening even for unreplicated pools, but running in a replicated configuration\nwill decrease the chances of this happening in the future.\n\nIf this error is encountered during 'zpool import', the pool is\nunrecoverable and cannot be imported.  The pool must be restored from an\nappropriate backup source.\n	"
140#
141# code: ZFS-8000-8A
142# keys: ereport.fs.zfs.object.corrupt_data
143#
144msgid "ZFS-8000-8A.type"
145msgstr "Error"
146msgid "ZFS-8000-8A.severity"
147msgstr "Critical"
148msgid "ZFS-8000-8A.description"
149msgstr "A file or directory could not be read due to corrupt data.  Refer to %s for more information."
150msgid "ZFS-8000-8A.response"
151msgstr "No automated response will be taken."
152msgid "ZFS-8000-8A.impact"
153msgstr "The file or directory is unavailable."
154msgid "ZFS-8000-8A.action"
155msgstr "\nRun 'zpool status -x' to determine which pool is damaged:\n\n\n# zpool status -x\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://www.sun.com/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: 1 data errors, use '-v' for a list\n\n\nUnfortunately, the data cannot be repaired, and the only choice to\nrepair the data is to restore the pool from backup.  Applications attempting to\naccess the corrupted data will get an error (EIO), and data may be permanently\nlost.\n\nOn recent versions of Solaris, the list of affected files can be\nretrieved by using the '-v' option to 'zpool status':\n\n\n# zpool status -xv\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://www.sun.com/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: Permanent errors have been detected in the following files:\n\n        /export/example/foo\n\n\nDamaged files may or may not be able to be removed depending on the\ntype of corruption.  If the corruption is within the plain data, the file should\nbe removable.  If the corruption is in the file metadata, then the file cannot\nbe removed, though it can be moved to an alternate location.  In either case,\nthe data should be restored from a backup source.  It is also possible for the\ncorruption to be within pool-wide metadata, resulting in entire datasets being\nunavailable.  If this is the case, the only option is to destroy the pool and\nre-create the datasets from backup.\n	"
156#
157# code: ZFS-8000-9P
158# keys: ereport.fs.zfs.device.failing
159#
160msgid "ZFS-8000-9P.type"
161msgstr "Error"
162msgid "ZFS-8000-9P.severity"
163msgstr "Minor"
164msgid "ZFS-8000-9P.description"
165msgstr "A device has experienced uncorrectable errors in a\n	    replicated configuration.  Refer to %s for more information."
166msgid "ZFS-8000-9P.response"
167msgstr "ZFS has attempted to repair the affected data."
168msgid "ZFS-8000-9P.impact"
169msgstr "The system is unaffected, though errors may indicate future\n	    failure.  Future errors may cause ZFS to automatically fault\n	    the device."
170msgid "ZFS-8000-9P.action"
171msgstr "\nRun 'zpool status -x' to determine which pool has experienced\nerrors:\n\n\n# zpool status\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an unrecoverable error.  An\n        attempt was made to correct the error.  Applications are unaffected.\naction: Determine if the device needs to be replaced, and clear the errors\n        using 'zpool online' or replace the device with 'zpool replace'.\n   see: http://www.sun.com/msg/ZFS-8000-9P\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     0\n          mirror              ONLINE       0     0     0\n            c0t0d0            ONLINE       0     0     2\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known data errors\n\n\nFind the device with a non-zero error count for READ, WRITE, or CKSUM.\nThis indicates that the device has experienced a read I/O error, write I/O\nerror, or checksum validation error.  Because the device is part of a mirror or\nRAID-Z device, ZFS was able to recover from the error and subsequently repair\nthe damaged data.\n\nIf these errors persist over a period of time, ZFS may determine the\ndevice is faulty and mark it as such.  However, these error counts may or may\nnot indicate that the device is unusable.  It depends on how the errors were\ncaused, which the administrator can determine in advance of any ZFS diagnosis.\nFor example, the following cases will all produce errors that do not indicate\npotential device failure:\n\n\nA network attached device lost connectivity but has now\nrecovered\nA device suffered from a bit flip, an expected event over long\nperiods of time\nAn administrator accidentally wrote over a portion of the disk using\nanother program\n\n\nIn these cases, the presence of errors does not indicate that the\ndevice is likely to fail in the future, and therefore does not need to be\nreplaced.  If this is the case, then the device errors should be cleared using\n'zpool clear':\n\n\n# zpool clear test c0t0d0\n\n\nOn the other hand, errors may very well indicate that the device has\nfailed or is about to fail.  If there are continual I/O errors to a device that\nis otherwise attached and functioning on the system, it most likely needs to be\nreplaced.   The administrator should check the system log for any driver\nmessages that may indicate hardware failure.  If it is determined that the\ndevice needs to be replaced, then the 'zpool replace' command should be\nused:\n\n\n# zpool replace test c0t0d0 c0t0d2\n\n\nThis will attach the new device to the pool and begin resilvering data\nto it.  Once the resilvering process is complete, the old device will\nautomatically be removed from the pool, at which point it can safely be removed\nfrom the system.  If the device needs to be replaced in-place (because there are\nno available spare devices), the original device can be removed and replaced\nwith a new device, at which point a different form of 'zpool replace' can be\nused:\n\n\n# zpool replace test c0t0d0\n\n\nThis assumes that the original device at 'c0t0d0' has been replaced\nwith a new device under the same path, and will be replaced\nappropriately.\n\nYou can monitor the progress of the resilvering operation by using the\n'zpool status -x' command:\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices is currently being replaced.  The pool may not be\n	providing the necessary level of replication.\naction: Wait for the resilvering operation to complete\n scrub: resilver in progress, 0.14% done, 0h0m to go\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     0\n          mirror              ONLINE       0     0     0\n            replacing         ONLINE       0     0     0\n              c0t0d0          ONLINE       0     0     3\n              c0t0d2          ONLINE       0     0     0  58.5K resilvered\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known data errors\n\n	"
172#
173# code: ZFS-8000-A5
174# keys: ereport.fs.zfs.device.version_mismatch
175#
176msgid "ZFS-8000-A5.type"
177msgstr "Error"
178msgid "ZFS-8000-A5.severity"
179msgstr "Major"
180msgid "ZFS-8000-A5.description"
181msgstr "The on-disk version is not compatible with the running\n	    system.  Refer to %s for more information."
182msgid "ZFS-8000-A5.response"
183msgstr "No automated response will occur."
184msgid "ZFS-8000-A5.impact"
185msgstr "The pool is unavailable."
186msgid "ZFS-8000-A5.action"
187msgstr "\nIf this error is seen during 'zpool import', see the section below.\nOtherwise, run 'zpool status -x' to determine which pool is faulted:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: The ZFS version for the pool is incompatible with the software running\n        on this system.\naction: Destroy and re-create the pool.\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  incompatible version\n          mirror              ONLINE       0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known errors\n\n\nThe pool cannot be used on this system.  Either move the storage to the\nsystem where the pool was originally created, upgrade the current system\nsoftware to a more recent version, or destroy the pool and re-create it from\nbackup.\n\nIf this error is seen during import, the pool cannot be imported on the\ncurrent system.  The disks must be attached to the system which originally\ncreated the pool, and imported there.\n\nThe list of currently supported versions can be displayed using 'zpool\nupgrade -v'.\n	"
188#
189# code: ZFS-8000-CS
190# keys: fault.fs.zfs.pool
191#
192msgid "ZFS-8000-CS.type"
193msgstr "Fault"
194msgid "ZFS-8000-CS.severity"
195msgstr "Major"
196msgid "ZFS-8000-CS.description"
197msgstr "A ZFS pool failed to open.  Refer to %s for more information."
198msgid "ZFS-8000-CS.response"
199msgstr "No automated response will occur."
200msgid "ZFS-8000-CS.impact"
201msgstr "The pool data is unavailable"
202msgid "ZFS-8000-CS.action"
203msgstr "Run 'zpool status -x' and either attach the missing device or\n	    restore from backup."
204#
205# code: ZFS-8000-D3
206# keys: fault.fs.zfs.device
207#
208msgid "ZFS-8000-D3.type"
209msgstr "Fault"
210msgid "ZFS-8000-D3.severity"
211msgstr "Major"
212msgid "ZFS-8000-D3.description"
213msgstr "A ZFS device failed.  Refer to %s for more information."
214msgid "ZFS-8000-D3.response"
215msgstr "No automated response will occur."
216msgid "ZFS-8000-D3.impact"
217msgstr "Fault tolerance of the pool may be compromised."
218msgid "ZFS-8000-D3.action"
219msgstr "Run 'zpool status -x' and replace the bad device."
220#
221# code: ZFS-8000-EY
222# keys: ereport.fs.zfs.pool.hostname_mismatch
223#
224msgid "ZFS-8000-EY.type"
225msgstr "Error"
226msgid "ZFS-8000-EY.severity"
227msgstr "Major"
228msgid "ZFS-8000-EY.description"
229msgstr "The ZFS pool was last accessed by another system  Refer to %s for more information."
230msgid "ZFS-8000-EY.response"
231msgstr "No automated response will be taken."
232msgid "ZFS-8000-EY.impact"
233msgstr "ZFS filesystems are not available"
234msgid "ZFS-8000-EY.action"
235msgstr "\n\nThe pool has been written to from another host, and was not cleanly exported\nfrom the other system.  Actively importing a pool on multiple systems will\ncorrupt the pool and leave it in an unrecoverable state.  To determine which\nsystem last accessed the pool, run the 'zpool import' command:\n\n\n# zpool import\n  pool: test\n    id: 14702934086626715962\nstate:  ONLINE\nstatus: The pool was last accessed by another system.\naction: The pool can be imported using its name or numeric identifier and\n        the '-f' flag.\n   see: http://www.sun.com/msg/ZFS-8000-EY\nconfig:\n\n        test              ONLINE\n          c0t0d0          ONLINE\n\n# zpool import test\ncannot import 'test': pool may be in use from other system, it was last\naccessed by 'tank' (hostid: 0x1435718c) on Fri Mar  9 15:42:47 2007\nuse '-f' to import anyway\n\n\n\nIf you are certain that the pool is not being actively accessed by another\nsystem, then you can use the '-f' option to 'zpool import' to forcibly\nimport the pool.\n\n	"
236#
237# code: ZFS-8000-FD
238# keys: fault.fs.zfs.vdev.io
239#
240msgid "ZFS-8000-FD.type"
241msgstr "Fault"
242msgid "ZFS-8000-FD.severity"
243msgstr "Major"
244msgid "ZFS-8000-FD.description"
245msgstr "The number of I/O errors associated with a ZFS device exceeded\n	     acceptable levels.  Refer to %s for more information."
246msgid "ZFS-8000-FD.response"
247msgstr "The device has been offlined and marked as faulted.  An attempt\n	     will be made to activate a hot spare if available. "
248msgid "ZFS-8000-FD.impact"
249msgstr "Fault tolerance of the pool may be compromised."
250msgid "ZFS-8000-FD.action"
251msgstr "Run 'zpool status -x' and replace the bad device."
252#
253# code: ZFS-8000-GH
254# keys: fault.fs.zfs.vdev.checksum
255#
256msgid "ZFS-8000-GH.type"
257msgstr "Fault"
258msgid "ZFS-8000-GH.severity"
259msgstr "Major"
260msgid "ZFS-8000-GH.description"
261msgstr "The number of checksum errors associated with a ZFS device\nexceeded acceptable levels.  Refer to %s for more information."
262msgid "ZFS-8000-GH.response"
263msgstr "The device has been marked as degraded.  An attempt\nwill be made to activate a hot spare if available."
264msgid "ZFS-8000-GH.impact"
265msgstr "Fault tolerance of the pool may be compromised."
266msgid "ZFS-8000-GH.action"
267msgstr "Run 'zpool status -x' and replace the bad device."
268