Print this page
    
8490 Remove Sun/Solaris references from FMA messages
Reviewed by: Jason King <jason.king@joyent.com>
Reviewed by: Elijah Zupancic <elijah.zupancic@joyent.com>
Reviewed by: Robert Mustacchi <rm@joyent.com>
Reviewed by: Jerry Jelinek <jerry.jelinek@joyent.com>
    
      
        | Split | 
	Close | 
      
      | Expand all | 
      | Collapse all | 
    
    
          --- old/usr/src/cmd/fm/dicts/ZFS.po
          +++ new/usr/src/cmd/fm/dicts/ZFS.po
   1    1  #
   2    2  # Copyright 2009 Sun Microsystems, Inc.  All rights reserved.
   3    3  # Use is subject to license terms.
   4    4  #
   5    5  # CDDL HEADER START
   6    6  #
   7    7  # The contents of this file are subject to the terms of the
   8    8  # Common Development and Distribution License (the "License").
   9    9  # You may not use this file except in compliance with the License.
  10   10  #
  11   11  # You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE
  12   12  # or http://www.opensolaris.org/os/licensing.
  13   13  # See the License for the specific language governing permissions
  14   14  # and limitations under the License.
  15   15  #
  16   16  # When distributing Covered Code, include this CDDL HEADER in each
  17   17  # file and include the License file at usr/src/OPENSOLARIS.LICENSE.
  18   18  # If applicable, add the following below this CDDL HEADER, with the
  19   19  # fields enclosed by brackets "[]" replaced with your own identifying
  20   20  # information: Portions Copyright [yyyy] [name of copyright owner]
  21   21  #
  22   22  # CDDL HEADER END
  23   23  #
  24   24  #
  25   25  # DO NOT EDIT -- this file is generated by the Event Registry.
  26   26  #
  27   27  #
  28   28  # code: ZFS-8000-14
  29   29  # keys: ereport.fs.zfs.pool.corrupt_cache
  30   30  #
  31   31  msgid "ZFS-8000-14.type"
  32   32  msgstr "Error"
  33   33  msgid "ZFS-8000-14.severity"
  34   34  msgstr "Critical"
  35   35  msgid "ZFS-8000-14.description"
  36   36  msgstr "The ZFS cache file is corrupted  Refer to %s for more information."
  37   37  msgid "ZFS-8000-14.response"
  38   38  msgstr "No automated response will be taken."
  39   39  msgid "ZFS-8000-14.impact"
  40   40  msgstr "ZFS filesystems are not available"
  41   41  msgid "ZFS-8000-14.action"
  42   42  msgstr "\nZFS keeps a list of active pools on the filesystem to avoid having to\nscan all devices when the system is booted.  If this file is corrupted, then\nnormally active pools will not be automatically opened.  The pools can be\nrecovered using the 'zpool import' command:\n\n\n# zpool import\n  pool: test\n    id: 12743384782310107047\n state: ONLINE\naction: The pool can be imported using its name or numeric identifier.\nconfig:\n\n        test              ONLINE\n          c0t0d0          ONLINE\n\n\nThis will automatically scan /dev/dsk for any\ndevices part of a pool.  If devices have been made available in an alternate\nlocation, use the '-d' option to 'zpool import' to search for devices in a\ndifferent directory.\n\nOnce you have determined which pools are available for import, you can\nimport the pool explicitly by specifying the name or numeric\nidentifier:\n\n\n# zpool import test\n\n\nAlternately, you can import all available pools by specifying the\n'-a' option.  Once a pool has been imported, the ZFS cache will be repaired so\nthat the pool will appear normally in the future.\n       "
  43   43  #
  44   44  # code: ZFS-8000-2Q
  45   45  # keys: ereport.fs.zfs.device.missing_r
  46   46  #
  47   47  msgid "ZFS-8000-2Q.type"
  48   48  msgstr "Error"
  49   49  msgid "ZFS-8000-2Q.severity"
  50   50  msgstr "Major"
  51   51  msgid "ZFS-8000-2Q.description"
  52   52  msgstr "A device in a replicated configuration could not be\n       opened.  Refer to %s for more information."
  53   53  msgid "ZFS-8000-2Q.response"
  54   54  msgstr "A hot spare will be activated if available."
  55   55  msgid "ZFS-8000-2Q.impact"
  56   56  msgstr "The pool is no longer providing the configured level of\n           replication."
  57   57  msgid "ZFS-8000-2Q.action"
  58   58  msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool has\nexperienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices could not be opened.  Sufficient replicas exist for\n        the pool to continue functioning in a degraded state.\naction: Attach the missing device and online it using 'zpool online'.\n   see: http://illumos.org/msg/ZFS-8000-2Q\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  DEGRADED     0     0     0\n          mirror              DEGRADED     0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            FAULTED      0     0     0  cannot open\n\nerrors: No known data errors\n\n\nDetermine which device failed to open by looking for a FAULTED device\nwith an additional 'cannot open' message.  If this device has been inadvertently\nremoved from the system, attach the device and bring it online with 'zpool\nonline':\n\n\n# zpool online test c0t0d1\n\n\nIf the device is no longer available, the device can be replaced using\nthe 'zpool replace' command:\n\n\n# zpool replace test c0t0d1 c0t0d2\n\n\nIf the device has been replaced by another disk in the same physical\nslot, then the device can be replaced using a single argument to the 'zpool\nreplace' command:\n\n\n# zpool replace test c0t0d1\n\n\nExisting data will be resilvered to the new device.  Once the\nresilvering completes, the device will be removed from the pool.\n\nFor an exported pool\n\nIf this error is encountered during a 'zpool import', it means that one\nof the devices is not attached to the system:\n\n\n# zpool import\n  pool: test\n    id: 10121266328238932306\n state: DEGRADED\nstatus: One or more devices are missing from the system.\naction: The pool can be imported despite missing or damaged devices.  The\n        fault tolerance of the pool may be compromised if imported.\n   see: http://illumos.org/msg/ZFS-8000-2Q\nconfig:\n\n        test              DEGRADED\n          mirror          DEGRADED\n            c0t0d0        ONLINE\n            c0t0d1        FAULTED   cannot open\n\n\nUnlike when the pool is active on the system, the device cannot be\nreplaced while the pool is exported.  If the device can be attached to the\nsystem, attach the device and run 'zpool import' again.\n\nAlternatively, the pool can be imported as-is, though it will be placed\nin the DEGRADED state due to a missing device.  The device will be marked as\nUNAVAIL.  Once the pool has been imported, the missing device can be replaced as\ndescribed above.\n   "
  59   59  #
  60   60  # code: ZFS-8000-3C
  61   61  # keys: ereport.fs.zfs.device.missing_nr
  62   62  #
  63   63  msgid "ZFS-8000-3C.type"
  64   64  msgstr "Error"
  65   65  msgid "ZFS-8000-3C.severity"
  66   66  msgstr "Critical"
  67   67  msgid "ZFS-8000-3C.description"
  68   68  msgstr "A device could not be opened and no replicas are available.  Refer to %s for more information."
  69   69  msgid "ZFS-8000-3C.response"
  70   70  msgstr "No automated response will be taken."
  71   71  msgid "ZFS-8000-3C.impact"
  72   72  msgstr "The pool is no longer available"
  73   73  msgid "ZFS-8000-3C.action"
  74   74  msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: One or more devices could not be opened.  There are insufficient\n      replicas for the pool to continue functioning.\naction: Attach the missing device and online it using 'zpool online'.\n   see: http://illumos.org/msg/ZFS-8000-3C\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  insufficient replicas\n          c0t0d0              ONLINE       0     0     0\n          c0t0d1              FAULTED      0     0     0  cannot open\n\nerrors: No known data errors\n\n\nIf the device has been temporarily detached from the system, attach the\ndevice to the system and run 'zpool status' again.  The pool should\nautomatically detect the newly attached device and resume functioning.  You may\nhave to mount the filesystems in the pool explicitly using 'zfs\nmount -a'.\n\nIf the device is no longer available and cannot be reattached to the\nsystem, then the pool must be destroyed and re-created from a backup\nsource.\n\nFor an exported pool\n\nIf this error is encountered during a 'zpool import', it means that one\nof the devices is not attached to the system:\n\n\n# zpool import\n  pool: test\n    id: 10121266328238932306\n state: FAULTED\nstatus: One or more devices are missing from the system.\naction: The pool cannot be imported.  Attach the missing devices and\n    try again.\n   see: http://illumos.org/msg/ZFS-8000-3C\nconfig:\n\n        test              FAULTED   insufficient replicas\n          c0t0d0          ONLINE\n          c0t0d1          FAULTED   cannot open\n\n\nThe pool cannot be imported until the missing device is attached to the\nsystem.  If the device has been made available in an alternate location, use the\n'-d' option to 'zpool import' to search for devices in a different directory.\nIf the missing device is unavailable, then the pool cannot be imported.\n        "
  75   75  #
  76   76  # code: ZFS-8000-4J
  77   77  # keys: ereport.fs.zfs.device.corrupt_label_r
  78   78  #
  79   79  msgid "ZFS-8000-4J.type"
  80   80  msgstr "Error"
  81   81  msgid "ZFS-8000-4J.severity"
  82   82  msgstr "Major"
  83   83  msgid "ZFS-8000-4J.description"
  84   84  msgstr "A device could not be opened due to a missing or invalid\n          device label.  Refer to %s for more information."
  85   85  msgid "ZFS-8000-4J.response"
  86   86  msgstr "A hot spare will be activated if available."
  87   87  msgid "ZFS-8000-4J.impact"
  88   88  msgstr "The pool is no longer providing the configured level of\n           replication."
  89   89  msgid "ZFS-8000-4J.action"
  90   90  msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices could not be used because the label is missing or\n        invalid.  Sufficient replicas exist for the pool to continue\n        functioning in a degraded state.\naction: Replace the device using 'zpool replace'.\n   see: http://illumos.org/msg/ZFS-8000-4J\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  DEGRADED     0     0     0\n          mirror              DEGRADED     0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            FAULTED      0     0     0  corrupted data\n\nerrors: No known data errors\n\n\nIf the device has been temporarily detached from the system, attach the\ndevice to the system and run 'zpool status' again.  The pool should\nautomatically detect the newly attached device and resume functioning.\n\nIf the device is no longer available, it can be replaced using 'zpool\nreplace':\n\n\n# zpool replace test c0t0d1 c0t0d2\n\n\nIf the device has been replaced by another disk in the same physical\nslot, then the device can be replaced using a single argument to the 'zpool\nreplace' command:\n\n\n# zpool replace test c0t0d1\n\n\nZFS will begin migrating data to the new device as soon as the replace\nis issued.  Once the resilvering completes, the original device (if different\nfrom the replacement) will be removed, and the pool will be restored to the\nONLINE state.\n\nFor an exported pool\n\nIf this error is encountered while running 'zpool import', the pool can\nbe still be imported despite the failure:\n\n\n# zpool import\n  pool: test\n    id: 5187963178597328409\n state: DEGRADED\nstatus: One or more devices contains corrupted data.  The fault tolerance of\n     the pool may be compromised if imported.\naction: The pool can be imported using its name or numeric identifier.\n   see: http://illumos.org/msg/ZFS-8000-4J\nconfig:\n\n        test              DEGRADED\n          mirror          DEGRADED\n            c0t0d0        ONLINE\n            c0t0d1        FAULTED   corrupted data\n\n\nTo import the pool, run 'zpool import':\n\n\n# zpool import test\n\n\nOnce the pool has been imported, the damaged device can be replaced\naccording to the above procedure.\n       "
  91   91  #
  92   92  # code: ZFS-8000-5E
  93   93  # keys: ereport.fs.zfs.device.corrupt_label_nr
  94   94  #
  95   95  msgid "ZFS-8000-5E.type"
  96   96  msgstr "Error"
  97   97  msgid "ZFS-8000-5E.severity"
  98   98  msgstr "Critical"
  99   99  msgid "ZFS-8000-5E.description"
 100  100  msgstr "A device could not be opened due to a missing or invalid\n          device label and no replicas are available.  Refer to %s for more information."
 101  101  msgid "ZFS-8000-5E.response"
 102  102  msgstr "No automated response will be taken."
 103  103  msgid "ZFS-8000-5E.impact"
 104  104  msgstr "The pool is no longer available"
 105  105  msgid "ZFS-8000-5E.action"
 106  106  msgstr "\nFor an active pool\n\nIf this error was encountered while running 'zpool import', please see\nthe section below.  Otherwise, run 'zpool status -x' to determine which pool\nhas experienced a failure:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: One or more devices could not be used because the the label is missing \n        or invalid.  There are insufficient replicas for the pool to continue\n        functioning.\naction: Destroy and re-create the pool from a backup source.\n   see: http://illumos.org/msg/ZFS-8000-5E\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  insufficient replicas\n          c0t0d0              FAULTED      0     0     0  corrupted data\n          c0t0d1              ONLINE       0     0     0\n\nerrors: No known data errors\n\n\nThe device listed as FAULTED with 'corrupted data' cannot be opened due\nto a corrupt label.  ZFS will be unable to use the pool, and all data within the\npool is irrevocably lost.  The pool must be destroyed and recreated from an\nappropriate backup source.  Using replicated configurations will prevent this\nfrom happening in the future.\n\nFor an exported pool\n\nIf this error is encountered during 'zpool import', the action is the\nsame.  The pool cannot be imported - all data is lost and must be restored from\nan appropriate backup source.\n   "
 107  107  #
 108  108  # code: ZFS-8000-6X
 109  109  # keys: ereport.fs.zfs.pool.bad_guid_sum
 110  110  #
 111  111  msgid "ZFS-8000-6X.type"
 112  112  msgstr "Error"
 113  113  msgid "ZFS-8000-6X.severity"
 114  114  msgstr "Critical"
 115  115  msgid "ZFS-8000-6X.description"
 116  116  msgstr "One or more top level devices are missing.  Refer to %s for more information."
 117  117  msgid "ZFS-8000-6X.response"
 118  118  msgstr "No automated response will be taken."
 119  119  msgid "ZFS-8000-6X.impact"
 120  120  msgstr "The pool cannot be imported"
 121  121  msgid "ZFS-8000-6X.action"
 122  122  msgstr "\nRun 'zpool import' to list which pool cannot be imported:\n\n\n# zpool import\n  pool: test\n    id: 13783646421373024673\n state: FAULTED\nstatus: One or more devices are missing from the system.\naction: The pool cannot be imported.  Attach the missing\n      devices and try again.\n   see: http://illumos.org/msg/ZFS-8000-6X\nconfig:\n\n        test              FAULTED   missing device\n          c0t0d0          ONLINE\n\n        Additional devices are known to be part of this pool, though their\n        exact configuration cannot be determined.\n\n\nZFS attempts to store enough configuration data on the devices such\nthat the configuration is recoverable from any subset of devices.  In some\ncases, particularly when an entire toplevel virtual device is not attached to\nthe system, ZFS will be unable to determine the complete configuration.  It will\nalways detect that these devices are missing, even if it cannot identify all of\nthe devices.\n\nThe pool cannot be imported until the unknown missing device is\nattached to the system.  If the device has been made available in an alternate\nlocation, use the '-d' option to 'zpool import' to search for devices in a\ndifferent directory.  If the missing device is unavailable, then the pool cannot\nbe imported.\n      "
 123  123  #
 124  124  # code: ZFS-8000-72
 125  125  # keys: ereport.fs.zfs.pool.corrupt_pool
 126  126  #
 127  127  msgid "ZFS-8000-72.type"
 128  128  msgstr "Error"
 129  129  msgid "ZFS-8000-72.severity"
 130  130  msgstr "Critical"
 131  131  msgid "ZFS-8000-72.description"
 132  132  msgstr "The metadata required to open the pool is corrupt.  Refer to %s for more information."
 133  133  msgid "ZFS-8000-72.response"
 134  134  msgstr "No automated response will be taken."
 135  135  msgid "ZFS-8000-72.impact"
 136  136  msgstr "The pool is no longer available"
 137  137  msgid "ZFS-8000-72.action"
 138  138  msgstr "\nEven though all the devices are available, the on-disk data\nhas been corrupted such that the pool cannot be opened.  If a recovery\naction is presented, the pool can be returned to a usable state.\nOtherwise, all data within the pool is lost, and the pool must be\ndestroyed and restored from an appropriate backup source.  ZFS\nincludes built-in metadata replication to prevent this from happening\neven for unreplicated pools, but running in a replicated configuration\nwill decrease the chances of this happening in the future.\n\nIf this error is encountered during 'zpool import', see the\nsection below.  Otherwise, run 'zpool status -x' to determine which\npool is faulted and if a recovery option is available:\n\n\n# zpool status -x\n  pool: test\n    id: 13783646421373024673\n state: FAULTED\nstatus: The pool metadata is corrupted and cannot be opened.\naction: Recovery is possible, but will result in some data loss.\n        Returning the pool to its state as of Mon Sep 28 10:24:39 2009\n        should correct the problem.  Approximately 59 seconds of data\n        will have to be discarded, irreversibly.  Recovery can be\n        attempted by executing 'zpool clear -F test'.  A scrub of the pool\n        is strongly recommended following a successful recovery.\n   see: http://illumos.org/msg/ZFS-8000-72\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     2  corrupted data\n            c0t0d0            ONLINE       0     0     2\n            c0t0d1            ONLINE       0     0     2\n\n\nIf recovery is unavailable, the recommended action will be:\n\n\naction: Destroy the pool and restore from backup.\n\n\nIf this error is encountered during 'zpool import', and if no\nrecovery option is mentioned, the pool is unrecoverable and cannot be\nimported.  The pool must be restored from an appropriate backup\nsource.  If a recovery option is available, the output from 'zpool\nimport' will look something like the following:\n\n\n# zpool import share\ncannot import 'share': I/O error\n        Recovery is possible, but will result in some data loss.\n        Returning the pool to its state as of Sun Sep 27 12:31:07 2009\n        should correct the problem.  Approximately 53 seconds of data\n        will have to be discarded, irreversibly.  Recovery can be\n        attempted by executing 'zpool import -F share'.  A scrub of the pool\n        is strongly recommended following a successful recovery.\n\n\nRecovery actions are requested with the -F option to either\n'zpool clear' or 'zpool import'.  Recovery will result in some data\nloss, because it reverts the pool to an earlier state.  A dry-run\nrecovery check can be performed by adding the -n option, affirming if\nrecovery is possible without actually reverting the pool to its\nearlier state.\n"
 139  139  #
 140  140  # code: ZFS-8000-8A
 141  141  # keys: ereport.fs.zfs.object.corrupt_data
 142  142  #
 143  143  msgid "ZFS-8000-8A.type"
  
    | 
      ↓ open down ↓ | 
    143 lines elided | 
    
      ↑ open up ↑ | 
  
 144  144  msgstr "Error"
 145  145  msgid "ZFS-8000-8A.severity"
 146  146  msgstr "Critical"
 147  147  msgid "ZFS-8000-8A.description"
 148  148  msgstr "A file or directory could not be read due to corrupt data.  Refer to %s for more information."
 149  149  msgid "ZFS-8000-8A.response"
 150  150  msgstr "No automated response will be taken."
 151  151  msgid "ZFS-8000-8A.impact"
 152  152  msgstr "The file or directory is unavailable."
 153  153  msgid "ZFS-8000-8A.action"
 154      -msgstr "\nRun 'zpool status -x' to determine which pool is damaged:\n\n\n# zpool status -x\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://illumos.org/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: 1 data errors, use '-v' for a list\n\n\nUnfortunately, the data cannot be repaired, and the only choice to\nrepair the data is to restore the pool from backup.  Applications attempting to\naccess the corrupted data will get an error (EIO), and data may be permanently\nlost.\n\nOn recent versions of Solaris, the list of affected files can be\nretrieved by using the '-v' option to 'zpool status':\n\n\n# zpool status -xv\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://illumos.org/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: Permanent errors have been detected in the following files:\n\n        /export/example/foo\n\n\nDamaged files may or may not be able to be removed depending on the\ntype of corruption.  If the corruption is within the plain data, the file should\nbe removable.  If the corruption is in the file metadata, then the file cannot\nbe removed, though it can be moved to an alternate location.  In either case,\nthe data should be restored from a backup source.  It is also possible for the\ncorruption to be within pool-wide metadata, resulting in entire datasets being\nunavailable.  If this is the case, the only option is to destroy the pool and\nre-create the datasets from backup.\n       "
      154 +msgstr "\nRun 'zpool status -x' to determine which pool is damaged:\n\n\n# zpool status -x\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://illumos.org/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: 1 data errors, use '-v' for a list\n\n\nUnfortunately, the data cannot be repaired, and the only choice to\nrepair the data is to restore the pool from backup.  Applications attempting to\naccess the corrupted data will get an error (EIO), and data may be permanently\nlost.\n\nOn recent versions of illumos, the list of affected files can be\nretrieved by using the '-v' option to 'zpool status':\n\n\n# zpool status -xv\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an error and no valid replicas\n        are available.  Some filesystem data is corrupt, and applications\n        may have been affected.\naction: Destroy the pool and restore from backup.\n   see: http://illumos.org/msg/ZFS-8000-8A\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     2\n          c0t0d0              ONLINE       0     0     2\n          c0t0d1              ONLINE       0     0     0\n\nerrors: Permanent errors have been detected in the following files:\n\n        /export/example/foo\n\n\nDamaged files may or may not be able to be removed depending on the\ntype of corruption.  If the corruption is within the plain data, the file should\nbe removable.  If the corruption is in the file metadata, then the file cannot\nbe removed, though it can be moved to an alternate location.  In either case,\nthe data should be restored from a backup source.  It is also possible for the\ncorruption to be within pool-wide metadata, resulting in entire datasets being\nunavailable.  If this is the case, the only option is to destroy the pool and\nre-create the datasets from backup.\n       "
 155  155  #
 156  156  # code: ZFS-8000-9P
 157  157  # keys: ereport.fs.zfs.device.failing
 158  158  #
 159  159  msgid "ZFS-8000-9P.type"
 160  160  msgstr "Error"
 161  161  msgid "ZFS-8000-9P.severity"
 162  162  msgstr "Minor"
 163  163  msgid "ZFS-8000-9P.description"
 164  164  msgstr "A device has experienced uncorrectable errors in a\n        replicated configuration.  Refer to %s for more information."
 165  165  msgid "ZFS-8000-9P.response"
 166  166  msgstr "ZFS has attempted to repair the affected data."
 167  167  msgid "ZFS-8000-9P.impact"
 168  168  msgstr "The system is unaffected, though errors may indicate future\n       failure.  Future errors may cause ZFS to automatically fault\n          the device."
 169  169  msgid "ZFS-8000-9P.action"
 170  170  msgstr "\nRun 'zpool status -x' to determine which pool has experienced\nerrors:\n\n\n# zpool status\n  pool: test\n state: ONLINE\nstatus: One or more devices has experienced an unrecoverable error.  An\n        attempt was made to correct the error.  Applications are unaffected.\naction: Determine if the device needs to be replaced, and clear the errors\n        using 'zpool online' or replace the device with 'zpool replace'.\n   see: http://illumos.org/msg/ZFS-8000-9P\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     0\n          mirror              ONLINE       0     0     0\n            c0t0d0            ONLINE       0     0     2\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known data errors\n\n\nFind the device with a non-zero error count for READ, WRITE, or CKSUM.\nThis indicates that the device has experienced a read I/O error, write I/O\nerror, or checksum validation error.  Because the device is part of a mirror or\nRAID-Z device, ZFS was able to recover from the error and subsequently repair\nthe damaged data.\n\nIf these errors persist over a period of time, ZFS may determine the\ndevice is faulty and mark it as such.  However, these error counts may or may\nnot indicate that the device is unusable.  It depends on how the errors were\ncaused, which the administrator can determine in advance of any ZFS diagnosis.\nFor example, the following cases will all produce errors that do not indicate\npotential device failure:\n\n\nA network attached device lost connectivity but has now\nrecovered\nA device suffered from a bit flip, an expected event over long\nperiods of time\nAn administrator accidentally wrote over a portion of the disk using\nanother program\n\n\nIn these cases, the presence of errors does not indicate that the\ndevice is likely to fail in the future, and therefore does not need to be\nreplaced.  If this is the case, then the device errors should be cleared using\n'zpool clear':\n\n\n# zpool clear test c0t0d0\n\n\nOn the other hand, errors may very well indicate that the device has\nfailed or is about to fail.  If there are continual I/O errors to a device that\nis otherwise attached and functioning on the system, it most likely needs to be\nreplaced.   The administrator should check the system log for any driver\nmessages that may indicate hardware failure.  If it is determined that the\ndevice needs to be replaced, then the 'zpool replace' command should be\nused:\n\n\n# zpool replace test c0t0d0 c0t0d2\n\n\nThis will attach the new device to the pool and begin resilvering data\nto it.  Once the resilvering process is complete, the old device will\nautomatically be removed from the pool, at which point it can safely be removed\nfrom the system.  If the device needs to be replaced in-place (because there are\nno available spare devices), the original device can be removed and replaced\nwith a new device, at which point a different form of 'zpool replace' can be\nused:\n\n\n# zpool replace test c0t0d0\n\n\nThis assumes that the original device at 'c0t0d0' has been replaced\nwith a new device under the same path, and will be replaced\nappropriately.\n\nYou can monitor the progress of the resilvering operation by using the\n'zpool status -x' command:\n\n\n# zpool status -x\n  pool: test\n state: DEGRADED\nstatus: One or more devices is currently being replaced.  The pool may not be\n     providing the necessary level of replication.\naction: Wait for the resilvering operation to complete\n scrub: resilver in progress, 0.14% done, 0h0m to go\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  ONLINE       0     0     0\n          mirror              ONLINE       0     0     0\n            replacing         ONLINE       0     0     0\n              c0t0d0          ONLINE       0     0     3\n              c0t0d2          ONLINE       0     0     0  58.5K resilvered\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known data errors\n\n      "
 171  171  #
 172  172  # code: ZFS-8000-A5
 173  173  # keys: ereport.fs.zfs.device.version_mismatch
 174  174  #
 175  175  msgid "ZFS-8000-A5.type"
 176  176  msgstr "Error"
 177  177  msgid "ZFS-8000-A5.severity"
 178  178  msgstr "Major"
 179  179  msgid "ZFS-8000-A5.description"
 180  180  msgstr "The on-disk version is not compatible with the running\n            system.  Refer to %s for more information."
 181  181  msgid "ZFS-8000-A5.response"
 182  182  msgstr "No automated response will occur."
 183  183  msgid "ZFS-8000-A5.impact"
 184  184  msgstr "The pool is unavailable."
 185  185  msgid "ZFS-8000-A5.action"
 186  186  msgstr "\nIf this error is seen during 'zpool import', see the section below.\nOtherwise, run 'zpool status -x' to determine which pool is faulted:\n\n\n# zpool status -x\n  pool: test\n state: FAULTED\nstatus: The ZFS version for the pool is incompatible with the software running\n        on this system.\naction: Destroy and re-create the pool.\n scrub: none requested\nconfig:\n\n        NAME                  STATE     READ WRITE CKSUM\n        test                  FAULTED      0     0     0  incompatible version\n          mirror              ONLINE       0     0     0\n            c0t0d0            ONLINE       0     0     0\n            c0t0d1            ONLINE       0     0     0\n\nerrors: No known errors\n\n\nThe pool cannot be used on this system.  Either move the storage to the\nsystem where the pool was originally created, upgrade the current system\nsoftware to a more recent version, or destroy the pool and re-create it from\nbackup.\n\nIf this error is seen during import, the pool cannot be imported on the\ncurrent system.  The disks must be attached to the system which originally\ncreated the pool, and imported there.\n\nThe list of currently supported versions can be displayed using 'zpool\nupgrade -v'.\n "
 187  187  #
 188  188  # code: ZFS-8000-CS
 189  189  # keys: fault.fs.zfs.pool
 190  190  #
 191  191  msgid "ZFS-8000-CS.type"
 192  192  msgstr "Fault"
 193  193  msgid "ZFS-8000-CS.severity"
 194  194  msgstr "Major"
 195  195  msgid "ZFS-8000-CS.description"
 196  196  msgstr "A ZFS pool failed to open.  Refer to %s for more information."
 197  197  msgid "ZFS-8000-CS.response"
 198  198  msgstr "No automated response will occur."
 199  199  msgid "ZFS-8000-CS.impact"
 200  200  msgstr "The pool data is unavailable"
 201  201  msgid "ZFS-8000-CS.action"
 202  202  msgstr "Run 'zpool status -x' and attach any missing devices, follow\n     any provided recovery instructions or restore from backup."
 203  203  #
 204  204  # code: ZFS-8000-D3
 205  205  # keys: fault.fs.zfs.device
 206  206  #
 207  207  msgid "ZFS-8000-D3.type"
 208  208  msgstr "Fault"
 209  209  msgid "ZFS-8000-D3.severity"
 210  210  msgstr "Major"
 211  211  msgid "ZFS-8000-D3.description"
 212  212  msgstr "A ZFS device failed.  Refer to %s for more information."
 213  213  msgid "ZFS-8000-D3.response"
 214  214  msgstr "No automated response will occur."
 215  215  msgid "ZFS-8000-D3.impact"
 216  216  msgstr "Fault tolerance of the pool may be compromised."
 217  217  msgid "ZFS-8000-D3.action"
 218  218  msgstr "Run 'zpool status -x' and replace the bad device."
 219  219  #
 220  220  # code: ZFS-8000-EY
 221  221  # keys: ereport.fs.zfs.pool.hostname_mismatch
 222  222  #
 223  223  msgid "ZFS-8000-EY.type"
 224  224  msgstr "Error"
 225  225  msgid "ZFS-8000-EY.severity"
 226  226  msgstr "Major"
 227  227  msgid "ZFS-8000-EY.description"
 228  228  msgstr "The ZFS pool was last accessed by another system  Refer to %s for more information."
 229  229  msgid "ZFS-8000-EY.response"
 230  230  msgstr "No automated response will be taken."
 231  231  msgid "ZFS-8000-EY.impact"
 232  232  msgstr "ZFS filesystems are not available"
 233  233  msgid "ZFS-8000-EY.action"
 234  234  msgstr "\n\nThe pool has been written to from another host, and was not cleanly exported\nfrom the other system.  Actively importing a pool on multiple systems will\ncorrupt the pool and leave it in an unrecoverable state.  To determine which\nsystem last accessed the pool, run the 'zpool import' command:\n\n\n# zpool import\n  pool: test\n    id: 14702934086626715962\nstate:  ONLINE\nstatus: The pool was last accessed by another system.\naction: The pool can be imported using its name or numeric identifier and\n        the '-f' flag.\n   see: http://illumos.org/msg/ZFS-8000-EY\nconfig:\n\n        test              ONLINE\n          c0t0d0          ONLINE\n\n# zpool import test\ncannot import 'test': pool may be in use from other system, it was last\naccessed by 'tank' (hostid: 0x1435718c) on Fri Mar  9 15:42:47 2007\nuse '-f' to import anyway\n\n\n\nIf you are certain that the pool is not being actively accessed by another\nsystem, then you can use the '-f' option to 'zpool import' to forcibly\nimport the pool.\n\n "
 235  235  #
 236  236  # code: ZFS-8000-FD
 237  237  # keys: fault.fs.zfs.vdev.io
 238  238  #
 239  239  msgid "ZFS-8000-FD.type"
 240  240  msgstr "Fault"
 241  241  msgid "ZFS-8000-FD.severity"
 242  242  msgstr "Major"
 243  243  msgid "ZFS-8000-FD.description"
 244  244  msgstr "The number of I/O errors associated with a ZFS device exceeded\n             acceptable levels.  Refer to %s for more information."
 245  245  msgid "ZFS-8000-FD.response"
 246  246  msgstr "The device has been offlined and marked as faulted.  An attempt\n            will be made to activate a hot spare if available. "
 247  247  msgid "ZFS-8000-FD.impact"
 248  248  msgstr "Fault tolerance of the pool may be compromised."
 249  249  msgid "ZFS-8000-FD.action"
 250  250  msgstr "Run 'zpool status -x' and replace the bad device."
 251  251  #
 252  252  # code: ZFS-8000-GH
 253  253  # keys: fault.fs.zfs.vdev.checksum
 254  254  #
 255  255  msgid "ZFS-8000-GH.type"
 256  256  msgstr "Fault"
 257  257  msgid "ZFS-8000-GH.severity"
 258  258  msgstr "Major"
 259  259  msgid "ZFS-8000-GH.description"
 260  260  msgstr "The number of checksum errors associated with a ZFS device\nexceeded acceptable levels.  Refer to %s for more information."
 261  261  msgid "ZFS-8000-GH.response"
 262  262  msgstr "The device has been marked as degraded.  An attempt\nwill be made to activate a hot spare if available."
 263  263  msgid "ZFS-8000-GH.impact"
 264  264  msgstr "Fault tolerance of the pool may be compromised."
 265  265  msgid "ZFS-8000-GH.action"
 266  266  msgstr "Run 'zpool status -x' and replace the bad device."
 267  267  #
 268  268  # code: ZFS-8000-HC
 269  269  # keys: fault.fs.zfs.io_failure_wait
 270  270  #
 271  271  msgid "ZFS-8000-HC.type"
 272  272  msgstr "Error"
 273  273  msgid "ZFS-8000-HC.severity"
 274  274  msgstr "Major"
 275  275  msgid "ZFS-8000-HC.description"
 276  276  msgstr "The ZFS pool has experienced currently unrecoverable I/O\n          failures.  Refer to %s for more information."
 277  277  msgid "ZFS-8000-HC.response"
 278  278  msgstr "No automated response will be taken."
 279  279  msgid "ZFS-8000-HC.impact"
 280  280  msgstr "Read and write I/Os cannot be serviced."
 281  281  msgid "ZFS-8000-HC.action"
 282  282  msgstr "Make sure the affected devices are connected, then run\n            'zpool clear'."
 283  283  #
 284  284  # code: ZFS-8000-JQ
 285  285  # keys: fault.fs.zfs.io_failure_continue
 286  286  #
 287  287  msgid "ZFS-8000-JQ.type"
 288  288  msgstr "Error"
 289  289  msgid "ZFS-8000-JQ.severity"
 290  290  msgstr "Major"
 291  291  msgid "ZFS-8000-JQ.description"
 292  292  msgstr "The ZFS pool has experienced currently unrecoverable I/O\n          failures.  Refer to %s for more information."
 293  293  msgid "ZFS-8000-JQ.response"
 294  294  msgstr "No automated response will be taken."
 295  295  msgid "ZFS-8000-JQ.impact"
 296  296  msgstr "Read and write I/Os cannot be serviced."
 297  297  msgid "ZFS-8000-JQ.action"
 298  298  msgstr "Make sure the affected devices are connected, then run\n            'zpool clear'."
 299  299  #
 300  300  # code: ZFS-8000-K4
 301  301  # keys: fault.fs.zfs.log_replay
 302  302  #
 303  303  msgid "ZFS-8000-K4.type"
 304  304  msgstr "Error"
 305  305  msgid "ZFS-8000-K4.severity"
 306  306  msgstr "Major"
 307  307  msgid "ZFS-8000-K4.description"
 308  308  msgstr "A ZFS intent log device could not be read.  Refer to %s for more information."
 309  309  msgid "ZFS-8000-K4.response"
 310  310  msgstr "No automated response will be taken."
 311  311  msgid "ZFS-8000-K4.impact"
 312  312  msgstr "The intent log(s) cannot be replayed."
 313  313  msgid "ZFS-8000-K4.action"
 314  314  msgstr "Either restore the affected device(s) and run 'zpool online',\n     or ignore the intent log records by running 'zpool clear'."
 315  315  
  
    | 
      ↓ open down ↓ | 
    151 lines elided | 
    
      ↑ open up ↑ | 
  
XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX