Hi,

My zpool is reporting unrecoverable errors with the metadata:

   pool: rpool2
>   state: ONLINE
>  status: One or more devices has experienced an error resulting in data
>              corruption.  Applications may be effected.
>  action: Restore the file in question if possible. Otherwise restore the
> entire pool from backup.
>      see: http://www.sun.com/msg/ZFS-8000-8A
>
> (snip)
>
> "errors: Permanent errors have been detected in the following files:
>     <metadata>:<0x0>
>     <metadata>:<0x1>
>

It initially reported a DEGRADED pool, but after a reboot, the pool is now
ONLINE and a quick inspection indicates that my data is present and intact
(though the errors stop the file-systems in the pool from mounting at boot -
it drops into maintenance mode). My reading of
http://www.sun.com/msg/ZFS-8000-8A indicates I should destroy the pool and
start again, but http://www.crypticide.com/dropsafe/article/2162 gives me
some small hope that this might be fixable...

The pool has been 'a little flakey' since I built it two months back.  I've
been getting small numbers of read and checksum errors on a few of the disks
each day.  Initially I replaced the disks, but they would always pass all
testing, so lately I've just been clearing the errors each day and looking
for another solution.  I thought I had found it when I discovered WD had a
firmware patch  (http://www.3ware.com/kb/article.aspx?id=15592 ,
http://blog.insanegenius.com/2009/09/western-digital-re4-gp-2tb-drive.html)
  which solved bugs in the drive spin-up behaviour which has been causing
problems for various hardware RAID controllers.   So, yesterday I shutdown
the machine, pulled 2 of the 4 troublesome disks, and applied the firmware
upgrade (04.05G05).   When I booted, the pool was degraded and showed
metadata errors.  After a shutdown and cold start, the pool was ONLINE but
still have metadata errors (so somewhat inconsistent guidance from ZPOOL
STATUS.

Can anyone explain what this 'metadata' is?

More Details:

   - This is a backup server so I can rebuild if necessary, but on principle
   I'd like to have a go at fixing it...
   - The zpool has 96 x 2TB drives divided in to RAIDZ2 sets of 8 (6+2).
   - The drives are Western Digital RE-4's  (WD2002FYPS).
   - Running OpenSolaris build snv_111b.
   - Drives are in two AIC JBODs connectected via SAS.
   - HBA is an LSI 3801E
   - Server is 1RU SuperMicro Intel.

Any advice appreciated!

:-)
Paul Tetley
NearMap Pty Ltd
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to