Matthew Angelo wrote:
Hi there,

Is there a way to get as much data as possible off an existing slightly corrupted zpool? I have a 2 disk stripe which I'm moving to new storage. I will be moving it to a ZFS Mirror, however at the moment I'm having problems with ZFS Panic'ing the system during a send | recv.

set your failmode to continue.  If you don't have a failmode parameter,
then you'll want to try a later version of Solaris.

I blogged about some techniques to deal with borken files a while back...
http://blogs.sun.com/relling/entry/holy_smokes_a_holey_file
http://blogs.sun.com/relling/entry/dd_tricks_for_holey_files
http://blogs.sun.com/relling/entry/more_on_holey_files

NB: if it is the receive that is panicing, then please file a bug.
-- richard


I don't know exactly how much data is valid. Everything appears to run as expected and applications aren't crashing.

Doing an $( ls -lR | grep -i "IO Error" ) returns roughly 10-15 files which are affected. Luckily, these files ls is returning aren't super critical.

Is it possible to tell ZFS to do a emergency "copy as much valid data off this file system"? I've tried disabling checkums on the corrupted source zpool. But even still, once ZFS runs into an error the zpool is FAULTED and the kernel panic's and the system crashes. Is it possible to tell the zpool to ignore any errors and continue without faulting the zpool?

We have a backup of the data, which is 2 months old. Is it slightly possible to bring this backup online, and 'sync as much as it can' between the two volumes? Could this just be a rsync job?

Thanks



[root]# zpool status -v apps
  pool: apps
 state: ONLINE
status: One or more devices has experienced an error resulting in data
        corruption.  Applications may be affected.
action: Restore the file in question if possible.  Otherwise restore the
        entire pool from backup.
   see: http://www.sun.com/msg/ZFS-8000-8A
 scrub: none requested
config:

        NAME        STATE     READ WRITE CKSUM
        apps        ONLINE       0     0     120
          c1t1d0    ONLINE       0     0     60
          c1t2d0    ONLINE       0     0     0
          c1t3d0    ONLINE       0     0     60

errors: Permanent errors have been detected in the following files:

        apps:<0x0>
        <0x1d2>:<0x0>

------------------------------------------------------------------------

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to