Opensolaris Aserver wrote:
We tried to replicate a snapshot via the built-in send receive zfs tools.
...
ZFS: bad checksum (read on <unknown> off 0: zio ffffffff3017b300 [L0 ZFS
plain file] 20000L/20000P DVA[0]=<0:3b98ed1e800:25800> fletcher2
uncompressed LE contiguous birth=806063 fill=1 cksum=a487e32d
...
errors: Permanent errors have been detected in the following files:
stor/[EMAIL PROTECTED]:01:00:/1003/kreos11/HB1030/C_Root/Documents
and Settings/bvp/My Documents/My
Pictures/confidential/tconfidential/confidential/96
...
Son we decided to destroy this snapshot, and then started another
Replication.
This time the server crashed again :-(
So, some of your data has been lost due to hardware failure, where the
hardware has "silently" corrupted your data. ZFS has detected this. If
you were to read this data (other than via 'zfs send'), you will get
EIO, and as you note, 'zfs status' shows what files are affected.
The 'zfs send' protocol isn't able to tell the other side "this part of
this file is corrupt", so it panics. This is a bug.
The reason you're seeing the panic when 'zfs send'-ing the next snapshot
is that the (corrupt) data is shared between multiple snapshots.
You can work around this by deleting or overwriting the files, then
taking and sending a new snapshot.
--matt
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss