Opensolaris Aserver wrote:
We tried to replicate a snapshot via the built-in send receive zfs tools.
...
ZFS: bad checksum (read on <unknown> off 0: zio ffffffff3017b300 [L0 ZFS plain file] 20000L/20000P DVA[0]=<0:3b98ed1e800:25800> fletcher2 uncompressed LE contiguous birth=806063 fill=1 cksum=a487e32d
...
errors: Permanent errors have been detected in the following files:

stor/[EMAIL PROTECTED]:01:00:/1003/kreos11/HB1030/C_Root/Documents and Settings/bvp/My Documents/My Pictures/confidential/tconfidential/confidential/96
...
Son we decided to destroy this snapshot, and then started another Replication.

This time the server crashed again :-(

So, some of your data has been lost due to hardware failure, where the hardware has "silently" corrupted your data. ZFS has detected this. If you were to read this data (other than via 'zfs send'), you will get EIO, and as you note, 'zfs status' shows what files are affected.

The 'zfs send' protocol isn't able to tell the other side "this part of this file is corrupt", so it panics. This is a bug.

The reason you're seeing the panic when 'zfs send'-ing the next snapshot is that the (corrupt) data is shared between multiple snapshots.

You can work around this by deleting or overwriting the files, then taking and sending a new snapshot.

--matt

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to