>>>>> "mb" == Marc Bevand <[EMAIL PROTECTED]> writes:

    mb> Ask your hardware vendor. The hardware corrupted your data,
    mb> not ZFS.

You absolutely do NOT have adequate basis to make this statement.

I would further argue that you are probably wrong, and that I think
based on what we know that the pool was probably corrupted by a bug in
ZFS.  Simply because ZFS is (a) able to detect problems with hardware
when they exist, and (b) ringing an alarm bell of some sort, does NOT
exhonerate ZFS.  and AIUI that is your position.

Further, ZFS's ability to use zpool-level redundancy heal problems
created by its own bugs is not a cause for celebration or an
improvement over filesystems without bugs.  The virtue of the
self-healing is for when hardware actually does fail.  If self-healing
also helps with corruption created by bugs in ZFS, that does not shift
blame for unhealed bug-corruption back to the hardware, nor make ZFS
more robust than a different filesystem without corruption bugs.

    mb> Other filesystems would have returned silently corrupted
    mb> data and it would have maybe taken you days/weeks to
    mb> troubleshoot

possibly.  very likely, other filesystems would have handled it fine.

Boris, have a look at the two links I posted earlier about ``simon
sez, import!'' incantations, and required patches.

  http://opensolaris.org/jive/message.jspa?messageID=192572#194209
  http://sunsolve.sun.com/search/document.do?assetkey=1-66-233602-1

panic-on-import, sounds a lot like your problem.  Jonathan also posted
http://www.opensolaris.org/jive/thread.jspa?messageID=220125 which
seems to be incomplete instructions on how to choose a different
ueberblock which helped someone else with a corrupted pool, but the OP
in that thread never wrote it up in recipe form for ignorant sysadmins
like me to follow so it might not be widely useful.

In short, ZFS is unstable and prone to corruption, but may improve
substantially when patched up to the latest revision.  And many fixes
are available now, but some which are in SXCE right now will be
available in the stable binary-only Solaris not until u6 so we haven't
yet gained experience with how much improvement the patches provide.
And finally, there is no way to back up a ZFS filesystem with lots of
clones which is similarly robust to past Unix backup systems---your
best bet for space-efficient backups is to zfs send/recv data onto a
separate ZFS pool.

In more detail, I think there is some experience here that when a
single storage subsystem hosting both ZFS pools and vxfs filesystems
goes away, ZFS pools sometimes become corrupt while vxfs rolls its log
and continues.  so, in stable Sol10u5, ZFS is probably more prone to
metadata corruption causing whole-pool-failure than other logging
filesystems.  some fixes are around the corner, and others are
apparently the subject of some philosophical debate.

Attachment: pgpWGngZltSqj.pgp
Description: PGP signature

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to