>>>>> "re" == Richard Elling <[EMAIL PROTECTED]> writes:

    re> unrecoverable read as the dominant disk failure mode. [...]
    re> none of the traditional software logical volume managers nor
    re> the popular open source file systems (other than ZFS :-)
    re> address this problem.

Other LVM's should address unrecoverable read errors as well or better
than ZFS, because that's when the drive returns an error instead of
data.  Doing a good job with this error is mostly about not freezing
the whole filesystem for the 30sec it takes the drive to report the
error.  Either the drives should be loaded with special firmware that
returns errors earlier, or the software LVM should read redundant data
and collect the statistic if the drive is well outside its usual
response latency.  I would expect all the software volume managers
including ZFS fail to do this.  It's really hard to test without
somehow getting a drive that returns read errors frequently, but isn't
about to die within the month---maybe ZFS should have an error
injector at driver-level instead of block-level, and a model for
time-based errors.  One thing other LVM's seem like they may do better
than ZFS, based on not-quite-the-same-scenario tests, is not freeze
filesystems unrelated to the failing drive during the 30 seconds it's
waiting for the I/O request to return an error.

In terms of FUD about ``silent corruption'', there is none of it when
the drive clearly reports a sector is unreadable.  Yes, traditional
non-big-storage-vendor RAID5, and all software LVM's I know of except
ZFS, depend on the drives to report unreadable sectors.  And,
generally, drives do.  so let's be clear about that and not try to imply
that the ``dominant failure mode'' causes silent corruption for
everyone except ZFS and Netapp users---it doesn't.

The Netapp paper focused on when drives silently return incorrect
data, which is different than returning an error.  Both Netapp and ZFS
do checksums to protect from this.  However Netapp never claimed this
failure mode was more common than reported unrecoverable read errors,
just that it was more interesting.  I expect it's much *less* common.

Further, we know Netapp loaded special firmware into the enterprise
drives in that study because they wanted the larger sector size.  They
are likely also loading special firmware into the desktop drives to
make them return errors sooner than 30 seconds.  so, it's not
improbable that the Netapp drives are more prone to deliver silently
corrupt data instead of UNC/seek errors compared to off-the-shelf
drives.

Finally, for the Google paper, silent corruption ``didn't even make
the chart.''  so, saying something didn't make your chart and saying
that it doesn't happen are two different things, and your favoured
conclusion has a stake in maintaining that view, too.

Attachment: pgpsI3gZXY8Ue.pgp
Description: PGP signature

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to