I know I'm a bit late to contribute to this thread, but I'd still like to
add my $0.02.  My "gut feel" is that we (generally) don't yet understand the
subtleties of disk drive failure modes as they relate to 1.5 or 2Tb+ drives.
 Why?  Because those large drives have not been widely available until
relatively recently.

There's a tendency to extrapolate ones existing knowledge base and
understanding of how/why drives fail (or degrade) by basing our expected
outcome on some "extension" of our existing knowledge base.  In the case of
the current generation of high capacity drives, that may or may not be
appropriate.  We simply don't know!  Mainly because the hard drive
manufacturers, those engineering gods and providers of ever increasing
storage density, don't communicate their acquired and evolving knowledge as
it relates to disk reliability (or failure) mechanisms.

In this case I feel, as a user, it's best to take a very conservative
approach and err on the side of safety by using raidz3 when high capacity
drives are being deployed.  Over time, a consensus based understanding of
the failure modes will emerge and then, from a user perspective, we can have
a clearer understanding of the risks of data loss and its relation to
different ZFS pool configurations.

Personally, I was surprised at how easily I was able to "take out" a 1Tb WD
Caviar black drive by moving a 1U server with the drives spinning.  Earlier
drive generations (500Gb or smaller) tolerated this abuse with no signs of
degradation.  So I know that high capacity drives are a lot more sensitive
to mechanical "abuse" - I can only assume that 2Tb drives are probably even
more sensitive and that shock mounting, to reduce vibration induced by a
bunch of similar drives operating in the same "box", is probably a smart
move.

Likewise, my previous experience has seen how a given percentage of disk
drives would fail in the 2 or 3 week period following a temperature
"excursion" in a data center environment.  Sometimes everyone knows about
that event, and sometimes the folks doing A/C work over a holiday weekend
will "forget" to publish the details of what went wrong! :)   Again - the
same doubts continue to nag me: are the current 1.5Tb+ drives more likely to
suffer degradation due to a temperature excursion over a relatively small
time period?  If the drive firmware does its job and remaps damaged sectors
or tracks transparently, we, as the users, won't know - until it happens one
time too many!!

Regards,

-- 
Al Hopper  Logical Approach Inc,Plano,TX a...@logical-approach.com
                  Voice: 972.379.2133 Timezone: US CDT
OpenSolaris Governing Board (OGB) Member - Apr 2005 to Mar 2007
http://www.opensolaris.org/os/community/ogb/ogb_2005-2007/
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to