casper....@sun.com wrote:
I'm not saying that ZFS should consider doing this - doing a validation
for in-memory data is non-trivially expensive in performance terms, and
there's only so much you can do and still expect your machine to
survive. I mean, I've used the old NonStop stuff, and yes, you can
shoot them with a .45 and it likely will still run, but wacking them
with a bazooka still is guarantied to make them, well, Non-NonStop.
If we scrub the memory anyway, why not include the check of the ZFS
checksums which are already in memory?
OTOH, zfs gets a lot of mileage out of cheap hardware and we know what the
limitations are when you don't use ECC; the industry must start to require
that all chipsets support ECC.
Caspe
Reading the paper was interesting, as it highlighted all the places
where ZFS "skips" validation. There's a lot of places. In many ways,
fixing this would likely make ZFS similar to AppleTalk whose notorious
performance (relative to Ethernet) was caused by what many called the
"Are You Sure?" design. Double and Triple checking absolutely
everything has it's costs.
And, yes, we really should just force computer manufacturers to use ECC
in more places (not just RAM) - as densities and data volumes increase,
we are more likely to see errors, and without proper hardware checking,
we're really going out on a limb here to be able to trust what the
hardware says. And, let's face it - hardware error correction is /so/
much faster than doing it in software.
--
Erik Trimble
Java System Support
Mailstop: usca22-123
Phone: x17195
Santa Clara, CA
Timezone: US/Pacific (GMT-0800)
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss