Chill. It's a filesystem. If you don't like it, don't use it. Sincere Regards, -Tim
can you guess? wrote: >> can you guess? wrote: >> > > ... > > >>> Most of the balance of your post isn't addressed in >>> >> any detail because it carefully avoids the >> fundamental issues that I raised: >> >>> >>> >> Not true; and by selective quoting you have removed >> my specific >> responses to most of these issues. >> > > While I'm naturally reluctant to call you an outright liar, David, you have > hardly so far in this discussion impressed me as someone whose presentation > is so well-organized and responsive to specific points that I can easily > assume that I simply missed those responses. If you happen to have a copy of > that earlier post, I'd like to see it resubmitted (without modification). > > >>> 1. How much visible damage does a single-bit error >>> >> actually do to the kind of large photographic (e.g., >> RAW) file you are describing? If it trashes the rest >> of the file, as you state is the case with jpeg, then >> you might have a point (though you'd still have to >> address my second issue below), but if it results in >> a virtually invisible blemish they you most certainly >> don't. >> >>> >>> >> I addressed this quite specifically, for two cases >> (compressed raw vs. >> uncompressed raw) with different results. >> > > Then please do so where we all can see it. > > >>> 2. If you actually care about your data, you'd >>> >> have to be a fool to entrust it to *any* single copy, >> regardless of medium. And once you've got more than >> one copy, then you're protected (at the cost of very >> minor redundancy restoration effort in the unlikely >> event that any problem occurs) against the loss of >> any one copy due to a minor error - the only loss of >> non-negligible likelihood that ZFS protects against >> better than other file systems. >> >>> >>> >> You have to detect the problem first. ZFS is in a >> much better position >> to detect the problem due to block checksums. >> > > Bulls***, to quote another poster here who has since been strangely quiet. > The vast majority of what ZFS can detect (save for *extremely* rare > undetectable bit-rot and for real hardware (path-related) errors that studies > like CERN's have found to be very rare - and you have yet to provide even > anecdotal evidence to the contrary) can also be detected by scrubbing, and > it's arguably a lot easier to apply brute-force scrubbing (e.g., by > scheduling a job that periodically copies your data to the null device if > your system does not otherwise support the mechanism) than to switch your > file system. > > >>> If you're relying upon RAID to provide the multiple >>> >> copies - though this would also arguably be foolish, >> if only due to the potential for trashing all the >> copies simultaneously - you'd probably want to >> schedule occasional scrubs, just in case you lost a >> disk. But using RAID as a substitute for off-line >> redundancy is hardly suitable in the kind of >> archiving situations that you describe - and >> therefore ZFS has absolutely nothing of value to >> offer there: you should be using off-line copies, >> and occasionally checking all copies for readability >> (e.g., by copying them to the null device - again, >> something you could do for your on-line copy with a >> cron job and which you should do for your off-line >> copy/copies once in a while as well. >> >>> >>> >> You have to detect the problem first. >> > > And I just described how to above - in a manner that also handles the > off-line storage that you *should* be using for archival purposes (where ZFS > scrubbing is useless). > > ZFS block > >> checksums will detect >> problems that a simple read-only pass through most >> other filesystems >> will not detect. >> > > The only problems that ZFS will detect that a simple read-through pass will > not are those that I just enumerated above: *extremely* rare undetectable > bit-rot and real hardware (path-related) errors that studies like CERN's have > found to be very rare (like, none in the TB-sized installation under > discussion here). > > >>> In sum, your support of ZFS in this specific area >>> >> seems very much knee-jerk in nature rather than >> carefully thought out - exactly the kind of >> 'over-hyping' that I pointed out in my first post in >> this thread. >> >>> >>> >> And your opposition to ZFS appears knee-jerk and >> irrational, from this >> end. But telling you that will have no beneficial >> effect, any more than >> what you just told me about how my opinions appear to >> you. Couldn't we >> leave personalities out of this, in future? >> > > When someone appears to be arguing irrationally, it's at least worth trying > to straighten him out. But I'll stop - *if* you start addressing the very > specific and quantitative issues that you've been so assiduously skirting > until now. > > >>> ... >>> >>> >>> >>>>>> And yet I know many people who have lost data in >>>>>> >>>>>> >>>> ways >>>> >>>> >>>>>> that ZFS would >>>>>> have prevented. >>>>>> >>>>>> >>>>>> >>>>> Specifics would be helpful here. How many? Can >>>>> >> they >> >>>>> >>>>> >>>> reasonably be characterized as consumers (I'll >>>> >> remind >> >>>> you once more: *that's* the subject to which your >>>> comments purport to be responding)? Can the data >>>> >> loss >> >>>> reasonably be characterized as significant (to >>>> 'consumers')? Were the causes hardware problems >>>> >> that >> >>>> could reasonably have been avoided ('bad cables' >>>> might translate to 'improperly inserted, overly >>>> >> long, >> >>>> or severely kinked cables', for example - and such >>>> >> a >> >>>> poorly-constructed system will tend to have other >>>> problems that ZFS cannot address)? >>>> >>>> >>>>> >>>>> >>>>> >>>> "Reasonably avoided" is irrelevant; they *weren't* >>>> avoided. >>>> >>>> >>> While that observation has at least some merit, >>> >> I'll observe that you jumped directly to the last of >> my questions above while carefully ignoring the three >> questions that preceded it. >> >>> >>> >> You'll notice (since you responded) that I got to at >> least one of them >> by the end of the message. >> > > The only specific example you gave later was loss of an entire disk - > something which ZFS does not handle significantly better than conventional > RAID (for volatile data, and including scrubbing) and is a poorer choice for > handling than off-line copies (for archival data). > > The statement of yours to which I was responding was "And yet I know many > people who have lost data in ways that ZFS would have prevented." I suggest > it was not unreasonable for me to have interpreted that sentence as including > by implication " ... and that conventional storage arrangements would *not* > have prevented", since otherwise it's kind of pointless. If interpreted that > way, your later example of loss of an entire disk does not qualify as any > kind of answer to any of the questions that I posed (though it might have had > at least tangential relevance if you had stated that the data loss had been > not because the only copy disappeared but because it occurred on a RAID which > could not be successfully rebuilt due to read errors on the surviving disks - > even here, though, the question then becomes whether it was still possible to > copy most of the data off the degraded array, which it usually should be). > > And you cut out of your > >> quotes what I >> specifically said about cables; that's cheating. >> > > Not at all: there was no need to repeat it (unless you think people > generally don't bother reading your own posts and would like others to try to > help remedy that), because that was the one area in which you actually > responded to what I had asked. > > ... > > >>>> Nearly everybody I can think of who's used a >>>> >> computer >> >>>> for more than a >>>> couple of years has stories of stuff they've lost. >>>> >>>> >>> Of course they have - and usually in ways that ZFS >>> >> would have been no help whatsoever in mitigating. >> >>> >>> >> ZFS will help detect problems with marginal drives, >> cables, power >> supplies, controllers, memory, and motherboards. >> > > None of which the CERN study found in significant numbers (save for their > RAID controller's possible failure to report disk timeouts, but consumer > systems - once again, the subject under discussion here - don't use RAID > controllers). > > The > >> block checksumming >> will show corruptions earlier on, and less >> ambiguously, giving you a >> warning that there's a problem to find and fix that >> you mostly didn't >> get before. >> > > And mostly won't get afterward either, because the incidence of such errors > (especially after you eliminate those which conventional scrubbing will > expose) is so low. You just can't seem to grasp the fact that while this > kind of error does occur, it occurs in such insignificant numbers that > consumers *just won't care*. So while ZFS is indeed 'better' in this area, > it's just not *sufficiently* better to make any difference to consumers > (sure, once in a while one may get hit with something that ZFS could have > prevented, but for every such occurrence there'll be hundreds or thousands of > comparable problems that ZFS couldn't help at all with). > > This is getting sufficiently tedious that I'm done with this part of the > discussion unless you manage to respond with some actual substance. > > - bill > > > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss