> From: Jim Klimov [mailto:jimkli...@cos.ru] > Sent: Thursday, July 12, 2012 8:42 AM > To: Edward Ned Harvey > Subject: Re: [zfs-discuss] New fast hash algorithm - is it needed? > > 2012-07-11 18:03, Edward Ned Harvey пишет: > >> From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- > >> boun...@opensolaris.org] On Behalf Of Sašo Kiselkov > >> > >> As your dedup > >> ratio grows, so does the performance hit from dedup=verify. At, say, > >> dedupratio=10.0x, on average, every write results in 10 reads. > > > > Why? > > If you intend to write a block, and you discover it has a matching checksum, > > you only need to verify it matches one block. You don't need to verify it > > matches all the other blocks that have previously been verified to match > > each other. > > As Saso explained, if you wrote the same block 10 times > and detected it was already deduped, then by verifying > this detection 10 times you get about 10 extra reads.
(In this case, Jim, you wrote me off-list and I replied on-list, but in this case, I thought it would be ok because this message doesn't look private or exclusive to me. I apologize if I was wrong.) I get the miscommunication now - When you write the duplicate block for the 10th time, we all understand you're not going to go back and verify 10 blocks. (It seemed, at least to me, that's what Saso was saying. Which is why I told him, No you don't.) You're saying, that when you wrote the duplicate block the 2nd time, you verified... And then when you wrote it the 3rd time, you verified... And the 4th time, and the 5th time... By the time you write the 10th time, you have already verified 9 previous times, but you're still going to verify again. Normally you would expect writing duplicate blocks dedup'd to be faster than writing the non-dedup'd data, because you get to skip the actual write. (This idealistic performance gain might be pure vapor due to need to update metadata, but ignoring that technicality, continue hypothetically...) When you verify, supposedly you're giving up that hypothetical performance gain, because you have to do a read instead of a write. So at first blush, it seems like no net-gain for performance. But if the duplicate block gets written frequently (for example a block of all zeros) then there's a high probability the duplicate block is already in ARC, so you can actually skip reading from disk and just read from RAM. So, the "10 extra reads" will sometimes be true - if the duplicate block doesn't already exist in ARC. And the "10 extra reads" will sometimes be false - if the duplicate block is already in ARC. _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss