>> On Dec 1, 2013, at 8:00 AM, Scott Ribe <scott_r...@elevated-dev.com> wrote:
> 
> 
>> On Dec 1, 2013, at 8:36 AM, Graham Cox <graham....@bigpond.com> wrote:
>> 
>> Scanning my entire hard drive (excluding hidden files), which took several 
>> hours, sure I had plenty of collisions - but absolutely no false ones - they 
>> all turned out to be genuine duplicates of existing files. This is using the 
>> FNV-1a 64-bit hash + length approach.
> 
> I have a drive sitting here that has a few *million* image files; I'd be 
> willing to bet zero collisions.

That’s all well and good, but why are we still debating this approach when Mike 
Abdullah has posted a far superior scheme with an infinitely faster hash (file 
size), *zero* risk of data loss, and less performance penalty than using a 
low-collision hashing algorithm?

I still find it unconscionable to handle the user’s data with any scheme that 
could potentially lose it under normal operation, regardless of how 
infinitesimal the chance. You just don’t do it, *especially* not silently! Just 
because git does it doesn’t make it okay.

--Kyle Sluder
_______________________________________________

Cocoa-dev mailing list (Cocoa-dev@lists.apple.com)

Please do not post admin requests or moderator comments to the list.
Contact the moderators at cocoa-dev-admins(at)lists.apple.com

Help/Unsubscribe/Update your Subscription:
https://lists.apple.com/mailman/options/cocoa-dev/archive%40mail-archive.com

This email sent to arch...@mail-archive.com

Reply via email to