ok brandon, that's a thought. build the hash on the hard drive at the time of data creation. you mention collision resolution. so let me build my hash on the hard drive using my 6 million buckets but increase the size of each bucket from 5 slots to 20. right? i can't exactly recreate my vector/bignum hash on the hard drive because i can't dynamically resize the buckets like i can the bignums. this gives me a 4 gb file whereas my original was 1 gb. i have enough space for that so that's not a problem. so as my buckets fill up they head towards the average of 5 data items per bucket. so on average here's what happens with each hd hash record. i go to my hd hash and read 3.5 (think about it) items and 90% of the time i don't find my data so i do a write. in my process i do an initial write, then a read, a write, a read, a write. compare: 3.5 vs 2 reads; 1 vs 3 writes. the reads are more costly and if i exceed 20 items in a bucket the hd hash breaks. what do you think? is it worth it?
-- You received this message because you are subscribed to the Google Groups "Racket Users" group. To unsubscribe from this group and stop receiving emails from it, send an email to racket-users+unsubscr...@googlegroups.com. For more options, visit https://groups.google.com/d/optout.