On 13/04/2011 00:36, David Magda wrote:
On Apr 11, 2011, at 17:54, Brandon High wrote:

I suspect that the minimum memory for most moderately sized pools is
over 16GB. There has been a lot of discussion regarding how much
memory each dedup'd block requires, and I think it was about 250-270
bytes per block. 1TB of data (at max block size and no duplicate data)
will require about 2GB of memory to run effectively. (This seems high
to me, hopefully someone else can confirm.)

There was a  thread on the topic with the subject "Newbie ZFS Question: RAM for 
Dedup". I think it summarized pretty well by Erik Trimble:

bottom line: 270 bytes per record

so, for 4k record size, that  works out to be 67GB per 1 TB of unique data. 
128k record size means about 2GB per 1 TB.

dedup means buy a (big) SSD for L2ARC.

http://mail.opensolaris.org/pipermail/zfs-discuss/2010-October/045720.html

Remember that 270 bytes per block means you're allocating one 512-byte sector 
for most current disks (a 4K sector for each block  RSN).

See also:

http://mail.opensolaris.org/pipermail/zfs-discuss/2010-March/037978.html
http://mail.opensolaris.org/pipermail/zfs-discuss/2010-February/037300.html


Thanks for the info guys.

I decided that the overhead involved in managing (esp deleting) deduped datasets far outweighed the benefits it was bringing me. I'm currently remaking datasets non-dedup and now I know about the "hang", I am a lot more patient :D

Thanks

Matt
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to