On 1/28/2011 2:24 PM, Roy Sigurd Karlsbakk wrote:
I created a zfs pool with dedup with the following settings:
zpool create data c8t1d0
zfs create data/shared
zfs set dedup=on data/shared

The thing I was wondering about was it seems like ZFS only dedup at
the file level and not the block. When I make multiple copies of a
file to the store I see an increase in the deup ratio, but when I copy
similar files the ratio stays at 1.00x.
I've done some rather intensive tests on zfs dedup on this 12TB test system we 
have. I have concluded that with some 150B worth of L2ARC and 8GB ARC, ZFS 
dedup is unusable for volumes even at 2TB storage. It works, but it's dead slow 
in write terms, and the time to remove a dataset is still very long. I wouldn't 
recommend using ZFS dedup unless your name were Ahmed Nazif or Silvio 
Berlusconi, where the damage might be used for some good.

Vennlige hilsener / Best regards

roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--

If you want Dedup to perform well, you *absolutely* must have a L2ARC device which can hold the *entire* Dedup Table. Remember, the size of the DDT is not dependent on the size of your data pool, but in the number of zfs slabs which are contained in that pool (slab = record, for this purpose). Thus, 12TB worth of DVD iso images (record size about 128k) will consume 256 times less DDT space as will 12TB filled with text configuration files (average record size < 512b).

And, I doubt 8GB for ARC is sufficient, either, for a DDT consuming over 100GB of space.





--
Erik Trimble
Java System Support
Mailstop:  usca22-123
Phone:  x17195
Santa Clara, CA

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to