On 1/28/2011 2:24 PM, Roy Sigurd Karlsbakk wrote:
I created a zfs pool with dedup with the following settings:
zpool create data c8t1d0
zfs create data/shared
zfs set dedup=on data/shared
The thing I was wondering about was it seems like ZFS only dedup at
the file level and not the block. When I make multiple copies of a
file to the store I see an increase in the deup ratio, but when I copy
similar files the ratio stays at 1.00x.
I've done some rather intensive tests on zfs dedup on this 12TB test system we
have. I have concluded that with some 150B worth of L2ARC and 8GB ARC, ZFS
dedup is unusable for volumes even at 2TB storage. It works, but it's dead slow
in write terms, and the time to remove a dataset is still very long. I wouldn't
recommend using ZFS dedup unless your name were Ahmed Nazif or Silvio
Berlusconi, where the damage might be used for some good.
Vennlige hilsener / Best regards
roy
--
Roy Sigurd Karlsbakk
(+47) 97542685
r...@karlsbakk.net
http://blogg.karlsbakk.net/
--
If you want Dedup to perform well, you *absolutely* must have a L2ARC
device which can hold the *entire* Dedup Table. Remember, the size of
the DDT is not dependent on the size of your data pool, but in the
number of zfs slabs which are contained in that pool (slab = record, for
this purpose). Thus, 12TB worth of DVD iso images (record size about
128k) will consume 256 times less DDT space as will 12TB filled with
text configuration files (average record size < 512b).
And, I doubt 8GB for ARC is sufficient, either, for a DDT consuming over
100GB of space.
--
Erik Trimble
Java System Support
Mailstop: usca22-123
Phone: x17195
Santa Clara, CA
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss