Hi all As I've said here on the list a few times earlier, the last on the thread 'ZFS not usable (was ZFS Dedup question)', I've been doing some rather thorough testing on zfs dedup, and as you can see from the posts, it wasn't very satisfactory. The docs claim 1-2GB memory usage per terabyte stored, ARC or L2ARC, but as you can read from the post, I don't find this very likely.
So, is there anyone in here using dedup for large storage (2TB? 10TB? more?) and can document sustained high performance? The reason I ask, is if this is the case, something is badly wrong with my test setup. The test box is a supermicro thing with a Core2duo CPU, 8 gigs of RAM, 4 gigs of mirrored SLOG and some 150 gigs of L2ARC on 80GB x25-M drives. The data drives are 7 2TB drives in RAIDz2. We're getting down to 10-20MB/s on Bacula backup to this system, meaning streaming, which should be good for RAIDz2. Since the writes are local (bacula-sd running), async writes will be the main thing. Initial results show pretty good I/O perfrmance, but after about 2TB used, the I/O speed is down to the numbers I mentioned PS: I know those drives aren't optimal for this, but the box is a year old or so. Still, they should help out a bit. Vennlige hilsener / Best regards roy -- Roy Sigurd Karlsbakk (+47) 97542685 r...@karlsbakk.net http://blogg.karlsbakk.net/ -- I all pedagogikk er det essensielt at pensum presenteres intelligibelt. Det er et elementært imperativ for alle pedagoger å unngå eksessiv anvendelse av idiomer med fremmed opprinnelse. I de fleste tilfeller eksisterer adekvate og relevante synonymer på norsk. _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss