On Wed, May 4, 2011 at 12:29 PM, Erik Trimble <erik.trim...@oracle.com> wrote: > I suspect that NetApp does the following to limit their resource > usage: they presume the presence of some sort of cache that can be > dedicated to the DDT (and, since they also control the hardware, they can > make sure there is always one present). Thus, they can make their code
AFAIK, NetApp has more restrictive requirements about how much data can be dedup'd on each type of hardware. See page 29 of http://media.netapp.com/documents/tr-3505.pdf - Smaller pieces of hardware can only dedup 1TB volumes, and even the big-daddy filers will only dedup up to 16TB per volume, even if the volume size is 32TB (the largest volume available for dedup). NetApp solves the problem by putting rigid constraints around the problem, whereas ZFS lets you enable dedup for any size dataset. Both approaches have limitations, and it sucks when you hit them. -B -- Brandon High : bh...@freaks.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss