On 9/11/07, Gino <[EMAIL PROTECTED]> wrote:

> -ZFS performs badly with a lot of small files.
> (about 20 times slower that UFS with our millions file rsync procedures)

        We have seen just the opposite... we have a server with about
40 million files and only 4 TB of data. We have been benchmarking FSes
for creation and manipulation of large populations of small files and
ZFS is the only one we have found that continues to scale linearly
above one million files in one FS. UFS, VXFS, HFS+ (don't ask why),
NSS (on NW not Linux) all show exponential growth in response time as
you cross a certain knee (we are graphing time to create <n> zero
length files, then do a series of basic manipulations on them) in
number of files. For all the FSes we have tested that knee has been
under one million files, except for ZFS. I know this is not 'real
world' but it does reflect the response time issues we have been
trying to solve. I will see if my client (I am a consultant) will
allow me to post the results, as I am under NDA for most of the
details of what we are doing.

        On the other hand, we have seen serious issues using rsync to
migrate this data from the existing server to the Solaris 10 / ZFS
system, so perhaps your performance issues were rsync related and not
ZFS. In fact, so far the fastest and most reliable method for moving
the data is proving to be Veritas NetBackup (back it up on the source
server, restore to the new ZFS server).

        Now having said all that, we are probably never going to see
100 million files in one zpool, because the ZFS architecture lets us
use a more distributed model (many zpools and datasets within them)
and still present the end users with a single view of all the data.

-- 
Paul Kraus
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to