I've been writing to the Solaris NFS list since I was getting some bad performance copying via NFS (noticeably there) a large set of small files. We have various source trees, including a tree with many linux versions that I was copying to my ZFS NAS-to-be. On large files, it flies pretty well, and "zpool iostat 1" shows interesting patterns of writes in the low k's up to 102MB/sec and down again as buffered segments apparently are synced.
However, in the numerous small file case, we see consistently only transfers in the low k's per second. First, to give some background, we are utilizing iscsi, with the backend made up a directly exposed SATA disks via the target. I've put them in a 8 disk raidz: pool: poola0 state: ONLINE scrub: none requested config: NAME STATE READ WRITE CKSUM poola0 ONLINE 0 0 0 raidz ONLINE 0 0 0 c2t1d0 ONLINE 0 0 0 c2t2d0 ONLINE 0 0 0 c2t3d0 ONLINE 0 0 0 c2t4d0 ONLINE 0 0 0 c2t5d0 ONLINE 0 0 0 c2t6d0 ONLINE 0 0 0 c2t7d0 ONLINE 0 0 0 c2t8d0 ONLINE 0 0 0 Again, I can get some great numbers on large files (doing a dd with a large blocksize screams!), but as a test, I took a problematic tree of around 1 million files, and walked it with a find/ls: bash-3.00# time find . \! -name ".*" | wc -l 987423 real 53m52.285s user 0m2.624s sys 0m27.980s That was local to the system, and not even NFS. The original files, located on a EXT3 RAID50, accessed via a linux client (NFS v3): [EMAIL PROTECTED] old-servers]# time find . \! -name ".*" | wc -l 987423 real 1m4.255s user 0m0.914s sys 0m6.976s Woe.. Something just isn't right here. Are there explicit ways I can find out what's wrong with my setup? This is from a dtrace/zdb/mdb neophyte. All I have been tracking with are zpool iostats. _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss