Nope. The ZFS head (iscsi initiator) is a Sun Ultra 20 Workstation.
The clients are RHEL4 quad opterons running the x86_64 kernel series.


On 5/4/06, Neil Perrin <[EMAIL PROTECTED]> wrote:
Actually the nfs slowness could be caused by the bug below,
but it doesn't explain the "find ." times on a local zfs.

Neil Perrin wrote On 05/04/06 21:01,:
> Was this a 32 bit intel system by chance?
> If so this is quite likely caused by:
>
> 6413731 pathologically slower fsync on 32 bit systems
>
> This was fixed in snv_39.
>
> Joe Little wrote On 05/04/06 15:47,:
>
>> I've been writing to the Solaris NFS list since I was getting some bad
>> performance copying via NFS (noticeably there) a large set of small
>> files. We have various source trees, including a tree with many linux
>> versions that I was copying to my ZFS NAS-to-be. On large files, it
>> flies pretty well, and "zpool iostat 1" shows interesting patterns of
>> writes in the low k's up to 102MB/sec and down again as buffered
>> segments apparently are synced.
>>
>> However, in the numerous small file case, we see consistently only
>> transfers in the low k's per second. First, to give some background,
>> we are utilizing iscsi, with the backend made up a directly exposed
>> SATA disks via the target. I've put them in a 8 disk raidz:
>>
>>   pool: poola0
>>  state: ONLINE
>>  scrub: none requested
>> config:
>>
>>         NAME        STATE     READ WRITE CKSUM
>>         poola0      ONLINE       0     0     0
>>           raidz     ONLINE       0     0     0
>>             c2t1d0  ONLINE       0     0     0
>>             c2t2d0  ONLINE       0     0     0
>>             c2t3d0  ONLINE       0     0     0
>>             c2t4d0  ONLINE       0     0     0
>>             c2t5d0  ONLINE       0     0     0
>>             c2t6d0  ONLINE       0     0     0
>>             c2t7d0  ONLINE       0     0     0
>>             c2t8d0  ONLINE       0     0     0
>>
>> Again, I can get some great numbers on large files (doing a dd with a
>> large blocksize screams!), but as a test, I took a problematic tree of
>> around 1 million files, and walked it with a find/ls:
>>
>> bash-3.00# time find . \! -name ".*" | wc -l
>>   987423
>>
>> real    53m52.285s
>> user    0m2.624s
>> sys     0m27.980s
>>
>> That was local to the system, and not even NFS.
>>
>> The original files, located on a EXT3 RAID50, accessed via a linux
>> client (NFS v3):
>> [EMAIL PROTECTED] old-servers]# time find . \! -name ".*" | wc -l
>> 987423
>>
>> real    1m4.255s
>> user    0m0.914s
>> sys     0m6.976s
>>
>> Woe.. Something just isn't right here. Are there explicit ways I can
>> find out what's wrong with my setup? This is from a dtrace/zdb/mdb
>> neophyte. All I have been tracking with are zpool iostats.
>> _______________________________________________
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>

--

Neil

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to