On Tue, Sep 29, 2009 at 10:35 AM, Richard Elling
<richard.ell...@gmail.com> wrote:
>
> On Sep 29, 2009, at 2:03 AM, Bernd Nies wrote:
>
>> Hi,
>>
>> We have a Sun Storage 7410 with the latest release (which is based upon
>> opensolaris). The system uses a hybrid storage pool (23 1TB SATA disks in
>> RAIDZ2 and 1 18GB SSD as log device). The ZFS volumes are exported with
>> NFSv3 over TCP. NFS mount options are:
>>
>> rw,bg,vers=3,proto=tcp,hard,intr,rsize=32768,wsize=32768,forcedirectio
>>
>> We compare that system with our Netapp FAS 3140 and notice a high
>> performance decrease when multiple hosts write many small files in parrallel
>> (e.g. CVS checkout).
>>
>> Doing that on one single host, the write speed is quite similar on both
>> systems:
>>
>> Netapp FAS 3140:
>>   be...@linuxhost:~/tmp> time cvs -Q checkout myBigProject
>>   real    0m32.914s
>>   user    0m1.568s
>>   sys     0m3.060s
>>
>> Sun Storage 7410:
>>   be...@linuxhost:/share/nightlybuild/tmp> time cvs -Q checkout
>> myBigProject
>>   real    0m34.049s
>>   user    0m1.592s
>>   sys     0m3.184s
>>
>> Doing the same operation on 5 different hosts on the same NFS share in
>> different directories we notice a performance decrease which is proportional
>> to the number of writing hosts (5x slower) while the same operation on
>> Netapp FAS 3140 is less than 2x slower:
>>
>> Netapp FAS 3140:
>>   be...@linuxhost:~/tmp/1> time cvs -Q checkout myBigProject
>>   real    0m58.120s
>>   user    0m1.452s
>>   sys     0m2.976s
>>
>> Sun Storage 7410:
>>   be...@linuxhost:/share/nightlybuild/tmp/1> time cvs -Q checkout
>> myBigProject
>>   real    4m32.747s
>>   user    0m2.296s
>>   sys     0m4.224s
>>
>> Often we run into timeouts (CVS timeout is set to 60 minutes) when
>> building software during a nightly build process which makes this storage
>> unusable because the NFS writes are slowed down drastically. This happens
>> also when we run VMware machines on an ESX server on a NFS pool and Oracle
>> databases on NFS. Netapp and Oracle recommend using NFS as central storage
>> but we wanted a less expensive system because it is used only for
>> development and testing and not highly critical production data. But the
>> performance slowdown when more than one writing NFS client is involved is
>> too bad.
>>
>> What might here the bottleneck? Any ideas? The zfs log device? Are there
>> more than one zfs log device required for parallel performance? As many as
>> NFS clients?
>
> bingo!  One should suffice.
>
> BTW, not fair comparing a machine with an NVRAM cache to one
> without... add an SSD for the log to even things out.
>  -- richard

Not exactly true, look below at his pool configuration....

>> nfsserver# zpool status
>>  pool: pool-0
>> state: ONLINE
>> scrub: resilver completed after 0h0m with 0 errors on Wed Sep 23 04:27:21
>> 2009
>> config:
>>
>>       NAME                                         STATE     READ WRITE
>> CKSUM
>>       pool-0                                       ONLINE       0     0
>>   0
>>         raidz2                                     ONLINE       0     0
>>   0
>>           c3t5000C50014ED4D01d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F4EC09d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F4EE46d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F4F50Ed0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F4FB64d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F50A7Cd0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F50F57d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F52A59d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F52D83d0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F52E0Cd0                    ONLINE       0     0
>>   0
>>           c3t5000C50014F52F9Bd0                    ONLINE       0     0
>>   0
>>         raidz2                                     ONLINE       0     0
>>   0
>>           c3t5000C50014F54EB1d0                    ONLINE       0     0
>>   0  254K resilvered
>>           c3t5000C50014F54FC9d0                    ONLINE       0     0
>>   0  264K resilvered
>>           c3t5000C50014F512E3d0                    ONLINE       0     0
>>   0  264K resilvered
>>           c3t5000C50014F515C9d0                    ONLINE       0     0
>>   0  262K resilvered
>>           c3t5000C50014F549EAd0                    ONLINE       0     0
>>   0  262K resilvered
>>           c3t5000C50014F553EBd0                    ONLINE       0     0
>>   0  262K resilvered
>>           c3t5000C50014F5072Cd0                    ONLINE       0     0
>>   0  279K resilvered
>>           c3t5000C50014F5192Bd0                    ONLINE       0     0
>>   0  4.60M resilvered
>>           c3t5000C50014F5494Bd0                    ONLINE       0     0
>>   0  258K resilvered
>>           c3t5000C50014F5500Bd0                    ONLINE       0     0
>>   0  264K resilvered
>>           c3t5000C50014F51865d0                    ONLINE       0     0
>>   0  248K resilvered
>>       logs
>>         c3tATASTECZEUSIOPS018GBYTESSTM0000D905Cd0  ONLINE       0     0
>>   0
>>       spares
>>         c3t5000C50014F53925d0                      AVAIL
>>
>> errors: No known data errors

It appears he has a:

http://www.stec-inc.com/product/zeusiops.php

Which should be more then capable of providing the IOPS needed.

Was that pool rebuilding during the tests or did that happen afterwards?

-Ross
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to