On Tue, Sep 29, 2009 at 10:35 AM, Richard Elling <richard.ell...@gmail.com> wrote: > > On Sep 29, 2009, at 2:03 AM, Bernd Nies wrote: > >> Hi, >> >> We have a Sun Storage 7410 with the latest release (which is based upon >> opensolaris). The system uses a hybrid storage pool (23 1TB SATA disks in >> RAIDZ2 and 1 18GB SSD as log device). The ZFS volumes are exported with >> NFSv3 over TCP. NFS mount options are: >> >> rw,bg,vers=3,proto=tcp,hard,intr,rsize=32768,wsize=32768,forcedirectio >> >> We compare that system with our Netapp FAS 3140 and notice a high >> performance decrease when multiple hosts write many small files in parrallel >> (e.g. CVS checkout). >> >> Doing that on one single host, the write speed is quite similar on both >> systems: >> >> Netapp FAS 3140: >> be...@linuxhost:~/tmp> time cvs -Q checkout myBigProject >> real 0m32.914s >> user 0m1.568s >> sys 0m3.060s >> >> Sun Storage 7410: >> be...@linuxhost:/share/nightlybuild/tmp> time cvs -Q checkout >> myBigProject >> real 0m34.049s >> user 0m1.592s >> sys 0m3.184s >> >> Doing the same operation on 5 different hosts on the same NFS share in >> different directories we notice a performance decrease which is proportional >> to the number of writing hosts (5x slower) while the same operation on >> Netapp FAS 3140 is less than 2x slower: >> >> Netapp FAS 3140: >> be...@linuxhost:~/tmp/1> time cvs -Q checkout myBigProject >> real 0m58.120s >> user 0m1.452s >> sys 0m2.976s >> >> Sun Storage 7410: >> be...@linuxhost:/share/nightlybuild/tmp/1> time cvs -Q checkout >> myBigProject >> real 4m32.747s >> user 0m2.296s >> sys 0m4.224s >> >> Often we run into timeouts (CVS timeout is set to 60 minutes) when >> building software during a nightly build process which makes this storage >> unusable because the NFS writes are slowed down drastically. This happens >> also when we run VMware machines on an ESX server on a NFS pool and Oracle >> databases on NFS. Netapp and Oracle recommend using NFS as central storage >> but we wanted a less expensive system because it is used only for >> development and testing and not highly critical production data. But the >> performance slowdown when more than one writing NFS client is involved is >> too bad. >> >> What might here the bottleneck? Any ideas? The zfs log device? Are there >> more than one zfs log device required for parallel performance? As many as >> NFS clients? > > bingo! One should suffice. > > BTW, not fair comparing a machine with an NVRAM cache to one > without... add an SSD for the log to even things out. > -- richard
Not exactly true, look below at his pool configuration.... >> nfsserver# zpool status >> pool: pool-0 >> state: ONLINE >> scrub: resilver completed after 0h0m with 0 errors on Wed Sep 23 04:27:21 >> 2009 >> config: >> >> NAME STATE READ WRITE >> CKSUM >> pool-0 ONLINE 0 0 >> 0 >> raidz2 ONLINE 0 0 >> 0 >> c3t5000C50014ED4D01d0 ONLINE 0 0 >> 0 >> c3t5000C50014F4EC09d0 ONLINE 0 0 >> 0 >> c3t5000C50014F4EE46d0 ONLINE 0 0 >> 0 >> c3t5000C50014F4F50Ed0 ONLINE 0 0 >> 0 >> c3t5000C50014F4FB64d0 ONLINE 0 0 >> 0 >> c3t5000C50014F50A7Cd0 ONLINE 0 0 >> 0 >> c3t5000C50014F50F57d0 ONLINE 0 0 >> 0 >> c3t5000C50014F52A59d0 ONLINE 0 0 >> 0 >> c3t5000C50014F52D83d0 ONLINE 0 0 >> 0 >> c3t5000C50014F52E0Cd0 ONLINE 0 0 >> 0 >> c3t5000C50014F52F9Bd0 ONLINE 0 0 >> 0 >> raidz2 ONLINE 0 0 >> 0 >> c3t5000C50014F54EB1d0 ONLINE 0 0 >> 0 254K resilvered >> c3t5000C50014F54FC9d0 ONLINE 0 0 >> 0 264K resilvered >> c3t5000C50014F512E3d0 ONLINE 0 0 >> 0 264K resilvered >> c3t5000C50014F515C9d0 ONLINE 0 0 >> 0 262K resilvered >> c3t5000C50014F549EAd0 ONLINE 0 0 >> 0 262K resilvered >> c3t5000C50014F553EBd0 ONLINE 0 0 >> 0 262K resilvered >> c3t5000C50014F5072Cd0 ONLINE 0 0 >> 0 279K resilvered >> c3t5000C50014F5192Bd0 ONLINE 0 0 >> 0 4.60M resilvered >> c3t5000C50014F5494Bd0 ONLINE 0 0 >> 0 258K resilvered >> c3t5000C50014F5500Bd0 ONLINE 0 0 >> 0 264K resilvered >> c3t5000C50014F51865d0 ONLINE 0 0 >> 0 248K resilvered >> logs >> c3tATASTECZEUSIOPS018GBYTESSTM0000D905Cd0 ONLINE 0 0 >> 0 >> spares >> c3t5000C50014F53925d0 AVAIL >> >> errors: No known data errors It appears he has a: http://www.stec-inc.com/product/zeusiops.php Which should be more then capable of providing the IOPS needed. Was that pool rebuilding during the tests or did that happen afterwards? -Ross _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss