We just picked up the fastest SSD we could in the local biccamera, which turned out to be a CSSDーSM32NI, with supposedly 95MB/s write speed.

I put it in place, and replaced the slog over:

****  0m49.173s
****  0m48.809s

So, it is slower than the CF test. This is disappointing. Everyone else seems to use Intel X25-M, which have a write-speed of 170MB/s (2nd generation) so perhaps that is why it works better for them. It is curious that it is slower than the CF card. Perhaps because it shares with so many other SATA devices?

Oh and we'll probably have to get a 3.5" frame for it, as I doubt it'll stay standing after the next earthquake. :)

Lund


Jorgen Lundman wrote:

This thread started over in nfs-discuss, as it appeared to be an nfs problem initially. Or at the very least, interaction between nfs and zil.

Just summarising speeds we have found when untarring something. Always in a new/empty directory. Only looking at write speed. read is always very fast.

The reason we started to look at this was because the 7 year old netapp being phased out, could untar the test file in 11 seconds. The x4500/x4540 Suns took 5 minutes.

For all our tests, we used MTOS-4.261-ja.tar.gz, just a random tarball I had lying around, but it can be downloaded here if you want the same test. (http://www.movabletype.org/downloads/stable/MTOS-4.261-ja.tar.gz)

The command executed generally, is:

# mkdir .test34 && time gtar --directory=.test34 -zxf /tmp/MTOS-4.261-ja.tar.gz



Solaris 10 1/06 intel client: netapp 6.5.1 FAS960 server: NFSv3
****  0m11.114s

Solaris 10 6/06 intel client: x4500 OpenSolaris svn117 server: nfsv4
****  5m11.654s

Solaris 10 6/06 intel client: x4500 Solaris 10 10/08 server: nfsv3
****  8m55.911s

Solaris 10 6/06 intel client: x4500 Solaris 10 10/08 server: nfsv4
****  10m32.629s


Just untarring the tarball on the x4500 itself:

----------------------------: x4500 OpenSolaris svn117 server
****  0m0.478s

----------------------------: x4500 Solaris 10 10/08 server
****  0m1.361s



So ZFS itself is very fast. Replacing NFS with different protocols, identical setup, just changing tar with rsync, and nfsd with sshd.

The baseline test, using:
"rsync -are ssh /tmp/MTOS-4.261-ja /export/x4500/testXX"


Solaris 10 6/06 intel client: x4500 OpenSolaris svn117 : rsync on nfsv4
****  3m44.857s

Solaris 10 6/06 intel client: x4500 OpenSolaris svn117 : rsync+ssh
****  0m1.387s

So, get rid of nfsd and it goes from 3 minutes to 1 second!

Lets share it with smb, and mount it:


OsX 10.5.6 intel client: x4500 OpenSolaris svn117 : smb+untar
****  0m24.480s


Neat, even SMB can beat nfs in default settings.

This would then indicate to me that nfsd is broken somehow, but then we try again after only disabling ZIL.


Solaris 10 6/06 : x4500 OpenSolaris svn117 DISABLE ZIL: nfsv4
****  0m8.453s
****  0m8.284s
****  0m8.264s

Nice, so this is theoretically the fastest NFS speeds we can reach? We run postfix+dovecot for mail, which probably would be safe to not use ZIL. The other type is FTP/WWW/CGI, which has more active writes/updates. Probably not as good. Comments?


Enable ZIL, but disable zfscache (Just as a test, I have been told disabling zfscache is far more dangerous).


Solaris 10 6/06 : x4500 OpenSolaris svn117 DISABLE zfscacheflush: nfsv4
****  0m45.139s

Interesting. Anyway, enable ZIL and zfscacheflush again, and learn a whole lot about slog.

First I tried creating a 2G slog on the boot mirror:


Solaris 10 6/06 : x4500 OpenSolaris svn117 slog boot pool: nfsv4

****  1m59.970s


Some improvements. For a lark, I created a 2GB file in /tmp/ and changed the slog to that. (I know, having the slog in volatile RAM is pretty much the same as disabling ZIL. But it should give me theoretical maximum speed with ZIL enabled right?).


Solaris 10 6/06 : x4500 OpenSolaris svn117 slog /tmp/junk: nfsv4
****  0m8.916s


Nice! Same speed as ZIL disabled. Since this is a X4540, we thought we would test with a CF card attached. Alas the 600X (92MB/s) card are not out until next month, rats! So, we bought a 300X (40MB/s) card.


Solaris 10 6/06 : x4500 OpenSolaris svn117 slog 300X CFFlash: nfsv4
****  0m26.566s


Not too bad really. But you have to reboot to see a CF card, fiddle with BIOS for the boot order etc. Just not an easy add on a live system. A SATA emulated SSD DISK can be hot-swapped.


Also, I learned an interesting lesson about rebooting with slog at /tmp/junk.


I am hoping to pick up a SSD SATA device today and see what speeds we get out of that.

That rsync (1s) vs nfs(8s) I can accept as over-head on a much more complicated protocol, but why would it take 3 minutes to write the same data on the same pool, with rsync(1s) vs nfs(3m)? The ZIL was on, slog is default, but both writing the same way. Does nfsd add FD_SYNC to every close regardless as to whether the application did or not?
This I have not yet wrapped my head around.

For example, I know rsync and tar does not use fdsync (but dovecot does) on its close(), but does NFS make it fdsync anyway?


Sorry for the giant email.



--
Jorgen Lundman       | <lund...@lundman.net>
Unix Administrator   | +81 (0)3 -5456-2687 ext 1017 (work)
Shibuya-ku, Tokyo    | +81 (0)90-5578-8500          (cell)
Japan                | +81 (0)3 -3375-1767          (home)
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to