Joe Little writes: > On 6/22/06, Bill Moore <[EMAIL PROTECTED]> wrote: > > Hey Joe. We're working on some ZFS changes in this area, and if you > > could run an experiment for us, that would be great. Just do this: > > > > echo 'zil_disable/W1' | mdb -kw > > > > We're working on some fixes to the ZIL so it won't be a bottleneck when > > fsyncs come around. The above command will let us know what kind of > > improvement is on the table. After our fixes you could get from 30-80% > > of that improvement, but this would be a good data point. This change > > makes ZFS ignore the iSCSI/NFS fsync requests, but we still push out a > > txg every 5 seconds. So at most, your disk will be 5 seconds out of > > date compared to what it should be. It's a pretty small window, but it > > all depends on your appetite for such windows. :) > > > > After running the above command, you'll need to unmount/mount the > > filesystem in order for the change to take effect. > > > > If you don't have time, no big deal. > > > > > > --Bill > > > > > > On Thu, Jun 22, 2006 at 04:22:22PM -0700, Joe Little wrote: > > > On 6/22/06, Jeff Bonwick <[EMAIL PROTECTED]> wrote: > > > >> a test against the same iscsi targets using linux and XFS and the > > > >> NFS server implementation there gave me 1.25MB/sec writes. I was about > > > >> to throw in the towel and deem ZFS/NFS has unusable until B41 came > > > >> along and at least gave me 1.25MB/sec. > > > > > > > >That's still super slow -- is this over a 10Mb link or something? > > > > > > > >Jeff
I think the performance is in line with expectation for, small file, single threaded, open/write/close NFS workload (nfs must commit on close). Therefore I expect : (avg file size) / (I/O latency). Joe does this formula approach the 1.25 MB/s ? > > > > > > > > > > > > > > Nope, gig-e link (single e1000g, or aggregate, doesn't matter) to the > > > iscsi target, and single gig-e link (nge) to the NFS clients, who are > > > gig-e. Sun Ultra20 or AMD Quad Opteron, again with no difference. > > > > > > Again, the issue is the multiple fsyncs that NFS requires, and likely > > > the serialization of those iscsi requests. Apparently, there is a > > > basic latency in iscsi that one could improve upon with FC, but we are > > > definitely in the all ethernet/iscsi camp for multi-building storage > > > pool growth and don't have interest in a FC-based SAN. > > > _______________________________________________ > > > zfs-discuss mailing list > > > zfs-discuss@opensolaris.org > > > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > > > > Well, following Bill's advice and the previous note on disabling zil, > I ran my test on a B38 opteron initiator and if you do a time on the > copy from the client, 6250 8k files transfer at 6MB/sec now. If you > watch the entire commit on the backend using "zpool iostat 1" I see > that it takes a few more seconds, and the actual rate there is > 4MB/sec. Beats my best of 1.25MB/sec, and this is not B41. > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss Joe, you know this but for the benefit of others, I have to highlight that running any NFS server this way, may cause silent data corruption from client's point of view. Whenever a server keeps data in RAM this way and does not commit it to stable storage upon request from clients, that opens a time window for corruption. So a client writes to a page, then reads the same page, and if the server suffered a crash in between, the data may not match. So this is performance at the expense of data integrity. -r _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss