Cool stuff, Robert.  It'd be interesting to see some RAID-Z (single- and
double-parity) benchmarks as well, but understandably this takes time
;-)

The first thing to note is that the current Nevada bits have a number of
performance fixes not in S10u2, so there's going to be a natural bias
when comparing ZFS to ZFS between these systems.

Second, you may be able to get more performance from the ZFS filesystem
on the HW lun by tweaking the max pending # of reqeusts.  One thing
we've found is that ZFS currently has a hardcoded limit of how many
outstanding requests to send to the underlying vdev (35).  This works
well for most single devices, but large arrays can actually handle more,
and we end up leaving some performance on the floor.  Currently the only
way to tweak this variable is through 'mdb -kw'.  Try something like:

# mdb -kw
> ::spa -v
ADDR                 STATE NAME                                                
ffffffff82ef4140    ACTIVE pool

    ADDR             STATE     AUX          DESCRIPTION                        
    ffffffff9677a1c0 HEALTHY   -            root
    ffffffff9678d080 HEALTHY   -              raidz
    ffffffff9678db80 HEALTHY   -                /dev/dsk/c2d0s0
    ffffffff96778640 HEALTHY   -                /dev/dsk/c3d0s0
    ffffffff967780c0 HEALTHY   -                /dev/dsk/c4d0s0
    ffffffff9e495780 HEALTHY   -                /dev/dsk/c5d0s0
> ffffffff9678db80::print -a vdev_t vdev_queue.vq_max_pending
ffffffff9678df00 vdev_queue.vq_max_pending = 0x23
> ffffffff9678df00/Z0t60
0xffffffff9678df00:             0x23                    =       0x3c

This will change the max # of pending requests for the disk to 60,
instead of 35.  We're trying to figure out how to tweak and/or
dynamically detect the best value here, so any more data would be
useful.

- Eric

On Mon, Aug 07, 2006 at 08:22:24AM -0700, Robert Milkowski wrote:
> Hi.
> 
> 3510 with two HW controllers, configured on LUN in RAID-10 using 12
> disks in head unit (FC-AL 73GB 15K disks). Optimization set to random,
> stripe size 32KB. Connected to v440 using two links, however in tests
> only one link was used (no MPxIO).
> 
> I used filebench and varmail test with default parameters and run for
> 60s, test was run twice.
> 
> System is S10U2 with all available patches (all support patches), kernel -18.
> 
> 
> ZFS filesystem on HW lun with atime=off:
> 
> IO Summary:      499078 ops 8248.0 ops/s, (1269/1269 r/w)  40.6mb/s,    314us 
> cpu/op,   6.0ms latency
> IO Summary:      503112 ops 8320.2 ops/s, (1280/1280 r/w)  41.0mb/s,    296us 
> cpu/op,   5.9ms latency
> 
> Now the same LUN but ZFS was destroyed and UFS filesystem was created.
> UFS filesystem on HW lun with maxcontig=24 and noatime:
> 
> IO Summary:      401671 ops 6638.2 ops/s, (1021/1021 r/w)  32.7mb/s,    404us 
> cpu/op,   7.5ms latency
> IO Summary:      403194 ops 6664.5 ops/s, (1025/1025 r/w)  32.5mb/s,    406us 
> cpu/op,   7.5ms latency
> 
> 
> 
> Now another v440 server (the same config) with snv_44, connected several 3510 
> JBODS on two FC-loops however only one loop was used (no MPxIO). The same 
> disks (73GB FC-AL 15K).
> 
> ZFS filesystem with atime=off with ZFS raid-10 using 12 disks from one 
> enclosure:
> 
> IO Summary:      558331 ops 9244.1 ops/s, (1422/1422 r/w)  45.2mb/s,    312us 
> cpu/op,   5.2ms latency
> IO Summary:      537542 ops 8899.9 ops/s, (1369/1369 r/w)  43.5mb/s,    307us 
> cpu/op,   5.4ms latency

--
Eric Schrock, Solaris Kernel Development       http://blogs.sun.com/eschrock
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to