I see at least two differences:
1. duration 30s vs 100s (so not "SAME")
2. your manual test doesn't empty the cache
Of course, it is the latter that makes all the difference.
Hope this helps,
Phil
Sent from my iPhone
On 2 Mar 2010, at 08:38, Abdullah Al-Dahlawi <dahl...@ieee.org> wrote:
Greeting All
I am using Filebench benchmark in an "Interactive mode" to test ZFS
performance with randomread wordload.
My Filebench setting & run results are as follwos
---
---
---
---
---
---
---
---------------------------------------------------------------------
filebench> set $filesize=5g
filebench> set $dir=/hdd/fs32k
filebench> set $iosize=32k
filebench> set $workingset=10m
filebench> set $function=generic
filebench> set $filesystem=zfs
filebench> run 100
1062: 106.866: Creating/pre-allocating files and filesets
1062: 106.867: File largefile1: mbytes=5120
1062: 106.867: Re-using file largefile1.
1062: 106.867: Creating file largefile1...
1062: 108.612: Preallocated 1 of 1 of file largefile1 in 2 seconds
1062: 108.612: waiting for fileset pre-allocation to finish
1062: 108.612: Starting 1 rand-read instances
1063: 109.617: Starting 1 rand-thread threads
1062: 112.627: Running...
1062: 213.627: Run took 100 seconds...
1062: 213.628: Per-Operation Breakdown
rand-rate 0ops/s 0.0mb/s 0.0ms/op
0us/op-cpu
rand-read1 41845ops/s 1307.7mb/s 0.0ms/op
20us/op-cpu
1062: 213.628:
IO Summary: 4226337 ops, 41845.0 ops/s, (41845/0 r/w) 1307.7mb/
s, 21us cpu/op, 0.0ms latency
1062: 213.628: Shutting down processes
---
---
---
---
---
---
---
---
---------------------------------------------------------------------
The output looks GREAT so far .... notice the 1307.7 mb/s
**** HOWEVER *****
When I run the SAME workload using Filebench "config file" in batch
mode, the performance dropped significantly !!!!!!!!
Here is my config file & filebench results.
# Copyright 2007 Sun Microsystems, Inc. All rights reserved.
# Use is subject to license terms.
#
# ident "%Z%%M% %I% %E% SMI"
DEFAULTS {
runtime = 30;
dir = /hdd/fs32k;
$statsdir=/export/home/abdullah/bench.stat/woow87;
stats = /export/home/abdullah/bench.stat;
filesystem = zfs;
description = "ZFS-RR-WS-10M";
}
CONFIG rr32k {
function = generic;
personality = randomread;
filesize = 5g;
iosize = 32k;
nthreads = 1;
workingset=10m;
}
And the Run result ....
abdul...@hp_hdx_16:/usr/benchmarks/filebench/config# filebench rrws10m
parsing profile for config: rr32k
Creating Client Script /export/home/abdullah/bench.stat/HP_HDX_16-
zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k/thisrun.f
Running /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-
Mar_2_2010-03h_10m_46s/rr32k/thisrun.f
FileBench Version 1.4.4
1147: 0.004: Random Read Version 2.0 IO personality successfully
loaded
1147: 0.004: Creating/pre-allocating files and filesets
1147: 0.005: File largefile1: mbytes=5120
1147: 0.005: Re-using file largefile1.
1147: 0.005: Creating file largefile1...
1147: 1.837: Preallocated 1 of 1 of file largefile1 in 2 seconds
1147: 1.837: waiting for fileset pre-allocation to finish
1147: 1.837: Running '/usr/benchmarks/filebench/scripts/fs_flush
zfs /hdd/fs32k'
1147: 1.845: Change dir to /export/home/abdullah/bench.stat/
HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k
1147: 1.845: Starting 1 rand-read instances
1149: 2.850: Starting 1 rand-thread threads
1147: 5.860: Running...
1147: 36.159: Run took 30 seconds...
1147: 36.160: Per-Operation Breakdown
rand-rate 0ops/s 0.0mb/s 0.0ms/op
0us/op-cpu
rand-read1 88ops/s 2.7mb/s 11.4ms/op
35us/op-cpu
1147: 36.160:
IO Summary: 2660 ops, 87.8 ops/s, (88/0 r/w) 2.7mb/s,
443us cpu/op, 11.4ms latency
1147: 36.160: Stats dump to file 'stats.rr32k.out'
1147: 36.160: in statsdump stats.rr32k.out
1147: 36.415: Shutting down processes
Generating html for /export/home/abdullah/bench.stat/HP_HDX_16-zfs-
rrws10m-Mar_2_2010-03h_10m_46s
file = /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-
Mar_2_2010-03h_10m_46s/rr32k/stats.rr32k.out
---
---
---
---
---
---
---
---
---
---------------------------------------------------------------------
The output for the same workload is disappointing , notice that the
throughput dropped from 1307.7 mb/s to 2.7 mb/
s !!!!!!!!!!!!!!!!!!!!!!1
My ARC_max is 3G
Here is a snapshot of my arcstat output in case of high throughput ---
notice the 100% hits ratio
arcsz,read,hits,Hit%,miss,miss%,dhit,dh%,dmis,dm%,phit,ph%,pmis,pm
%,mhit,mh%,mmis,mm%,mfug,mrug,
1G, 31M, 31M, 99,111K, 0, 28M, 99, 99K, 0, 2M, 99, 12K,
0, 1M, 98, 13K, 1, 43, 43,
1G,147K,145K, 99, 1K, 0, 14K, 99, 2, 0,131K, 99, 1K,
0, 0, 0, 0, 0, 0, 0,
1G,166K,166K, 100, 0, 0, 37K,100, 0, 0,128K,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 41K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 41K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 10,100, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 41K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 41K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 42K, 42K, 100, 0, 0, 42K,100, 0, 0, 256,100, 0,
0, 0, 0, 0, 0, 0, 0,
and a snapshot in case of low throughput
notice the low hit ratio !!
arcsz,read,hits,Hit%,miss,miss%,dhit,dh%,dmis,dm%,phit,ph%,pmis,pm
%,mhit,mh%,mmis,mm%,mfug,mrug,
1G, 3, 3, 100, 0, 0, 3,100, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 40, 3, 7, 37, 92, 3, 7, 37, 92, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 113, 12, 10, 101, 89, 12, 10, 101, 89, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 105, 14, 13, 91, 86, 14, 13, 91, 86, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 108, 15, 13, 93, 86, 15, 13, 93, 86, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 99, 11, 11, 88, 88, 11, 11, 88, 88, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 103, 11, 10, 92, 89, 11, 10, 92, 89, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 101, 13, 12, 88, 87, 13, 12, 88, 87, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 107, 12, 11, 95, 88, 12, 11, 95, 88, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 99, 12, 12, 87, 87, 12, 12, 87, 87, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 100, 5, 5, 95, 95, 5, 5, 95, 95, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 114, 17, 14, 97, 85, 17, 14, 97, 85, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 106, 17, 16, 89, 83, 17, 16, 89, 83, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 107, 7, 6, 100, 93, 7, 6, 100, 93, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 100, 11, 11, 89, 89, 11, 11, 89, 89, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 99, 8, 8, 91, 91, 8, 8, 91, 91, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 101, 9, 8, 92, 91, 9, 8, 92, 91, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
1G, 101, 9, 8, 92, 91, 9, 8, 92, 91, 0, 0, 0,
0, 0, 0, 0, 0, 0, 0,
Any Feed Back !!!!!!!!!!!!
--
Abdullah Al-Dahlawi
PhD Candidate
George Washington University
Department. Of Electrical & Computer Engineering
----
Check The Fastest 500 Super Computers Worldwide
http://www.top500.org/list/2009/11/100
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss