Hi Phil

Good notice , but believe me this is not the problem .

I just cut & pasted the filebench "config" file after I rerun it with in
shorter time, the orginal one was SAME with 100 seconds.

and by the way I always enmpty the cache with zfs export & import before
every run

Filebench still behaving weird with this benchmark, I ran this benchmark for
over 50 times with the incompatible results !!!!

Thanks for reading the output carefully ...



On Tue, Mar 2, 2010 at 4:53 AM, Phil Harman <phil.har...@gmail.com> wrote:

> I see at least two differences:
>
> 1. duration 30s vs 100s (so not "SAME")
>
> 2. your manual test doesn't empty the cache
>
> Of course, it is the latter that makes all the difference.
>
> Hope this helps,
> Phil
>
> Sent from my iPhone
>
> On 2 Mar 2010, at 08:38, Abdullah Al-Dahlawi <dahl...@ieee.org> wrote:
>
> Greeting All
>
> I am using Filebench benchmark in an "Interactive mode" to test ZFS
> performance with randomread wordload.
> My Filebench setting & run results are as follwos
>
> ------------------------------------------------------------------------------------------
> filebench> set $filesize=5g
> filebench> set $dir=/hdd/fs32k
> filebench> set $iosize=32k
> filebench> set $workingset=10m
> filebench> set $function=generic
> filebench> set $filesystem=zfs
> filebench> run 100
>  1062: 106.866: Creating/pre-allocating files and filesets
>  1062: 106.867: File largefile1: mbytes=5120
>  1062: 106.867: Re-using file largefile1.
>  1062: 106.867: Creating file largefile1...
>  1062: 108.612: Preallocated 1 of 1 of file largefile1 in 2 seconds
>  1062: 108.612: waiting for fileset pre-allocation to finish
>  1062: 108.612: Starting 1 rand-read instances
>  1063: 109.617: Starting 1 rand-thread threads
>  1062: 112.627: Running...
>  1062: 213.627: Run took 100 seconds...
>  1062: 213.628: Per-Operation Breakdown
> rand-rate                   0ops/s   0.0mb/s      0.0ms/op
> 0us/op-cpu
> rand-read1              41845ops/s *1307.7mb/s*      0.0ms/op
> 20us/op-cpu
>
>  1062: 213.628:
> IO Summary:      4226337 ops, 41845.0 ops/s, (41845/0 r/w) 1307.7mb/s,
> 21us cpu/op,   0.0ms latency
>  1062: 213.628: Shutting down processes
>
> ---------------------------------------------------------------------------------------------
> The output looks GREAT so far .... notice the 1307.7 mb/s
>
> **** HOWEVER *****
>
> When I run the SAME workload using Filebench "config file" in batch mode,
> the performance dropped significantly !!!!!!!!
>
> Here is my config file & filebench results.
>
>
> # Copyright 2007 Sun Microsystems, Inc.  All rights reserved.
> # Use is subject to license terms.
> #
> # ident    "%Z%%M%    %I%    %E% SMI"
>
> DEFAULTS {
>     runtime = 30;
>         dir = /hdd/fs32k;
>         $statsdir=/export/home/abdullah/bench.stat/woow87;
>         stats = /export/home/abdullah/bench.stat;
>     filesystem = zfs;
>     description = "ZFS-RR-WS-10M";
> }
>
> CONFIG rr32k {
>     function = generic;
>     personality = randomread;
>     filesize = 5g;
>     iosize = 32k;
>     nthreads = 1;
>      workingset=10m;
> }
>
> And the Run result ....
>
> abdul...@hp_hdx_16:/usr/benchmarks/filebench/config# filebench rrws10m
> parsing profile for config: rr32k
> Creating Client Script
> /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k/thisrun.f
> Running
> /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k/thisrun.f
> FileBench Version 1.4.4
>  1147: 0.004: Random Read Version 2.0 IO personality successfully loaded
>  1147: 0.004: Creating/pre-allocating files and filesets
>  1147: 0.005: File largefile1: mbytes=5120
>  1147: 0.005: Re-using file largefile1.
>  1147: 0.005: Creating file largefile1...
>  1147: 1.837: Preallocated 1 of 1 of file largefile1 in 2 seconds
>  1147: 1.837: waiting for fileset pre-allocation to finish
>  1147: 1.837: Running '/usr/benchmarks/filebench/scripts/fs_flush zfs
> /hdd/fs32k'
>  1147: 1.845: Change dir to
> /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k
>  1147: 1.845: Starting 1 rand-read instances
>  1149: 2.850: Starting 1 rand-thread threads
>  1147: 5.860: Running...
>  1147: 36.159: Run took 30 seconds...
>  1147: 36.160: Per-Operation Breakdown
> rand-rate                   0ops/s   0.0mb/s      0.0ms/op
> 0us/op-cpu
> rand-read1                 88ops/s   *2.7mb/s** *    11.4ms/op
> 35us/op-cpu
>
>  1147: 36.160:
> IO Summary:       2660 ops,  87.8 ops/s, (88/0 r/w)   2.7mb/s,    443us
> cpu/op,  11.4ms latency
>  1147: 36.160: Stats dump to file 'stats.rr32k.out'
>  1147: 36.160: in statsdump stats.rr32k.out
>  1147: 36.415: Shutting down processes
> Generating html for
> /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s
> file =
> /export/home/abdullah/bench.stat/HP_HDX_16-zfs-rrws10m-Mar_2_2010-03h_10m_46s/rr32k/stats.rr32k.out
>
> ------------------------------------------------------------------------------------------------
>
> The output for the same workload is disappointing , notice that the
> throughput dropped from 1307.7 mb/s to 2.7 mb/s !!!!!!!!!!!!!!!!!!!!!!1
>
> My ARC_max is 3G
>
> Here is a snapshot of my arcstat output in case of high throughput ---
> notice the 100% hits ratio
>
>
>
> arcsz,read,hits,Hit%,miss,miss%,dhit,dh%,dmis,dm%,phit,ph%,pmis,pm%,mhit,mh%,mmis,mm%,mfug,mrug,
>    1G, 31M, 31M,  99,111K,    0, 28M, 99, 99K,  0,  2M, 99, 12K,  0,  1M,
> 98, 13K,  1,  43,  43,
>    1G,147K,145K,  99,  1K,    0, 14K, 99,   2,  0,131K, 99,  1K,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,166K,166K, 100,   0,    0, 37K,100,   0,  0,128K,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 41K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 41K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,
> 10,100,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 41K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 41K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 42K, 42K, 100,   0,    0, 42K,100,   0,  0, 256,100,   0,  0,   0,
> 0,   0,  0,   0,   0,
>
> and a snapshot in case of low throughput
> notice the low hit ratio !!
>
> arcsz,read,hits,Hit%,miss,miss%,dhit,dh%,dmis,dm%,phit,ph%,pmis,pm%,mhit,mh%,mmis,mm%,mfug,mrug,
>    1G,   3,   3, 100,   0,    0,   3,100,   0,  0,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,   0,   0,   0,   0,    0,   0,  0,   0,  0,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,   0,   0,   0,   0,    0,   0,  0,   0,  0,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,   0,   0,   0,   0,    0,   0,  0,   0,  0,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,  40,   3,   7,  37,   92,   3,  7,  37, 92,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 113,  12,  10, 101,   89,  12, 10, 101, 89,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 105,  14,  13,  91,   86,  14, 13,  91, 86,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 108,  15,  13,  93,   86,  15, 13,  93, 86,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,  99,  11,  11,  88,   88,  11, 11,  88, 88,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 103,  11,  10,  92,   89,  11, 10,  92, 89,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 101,  13,  12,  88,   87,  13, 12,  88, 87,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 107,  12,  11,  95,   88,  12, 11,  95, 88,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,  99,  12,  12,  87,   87,  12, 12,  87, 87,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 100,   5,   5,  95,   95,   5,  5,  95, 95,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 114,  17,  14,  97,   85,  17, 14,  97, 85,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 106,  17,  16,  89,   83,  17, 16,  89, 83,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 107,   7,   6, 100,   93,   7,  6, 100, 93,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 100,  11,  11,  89,   89,  11, 11,  89, 89,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G,  99,   8,   8,  91,   91,   8,  8,  91, 91,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 101,   9,   8,  92,   91,   9,  8,  92, 91,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>    1G, 101,   9,   8,  92,   91,   9,  8,  92, 91,   0,  0,   0,  0,   0,
> 0,   0,  0,   0,   0,
>
>
> Any Feed Back !!!!!!!!!!!!
>
> --
> Abdullah Al-Dahlawi
> PhD Candidate
> George Washington University
> Department. Of Electrical & Computer Engineering
> ----
> Check The Fastest 500 Super Computers Worldwide
> <http://www.top500.org/list/2009/11/100>
> http://www.top500.org/list/2009/11/100
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>


-- 
Abdullah Al-Dahlawi
PhD Candidate
George Washington University
Department. Of Electrical & Computer Engineering
----
Check The Fastest 500 Super Computers Worldwide
http://www.top500.org/list/2009/11/100
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to