Hi,

After comparison we found there is nothing much difference between format 1
and format 2.
format 1 is even worse for randrw.

format 1 result:

# sysbench --test=fileio --file-total-size=5G --file-test-mode=rndrw
--init-rng=on --max-time=300 --max-requests=0  run
sysbench 0.4.12:  multi-threaded system evaluation benchmark

Running the test with following options:
Number of threads: 1
Initializing random number generator from timer.


Extra file open flags: 0
128 files, 40Mb each
5Gb total file size
Block size 16Kb
Number of random requests for random IO: 0
Read/Write ratio for combined random IO test: 1.50
Periodic FSYNC enabled, calling fsync() each 100 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test
Threads started!

Time limit exceeded, exiting...
Done.

Operations performed:  6240 Read, 4160 Write, 13301 Other = 23701 Total
Read 97.5Mb  Written 65Mb  Total transferred 162.5Mb  (554.64Kb/sec)
   34.67 Requests/sec executed

Test execution summary:
    total time:                          300.0118s
    total number of events:              10400
    total time taken by event execution: 8.3638
    per-request statistics:
         min:                                  0.01ms
         avg:                                  0.80ms
         max:                                181.63ms
         approx.  95 percentile:               1.72ms

Threads fairness:
    events (avg/stddev):           10400.0000/0.00
    execution time (avg/stddev):   8.3638/0.00



format 2 result:

# sysbench --test=fileio --file-total-size=5G --file-test-mode=rndrw
--init-rng=on --max-time=300 --max-requests=0  run
sysbench 0.4.12:  multi-threaded system evaluation benchmark

Running the test with following options:
Number of threads: 1
Initializing random number generator from timer.


Extra file open flags: 0
128 files, 40Mb each
5Gb total file size
Block size 16Kb
Number of random requests for random IO: 0
Read/Write ratio for combined random IO test: 1.50
Periodic FSYNC enabled, calling fsync() each 100 requests.
Calling fsync() at the end of test, Enabled.
Using synchronous I/O mode
Doing random r/w test
Threads started!
Time limit exceeded, exiting...
Done.

Operations performed:  7260 Read, 4840 Write, 15363 Other = 27463 Total
Read 113.44Mb  Written 75.625Mb  Total transferred 189.06Mb  (645.15Kb/sec)
   40.32 Requests/sec executed

Test execution summary:
    total time:                          300.0843s
    total number of events:              12100
    total time taken by event execution: 9.8130
    per-request statistics:
         min:                                  0.01ms
         avg:                                  0.81ms
         max:                                209.24ms
         approx.  95 percentile:               1.64ms

Threads fairness:
    events (avg/stddev):           12100.0000/0.00
    execution time (avg/stddev):   9.8130/0.00


2016-05-25 15:31 GMT+08:00 Adrian Saul <adrian.s...@tpgtelecom.com.au>:

>
>
> Are you using image-format 2 RBD images?
>
>
>
> We found a major performance hit using format 2 images under 10.2.0 today
> in some testing.  When we switched to using format 1 images we literally
> got 10x random write IOPS performance (1600 IOPs up to 30000 IOPS for the
> same test).
>
>
>
>
>
>
>
> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf
> Of *Ken Peng
> *Sent:* Wednesday, 25 May 2016 5:02 PM
> *To:* ceph-users@lists.ceph.com
> *Subject:* [ceph-users] seqwrite gets good performance but random rw gets
> worse
>
>
>
> Hello,
>
> We have a cluster with 20+ hosts and 200+ OSDs, each 4T SATA disk for an
> OSD, no SSD cache.
>
> OS is Ubuntu 16.04 LTS, ceph version 10.2.0
>
> Both data network and cluster network are 10Gbps.
>
> We run ceph as block storage service only (rbd client within VM).
>
> For testing within a VM with sysbench tool, we see that the seqwrite has a
> relatively good performance, it can reach 170.37Mb/sec, but random
> read/write always gets bad result, it can be only 474.63Kb/sec (shown as
> below).
>
> Can you help give the idea why the random IO is so worse? Thanks.
>
> This is what sysbench outputs,
>
> # sysbench --test=fileio --file-total-size=5G prepare
> sysbench 0.4.12:  multi-threaded system evaluation benchmark
>
> 128 files, 40960Kb each, 5120Mb total
> Creating files for the test...
>
>
> # sysbench --test=fileio --file-total-size=5G --file-test-mode=seqwr
> --init-rng=on --max-time=300 --max-requests=0 run
> sysbench 0.4.12:  multi-threaded system evaluation benchmark
>
> Running the test with following options:
> Number of threads: 1
> Initializing random number generator from timer.
>
>
> Extra file open flags: 0
> 128 files, 40Mb each
> 5Gb total file size
> Block size 16Kb
> Periodic FSYNC enabled, calling fsync() each 100 requests.
> Calling fsync() at the end of test, Enabled.
> Using synchronous I/O mode
> Doing sequential write (creation) test
> Threads started!
> Done.
>
> Operations performed:  0 Read, 327680 Write, 128 Other = 327808 Total
> Read 0b  Written 5Gb  Total transferred 5Gb  (170.37Mb/sec)
> 10903.42 Requests/sec executed
>
> Test execution summary:
>     total time:                          30.0530s
>     total number of events:              327680
>     total time taken by event execution: 28.5936
>     per-request statistics:
>          min:                                  0.01ms
>          avg:                                  0.09ms
>          max:                                192.84ms
>          approx.  95 percentile:               0.03ms
>
> Threads fairness:
>     events (avg/stddev):           327680.0000/0.00
>     execution time (avg/stddev):   28.5936/0.00
>
>
>
> # sysbench --test=fileio --file-total-size=5G --file-test-mode=rndrw
> --init-rng=on --max-time=300 --max-requests=0 run
> sysbench 0.4.12:  multi-threaded system evaluation benchmark
>
> Running the test with following options:
> Number of threads: 1
> Initializing random number generator from timer.
>
>
> Extra file open flags: 0
> 128 files, 40Mb each
> 5Gb total file size
> Block size 16Kb
> Number of random requests for random IO: 0
> Read/Write ratio for combined random IO test: 1.50
> Periodic FSYNC enabled, calling fsync() each 100 requests.
> Calling fsync() at the end of test, Enabled.
> Using synchronous I/O mode
> Doing random r/w test
> Threads started!
>
> Time limit exceeded, exiting...
> Done.
>
> Operations performed:  5340 Read, 3560 Write, 11269 Other = 20169 Total
> Read 83.438Mb  Written 55.625Mb  Total transferred 139.06Mb  (474.63Kb/sec)
>    29.66 Requests/sec executed
>
> Test execution summary:
>     total time:                          300.0216s
>     total number of events:              8900
>     total time taken by event execution: 6.4774
>     per-request statistics:
>          min:                                  0.01ms
>          avg:                                  0.73ms
>          max:                                 90.18ms
>          approx.  95 percentile:               1.60ms
>
> Threads fairness:
>     events (avg/stddev):           8900.0000/0.00
>     execution time (avg/stddev):   6.4774/0.00
> Confidentiality: This email and any attachments are confidential and may
> be subject to copyright, legal or some other professional privilege. They
> are intended solely for the attention and use of the named addressee(s).
> They may only be copied, distributed or disclosed with the consent of the
> copyright owner. If you have received this email by mistake or by breach of
> the confidentiality clause, please notify the sender immediately by return
> email and delete or destroy all copies of the email. Any confidentiality,
> privilege or copyright is not waived or lost because this email has been
> sent to you by mistake.
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to