Hi, Maged Not a big difference in both cases.
Performance of 4 nodes pool with 5x PM863a each is: 4k bs - 33-37kIOPS krbd 128 threads and 42-51kIOPS vs 1024 threads (fio numjobs 128-256-512) the same situation happens when we try to increase rbd workload, 3 rbd gets the same iops #. Dead end & limit ) Thank you! 2018-03-12 21:49 GMT+03:00 Maged Mokhtar <mmokh...@petasan.org>: > Hi, > > Try increasing the queue depth from default 128 to 1024: > > rbd map image-XX -o queue_depth=1024 > > > Also if you run multiple rbd images/fio tests, do you get higher combined > performance ? > > Maged > > > On 2018-03-12 17:16, Sergey Kotov wrote: > > Dear moderator, i subscribed to ceph list today, could you please post my > message? > > ---------- Forwarded message ---------- > From: Sergey Kotov <grayc...@gmail.com> > Date: 2018-03-06 10:52 GMT+03:00 > Subject: [ceph bad performance], can't find a bottleneck > To: ceph-users@lists.ceph.com > Cc: Житенев Алексей <zhitene...@wipline.ru>, Anna Anikina < > anik...@gmail.com> > > > Good day. > > Can you please help us to find bottleneck in our ceph installations. > We have 3 SSD-only clusters with different HW, but situation is the same - > overall i/o operations between client & ceph lower than 1/6 of summary > performance all ssd. > > For example - > One of our cluster has 4-nodes with ssd Toshiba 2Tb Enterprise drives, > installed on Ubuntu server 16.04. > Servers are connected to the 10G switches. Latency between modes is about > 0.1ms. Ethernet utilisation is low. > > # uname -a > Linux storage01 4.4.0-101-generic #124-Ubuntu SMP Fri Nov 10 18:29:59 UTC > 2017 x86_64 x86_64 x86_64 GNU/Linux > > # ceph osd versions > { > "ceph version 12.2.1 (3e7492b9ada8bdc9a5cd0feafd42fbca27f9c38e) > luminous (stable)": 55 > } > > > When we map rbd image direct on the storage nodes via krbd, performance is > not good enough. > We use fio for testing. Even we run randwrite with 4k block size test in > multiple thread mode, our drives don't have utilisation higher then 30% and > lat is ok. > > At the same time iostat tool displays 100% utilisation on /dev/rbdX. > > Also we can't enable rbd_cache, because of using scst iscsi over rbd > mapped images. > > How can we resolve the issue? > > Ceph config: > > [global] > fsid = beX482fX-6a91-46dX-ad22-21a8a2696abX > mon_initial_members = storage01, storage02, storage03 > mon_host = X.Y.Z.1,X.Y.Z.2,X.Y.Z.3 > auth_cluster_required = cephx > auth_service_required = cephx > auth_client_required = cephx > public_network = X.Y.Z.0/24 > filestore_xattr_use_omap = true > osd_pool_default_size = 2 > osd_pool_default_min_size = 1 > osd_pool_default_pg_num = 1024 > osd_journal_size = 10240 > osd_mkfs_type = xfs > filestore_op_threads = 16 > filestore_wbthrottle_enable = False > throttler_perf_counter = False > osd crush update on start = false > > [osd] > osd_scrub_begin_hour = 1 > osd_scrub_end_hour = 6 > osd_scrub_priority = 1 > > osd_enable_op_tracker = False > osd_max_backfills = 1 > osd heartbeat grace = 20 > osd heartbeat interval = 5 > osd recovery max active = 1 > osd recovery max single start = 1 > osd recovery op priority = 1 > osd recovery threads = 1 > osd backfill scan max = 16 > osd backfill scan min = 4 > osd max scrubs = 1 > osd scrub interval randomize ratio = 1.0 > osd disk thread ioprio class = idle > osd disk thread ioprio priority = 0 > osd scrub chunk max = 1 > osd scrub chunk min = 1 > osd deep scrub stride = 1048576 > osd scrub load threshold = 5.0 > osd scrub sleep = 0.1 > > [client] > rbd_cache = false > > > Sample fio tests: > > root@storage04:~# fio --name iops --rw randread --bs 4k --filename > /dev/rbd2 --numjobs 12 --ioengine=libaio --group_reporting > iops: (g=0): rw=randread, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=1 > ... > fio-2.2.10 > Starting 12 processes > ^Cbs: 12 (f=12): [r(12)] [1.2% done] [128.4MB/0KB/0KB /s] [32.9K/0/0 iops] > [eta 16m:49s] > fio: terminating on signal 2 > > iops: (groupid=0, jobs=12): err= 0: pid=29812: Sun Feb 11 23:59:19 2018 > read : io=1367.8MB, bw=126212KB/s, iops=31553, runt= 11097msec > slat (usec): min=1, max=59700, avg=375.92, stdev=495.19 > clat (usec): min=0, max=377, avg= 1.12, stdev= 3.16 > lat (usec): min=1, max=59702, avg=377.61, stdev=495.32 > clat percentiles (usec): > | 1.00th=[ 0], 5.00th=[ 0], 10.00th=[ 1], 20.00th=[ 1], > | 30.00th=[ 1], 40.00th=[ 1], 50.00th=[ 1], 60.00th=[ 1], > | 70.00th=[ 1], 80.00th=[ 1], 90.00th=[ 1], 95.00th=[ 2], > | 99.00th=[ 2], 99.50th=[ 2], 99.90th=[ 73], 99.95th=[ 78], > | 99.99th=[ 115] > bw (KB /s): min= 8536, max=11944, per=8.33%, avg=10516.45, > stdev=635.32 > lat (usec) : 2=91.74%, 4=7.93%, 10=0.14%, 20=0.09%, 50=0.01% > lat (usec) : 100=0.07%, 250=0.03%, 500=0.01% > cpu : usr=1.32%, sys=3.69%, ctx=329556, majf=0, minf=134 > IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=350144/w=0/d=0, short=r=0/w=0/d=0, > drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=1 > > Run status group 0 (all jobs): > READ: io=1367.8MB, aggrb=126212KB/s, minb=126212KB/s, maxb=126212KB/s, > mint=11097msec, maxt=11097msec > > Disk stats (read/write): > rbd2: ios=323072/0, merge=0/0, ticks=124268/0, in_queue=124680, > util=99.31% > > > root@storage04:~# fio --name iops --rw randwrite --bs 4k --filename > /dev/rbd2 --numjobs 12 --ioengine=libaio --group_reporting > iops: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=1 > ... > fio-2.2.10 > Starting 12 processes > ^Cbs: 12 (f=12): [w(12)] [25.0% done] [0KB/713.5MB/0KB /s] [0/183K/0 iops] > [eta 00m:45s] > fio: terminating on signal 2 > fio: job 'iops' hasn't exited in 60 seconds, it appears to be stuck. Doing > forceful exit of this job. > > iops: (groupid=0, jobs=12): err= 0: pid=30327: Mon Feb 12 00:01:04 2018 > write: io=32020MB, bw=2172.5MB/s, iops=556149, runt= 14739msec > slat (usec): min=1, max=255011, avg=20.00, stdev=1110.12 > clat (usec): min=0, max=4426, avg= 0.35, stdev= 2.32 > lat (usec): min=1, max=255015, avg=20.43, stdev=1110.18 > clat percentiles (usec): > | 1.00th=[ 0], 5.00th=[ 0], 10.00th=[ 0], 20.00th=[ 0], > | 30.00th=[ 0], 40.00th=[ 0], 50.00th=[ 0], 60.00th=[ 0], > | 70.00th=[ 1], 80.00th=[ 1], 90.00th=[ 1], 95.00th=[ 1], > | 99.00th=[ 1], 99.50th=[ 1], 99.90th=[ 2], 99.95th=[ 3], > | 99.99th=[ 81] > bw (KB /s): min=16895, max=641688, per=8.68%, avg=193110.20, > stdev=150217.09 > lat (usec) : 2=99.88%, 4=0.08%, 10=0.01%, 20=0.01%, 50=0.01% > lat (usec) : 100=0.01%, 250=0.01%, 500=0.01%, 750=0.01%, 1000=0.01% > lat (msec) : 10=0.01% > cpu : usr=5.51%, sys=23.27%, ctx=8723, majf=0, minf=3528 > IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=8197085/d=0, short=r=0/w=0/d=0, > drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=1 > > Run status group 0 (all jobs): > WRITE: io=32020MB, aggrb=2172.5MB/s, minb=2172.5MB/s, maxb=2172.5MB/s, > mint=14739msec, maxt=14739msec > > Disk stats (read/write): > rbd2: ios=0/30044, merge=0/330950, ticks=0/1601972, in_queue=1603328, > util=87.57% > > > root@storage04:~# fio --name iops --rw randwrite --bs 4k --filename > /dev/rbd2 --numjobs 12 --ioengine=libaio --group_reporting --direct=1 > iops: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, iodepth=1 > ... > fio-2.2.10 > Starting 12 processes > ^Cbs: 12 (f=12): [w(12)] [0.5% done] [0KB/14304KB/0KB /s] [0/3576/0 iops] > [eta 03h:42m:23s]] > fio: terminating on signal 2 > > iops: (groupid=0, jobs=12): err= 0: pid=31552: Mon Feb 12 00:02:39 2018 > write: io=617132KB, bw=9554.8KB/s, iops=2388, runt= 64589msec > slat (usec): min=3, max=463, avg=11.56, stdev=14.17 > clat (usec): min=930, max=999957, avg=5004.58, stdev=30299.43 > lat (usec): min=938, max=999975, avg=5016.50, stdev=30299.58 > clat percentiles (usec): > | 1.00th=[ 1192], 5.00th=[ 1304], 10.00th=[ 1400], 20.00th=[ 1512], > | 30.00th=[ 1608], 40.00th=[ 1720], 50.00th=[ 1832], 60.00th=[ 1960], > | 70.00th=[ 2160], 80.00th=[ 2480], 90.00th=[ 4256], 95.00th=[12096], > | 99.00th=[51968], 99.50th=[81408], 99.90th=[733184], > 99.95th=[864256], > | 99.99th=[946176] > bw (KB /s): min= 4, max= 2251, per=9.30%, avg=888.88, stdev=453.38 > lat (usec) : 1000=0.02% > lat (msec) : 2=62.38%, 4=26.86%, 10=5.16%, 20=1.82%, 50=2.66% > lat (msec) : 100=0.75%, 250=0.22%, 500=0.03%, 750=0.01%, 1000=0.09% > cpu : usr=0.24%, sys=0.54%, ctx=154351, majf=0, minf=149 > IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=154283/d=0, short=r=0/w=0/d=0, > drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=1 > > Run status group 0 (all jobs): > WRITE: io=617132KB, aggrb=9554KB/s, minb=9554KB/s, maxb=9554KB/s, > mint=64589msec, maxt=64589msec > > Disk stats (read/write): > rbd2: ios=0/154277, merge=0/0, ticks=0/767860, in_queue=768740, > util=99.94% > > > root@storage04:~# fio --name iops --rw randwrite --bs 4k --filename > /dev/rbd2 --numjobs 12 --iodepth=32 --ioengine=libaio --group_reporting > --direct=1 --runtime=100 > iops: (g=0): rw=randwrite, bs=4K-4K/4K-4K/4K-4K, ioengine=libaio, > iodepth=32 > ... > fio-2.2.10 > Starting 12 processes > Jobs: 12 (f=12): [w(12)] [100.0% done] [0KB/49557KB/0KB /s] [0/12.4K/0 > iops] [eta 00m:00s] > iops: (groupid=0, jobs=12): err= 0: pid=32632: Mon Feb 12 00:04:51 2018 > write: io=3093.6MB, bw=31675KB/s, iops=7918, runt=100009msec > slat (usec): min=2, max=973942, avg=1510.77, stdev=18783.09 > clat (msec): min=1, max=1719, avg=46.97, stdev=111.64 > lat (msec): min=1, max=1719, avg=48.49, stdev=113.62 > clat percentiles (msec): > | 1.00th=[ 5], 5.00th=[ 8], 10.00th=[ 10], 20.00th=[ 13], > | 30.00th=[ 16], 40.00th=[ 18], 50.00th=[ 21], 60.00th=[ 25], > | 70.00th=[ 31], 80.00th=[ 43], 90.00th=[ 68], 95.00th=[ 131], > | 99.00th=[ 742], 99.50th=[ 824], 99.90th=[ 963], 99.95th=[ 988], > | 99.99th=[ 1045] > bw (KB /s): min= 6, max= 8320, per=9.51%, avg=3011.84, > stdev=1956.64 > lat (msec) : 2=0.07%, 4=0.55%, 10=10.28%, 20=37.47%, 50=35.89% > lat (msec) : 100=9.47%, 250=3.45%, 500=0.61%, 750=1.26%, 1000=0.92% > lat (msec) : 2000=0.03% > cpu : usr=0.28%, sys=0.58%, ctx=265641, majf=0, minf=146 > IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=100.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.1%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=791939/d=0, short=r=0/w=0/d=0, > drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=32 > > Run status group 0 (all jobs): > WRITE: io=3093.6MB, aggrb=31674KB/s, minb=31674KB/s, maxb=31674KB/s, > mint=100009msec, maxt=100009msec > > Disk stats (read/write): > rbd2: ios=0/789529, merge=0/0, ticks=0/12345428, in_queue=12383232, > util=100.00% > > > > root@storage04:~# fio --name iops --rw randwrite --bs 21k --filename > /dev/rbd2 --numjobs 64 --iodepth=4 --ioengine=libaio --group_reporting > --direct=1 --runtime=100 > iops: (g=0): rw=randwrite, bs=21K-21K/21K-21K/21K-21K, ioengine=libaio, > iodepth=4 > ... > fio-2.2.10 > Starting 64 processes > ^Cbs: 64 (f=64): [w(64)] [28.7% done] [0KB/98028KB/0KB /s] [0/4668/0 iops] > [eta 01m:12s] > fio: terminating on signal 2 > Jobs: 9 (f=9): [E(1),w(1),_(2),w(1),_(3),w(2) > ,_(2),w(1),_(15),w(1),_(7),w(1),_(13),w(1),_(2),w(1),_(10)] [1.9% done] > [0KB/4804KB/0KB /s] [0/228/0 iops] [eta 27m:16s] > iops: (groupid=0, jobs=64): err= 0: pid=5221: Mon Feb 12 00:10:52 2018 > write: io=1967.9MB, bw=67393KB/s, iops=3209, runt= 29900msec > slat (usec): min=3, max=1211.3K, avg=14387.91, stdev=58236.34 > clat (msec): min=1, max=1711, avg=63.67, stdev=141.68 > lat (msec): min=1, max=1711, avg=78.06, stdev=159.64 > clat percentiles (usec): > | 1.00th=[ 1944], 5.00th=[ 2864], 10.00th=[ 4192], 20.00th=[ 7136], > | 30.00th=[11200], 40.00th=[16768], 50.00th=[23680], 60.00th=[33536], > | 70.00th=[46848], 80.00th=[66048], 90.00th=[110080], > 95.00th=[205824], > | 99.00th=[790528], 99.50th=[872448], 99.90th=[1253376], 99.95th=[ > 1368064], > | 99.99th=[1531904] > bw (KB /s): min= 23, max= 3906, per=1.77%, avg=1189.80, stdev=837.25 > lat (msec) : 2=1.25%, 4=8.01%, 10=18.08%, 20=17.69%, 50=27.03% > lat (msec) : 100=16.47%, 250=6.99%, 500=0.94%, 750=2.15%, 1000=1.20% > lat (msec) : 2000=0.19% > cpu : usr=0.06%, sys=0.11%, ctx=173237, majf=0, minf=787 > IO depths : 1=0.1%, 2=0.1%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=95955/d=0, short=r=0/w=0/d=0, drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=4 > > Run status group 0 (all jobs): > WRITE: io=1967.9MB, aggrb=67393KB/s, minb=67393KB/s, maxb=67393KB/s, > mint=29900msec, maxt=29900msec > > Disk stats (read/write): > rbd2: ios=0/95945, merge=0/0, ticks=0/3672960, in_queue=3685732, > util=99.66% > > > root@storage04:~# fio --name iops --rw randwrite --bs 21k --filename > /dev/rbd2 --numjobs 32 --iodepth=4 --ioengine=libaio --group_reporting > --direct=1 --runtime=100 > iops: (g=0): rw=randwrite, bs=21K-21K/21K-21K/21K-21K, ioengine=libaio, > iodepth=4 > ... > fio-2.2.10 > Starting 32 processes > ^Cbs: 32 (f=32): [w(32)] [11.9% done] [0KB/40866KB/0KB /s] [0/1946/0 iops] > [eta 01m:29s] > fio: terminating on signal 2 > Jobs: 32 (f=30): [w(32)] [12.9% done] [0KB/53676KB/0KB /s] [0/2556/0 iops] > [eta 01m:28s] > iops: (groupid=0, jobs=32): err= 0: pid=6344: Mon Feb 12 00:12:01 2018 > write: io=820680KB, bw=66912KB/s, iops=3186, runt= 12265msec > slat (usec): min=3, max=842, avg=13.14, stdev=15.18 > clat (msec): min=1, max=1270, avg=38.67, stdev=114.69 > lat (msec): min=1, max=1270, avg=38.68, stdev=114.69 > clat percentiles (usec): > | 1.00th=[ 1976], 5.00th=[ 2416], 10.00th=[ 3024], 20.00th=[ 4384], > | 30.00th=[ 6048], 40.00th=[ 8032], 50.00th=[10688], 60.00th=[14400], > | 70.00th=[19584], 80.00th=[30080], 90.00th=[57088], 95.00th=[112128], > | 99.00th=[765952], 99.50th=[839680], 99.90th=[954368], > 99.95th=[987136], > | 99.99th=[1089536] > bw (KB /s): min= 30, max= 6300, per=3.58%, avg=2397.65, > stdev=1548.34 > lat (msec) : 2=1.16%, 4=16.01%, 10=30.33%, 20=22.95%, 50=17.40% > lat (msec) : 100=6.82%, 250=1.96%, 500=1.46%, 750=0.82%, 1000=1.07% > lat (msec) : 2000=0.03% > cpu : usr=0.13%, sys=0.21%, ctx=38649, majf=0, minf=372 > IO depths : 1=0.1%, 2=0.2%, 4=99.8%, 8=0.0%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=39080/d=0, short=r=0/w=0/d=0, drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=4 > > Run status group 0 (all jobs): > WRITE: io=820680KB, aggrb=66912KB/s, minb=66912KB/s, maxb=66912KB/s, > mint=12265msec, maxt=12265msec > > Disk stats (read/write): > rbd2: ios=0/39077, merge=0/0, ticks=0/1504048, in_queue=1507272, > util=99.17% > > > > root@storage04:~# fio --name iops --rw randwrite --bs 21k --filename > /dev/rbd2 --numjobs 8 --iodepth=12 --ioengine=libaio --group_reporting > --direct=1 --runtime=100 > iops: (g=0): rw=randwrite, bs=21K-21K/21K-21K/21K-21K, ioengine=libaio, > iodepth=12 > ... > fio-2.2.10 > Starting 8 processes > ^Cbs: 8 (f=8): [w(8)] [20.8% done] [0KB/34650KB/0KB /s] [0/1650/0 iops] > [eta 01m:20s] > fio: terminating on signal 2 > Jobs: 2 (f=2): [_(5),w(2),_(1)] [10.1% done] [0KB/1554KB/0KB /s] [0/74/0 > iops] [eta 03m:24s] > iops: (groupid=0, jobs=8): err= 0: pid=6770: Mon Feb 12 00:12:32 2018 > write: io=979755KB, bw=44563KB/s, iops=2122, runt= 21986msec > slat (usec): min=3, max=551, avg=12.92, stdev=14.78 > clat (msec): min=1, max=1531, avg=44.14, stdev=136.44 > lat (msec): min=1, max=1531, avg=44.15, stdev=136.44 > clat percentiles (usec): > | 1.00th=[ 1944], 5.00th=[ 2320], 10.00th=[ 2672], 20.00th=[ 3568], > | 30.00th=[ 4768], 40.00th=[ 6240], 50.00th=[ 8640], 60.00th=[12224], > | 70.00th=[18816], 80.00th=[31872], 90.00th=[67072], 95.00th=[154624], > | 99.00th=[839680], 99.50th=[913408], 99.90th=[1073152], 99.95th=[ > 1368064], > | 99.99th=[1466368] > bw (KB /s): min= 26, max=18023, per=14.47%, avg=6449.72, > stdev=5032.09 > lat (msec) : 2=1.37%, 4=22.49%, 10=30.69%, 20=16.94%, 50=15.12% > lat (msec) : 100=6.72%, 250=3.13%, 500=0.73%, 750=1.14%, 1000=1.41% > lat (msec) : 2000=0.25% > cpu : usr=0.32%, sys=0.61%, ctx=44124, majf=0, minf=97 > IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=99.9%, 16=0.0%, 32=0.0%, > >=64=0.0% > submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, > >=64=0.0% > complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.1%, 32=0.0%, 64=0.0%, > >=64=0.0% > issued : total=r=0/w=46655/d=0, short=r=0/w=0/d=0, drop=r=0/w=0/d=0 > latency : target=0, window=0, percentile=100.00%, depth=12 > > Run status group 0 (all jobs): > WRITE: io=979755KB, aggrb=44562KB/s, minb=44562KB/s, maxb=44562KB/s, > mint=21986msec, maxt=21986msec > > Disk stats (read/write): > rbd2: ios=0/46652, merge=0/0, ticks=0/2037428, in_queue=2043236, > util=99.65% > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com