Hello, did you test the performance of your individual drives?
Here is a small snippet: ----------------- DRIVE=/dev/XXX smartctl --a $DRIVE for i in 1 2 4 8 16; do echo "Test $i"; fio --filename=$DRIVE --direct=1 --sync=1 --rw=write --bs=4k --numjobs=$i --iodepth=1 --runtime=60 --time_based --group_reporting --name=journal-test; done ----------------- Please share the results that we know what's possible with your hardware. -- Martin Verges Managing director Mobile: +49 174 9335695 E-Mail: martin.ver...@croit.io Chat: https://t.me/MartinVerges croit GmbH, Freseniusstr. 31h, 81247 Munich CEO: Martin Verges - VAT-ID: DE310638492 Com. register: Amtsgericht Munich HRB 231263 Web: https://croit.io YouTube: https://goo.gl/PGE1Bx Vitaliy Filippov <vita...@yourcmc.ru> schrieb am Sa., 9. März 2019, 21:09: > There are 2: > > fio -ioengine=rbd -direct=1 -name=test -bs=4k -iodepth=1 -rw=randwrite > -pool=bench -rbdname=testimg > > fio -ioengine=rbd -direct=1 -name=test -bs=4k -iodepth=128 -rw=randwrite > -pool=bench -rbdname=testimg > > The first measures your min possible latency - it does not scale with the > number of OSDs at all, but it's usually what real applications like > DBMSes > need. > > The second measures your max possible random write throughput which you > probably won't be able to utilize if you don't have enough VMs all > writing > in parallel. > > -- > With best regards, > Vitaliy Filippov > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com