On Wed, Jul 18, 2018 at 1:08 PM Nikola Ciprich <nikola.cipr...@linuxbox.cz>
wrote:

> > Care to share your "bench-rbd" script (on pastebin or similar)?
> sure, no problem.. it's so short I hope nobody will get offended if I
> paste it right
> here :)
>
> #!/bin/bash
>
> #export LD_PRELOAD="/usr/lib64/libtcmalloc.so.4"
> numjobs=8
> pool=nvme
> vol=xxx
> time=30
>
> opts="--randrepeat=1 --ioengine=rbd --direct=1 --numjobs=${numjobs}
> --gtod_reduce=1 --name=test --pool=${pool} --rbdname=${vol} --invalidate=0
> --bs=4k --iodepth=64 --time_based --runtime=$time --group_reporting"
>

So that "--numjobs" parameter is what I was referring to when I said
multiple jobs will cause a huge performance it. This causes fio to open the
same image X images, so with (nearly) each write operation, the
exclusive-lock is being moved from client-to-client. Instead of multiple
jobs against the same image, you should use multiple images.


> sopts="--randrepeat=1 --ioengine=rbd --direct=1 --numjobs=1
> --gtod_reduce=1 --name=test --pool=${pool} --rbdname=${vol} --invalidate=0
> --bs=256k --iodepth=64 --time_based --runtime=$time --group_reporting"
>
> #fio $sopts --readwrite=read --output=rbd-fio-seqread.log
> echo
>
> #fio $sopts --readwrite=write --output=rbd-fio-seqwrite.log
> echo
>
> fio $opts --readwrite=randread --output=rbd-fio-randread.log
> echo
>
> fio $opts --readwrite=randwrite --output=rbd-fio-randwrite.log
> echo
>
>
> hope it's of some use..
>
> n.
>
>
> --
> -------------------------------------
> Ing. Nikola CIPRICH
> LinuxBox.cz, s.r.o.
> 28. rijna 168, 709 00 Ostrava
>
> tel.:   +420 591 166 214
> fax:    +420 596 621 273
> mobil:  +420 777 093 799
>
> www.linuxbox.cz
>
> mobil servis: +420 737 238 656
> email servis: ser...@linuxbox.cz
> -------------------------------------
>


-- 
Jason
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to