Why do you think that is slow? That's 4.5k write iops and 13.5k read iops
at the same time, that's amazing for a total of 30 HDDs.

It's actually way faster than you'd expect for 30 HDDs, so these DB devices
are really helping there :)


Paul

-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90


On Thu, Jan 2, 2020 at 12:14 PM Ignazio Cassano <ignaziocass...@gmail.com>
wrote:

> Hi Stefan, using fio with bs=64k I got very good performances.
> I am not skilled on storage, but linux file system block size is 4k.
> So, How can I modify the configuration on ceph to obtain best performances
> with bs=4k ?
> Regards
> Ignazio
>
>
>
> Il giorno gio 2 gen 2020 alle ore 10:59 Stefan Kooman <ste...@bit.nl> ha
> scritto:
>
>> Quoting Ignazio Cassano (ignaziocass...@gmail.com):
>> > Hello All,
>> > I installed ceph luminous with openstack, an using fio in a virtual
>> machine
>> > I got slow random writes:
>> >
>> > fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1
>> --name=test
>> > --filename=random_read_write.fio --bs=4k --iodepth=64 --size=4G
>> > --readwrite=randrw --rwmixread=75
>>
>> Do you use virtio-scsi with a SCSI queue per virtual CPU core? How many
>> cores do you have? I suspect that the queue depth is hampering
>> throughput here ... but is throughput performance really interesting
>> anyway for your use case? Low latency generally matters most.
>>
>> Gr. Stefan
>>
>>
>> --
>> | BIT BV  https://www.bit.nl/        Kamer van Koophandel 09090351
>> | GPG: 0xD14839C6                   +31 318 648 688 / i...@bit.nl
>>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to