Effectively performances are not so bad but they decrease a lot if you run
the same test with 2/3 instances at the same time.
With iscsi on an emc unity with sas disks, performances  are a little more
high.
But they do not decrease so much when you run the same test with 2/3
instances at the same time.
Ignazio

Il Gio 2 Gen 2020, 11:19 Sinan Polat <si...@turka.nl> ha scritto:

> Hi,
>
> Your performance is not that bad, is it? What performance do you expect?
>
> I just ran the same test.
> 12 Node, SATA SSD Only:
>    READ: bw=63.8MiB/s (66.9MB/s), 63.8MiB/s-63.8MiB/s (66.9MB/s-66.9MB/s),
> io=3070MiB (3219MB), run=48097-48097msec
>   WRITE: bw=21.3MiB/s (22.4MB/s), 21.3MiB/s-21.3MiB/s (22.4MB/s-22.4MB/s),
> io=1026MiB (1076MB), run=48097-48097msec
>
> 6 Node, SAS Only:
>    READ: bw=22.1MiB/s (23.2MB/s), 22.1MiB/s-22.1MiB/s (23.2MB/s-23.2MB/s),
> io=3070MiB (3219MB), run=138650-138650msec
>   WRITE: bw=7578KiB/s (7759kB/s), 7578KiB/s-7578KiB/s (7759kB/s-7759kB/s),
> io=1026MiB (1076MB), run=138650-138650msec
>
> This is OpenStack Queens with Ceph FileStore (Luminous).
>
> Kind regards,
> Sinan Polat
>
> > Op 2 januari 2020 om 10:59 schreef Stefan Kooman <ste...@bit.nl>:
> >
> >
> > Quoting Ignazio Cassano (ignaziocass...@gmail.com):
> > > Hello All,
> > > I installed ceph luminous with openstack, an using fio in a virtual
> machine
> > > I got slow random writes:
> > >
> > > fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1
> --name=test
> > > --filename=random_read_write.fio --bs=4k --iodepth=64 --size=4G
> > > --readwrite=randrw --rwmixread=75
> >
> > Do you use virtio-scsi with a SCSI queue per virtual CPU core? How many
> > cores do you have? I suspect that the queue depth is hampering
> > throughput here ... but is throughput performance really interesting
> > anyway for your use case? Low latency generally matters most.
> >
> > Gr. Stefan
> >
> >
> > --
> > | BIT BV  https://www.bit.nl/        Kamer van Koophandel 09090351
> > | GPG: 0xD14839C6                   +31 318 648 688 / i...@bit.nl
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@lists.ceph.com
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to