Effectively performances are not so bad but they decrease a lot if you run the same test with 2/3 instances at the same time. With iscsi on an emc unity with sas disks, performances are a little more high. But they do not decrease so much when you run the same test with 2/3 instances at the same time. Ignazio
Il Gio 2 Gen 2020, 11:19 Sinan Polat <si...@turka.nl> ha scritto: > Hi, > > Your performance is not that bad, is it? What performance do you expect? > > I just ran the same test. > 12 Node, SATA SSD Only: > READ: bw=63.8MiB/s (66.9MB/s), 63.8MiB/s-63.8MiB/s (66.9MB/s-66.9MB/s), > io=3070MiB (3219MB), run=48097-48097msec > WRITE: bw=21.3MiB/s (22.4MB/s), 21.3MiB/s-21.3MiB/s (22.4MB/s-22.4MB/s), > io=1026MiB (1076MB), run=48097-48097msec > > 6 Node, SAS Only: > READ: bw=22.1MiB/s (23.2MB/s), 22.1MiB/s-22.1MiB/s (23.2MB/s-23.2MB/s), > io=3070MiB (3219MB), run=138650-138650msec > WRITE: bw=7578KiB/s (7759kB/s), 7578KiB/s-7578KiB/s (7759kB/s-7759kB/s), > io=1026MiB (1076MB), run=138650-138650msec > > This is OpenStack Queens with Ceph FileStore (Luminous). > > Kind regards, > Sinan Polat > > > Op 2 januari 2020 om 10:59 schreef Stefan Kooman <ste...@bit.nl>: > > > > > > Quoting Ignazio Cassano (ignaziocass...@gmail.com): > > > Hello All, > > > I installed ceph luminous with openstack, an using fio in a virtual > machine > > > I got slow random writes: > > > > > > fio --randrepeat=1 --ioengine=libaio --direct=1 --gtod_reduce=1 > --name=test > > > --filename=random_read_write.fio --bs=4k --iodepth=64 --size=4G > > > --readwrite=randrw --rwmixread=75 > > > > Do you use virtio-scsi with a SCSI queue per virtual CPU core? How many > > cores do you have? I suspect that the queue depth is hampering > > throughput here ... but is throughput performance really interesting > > anyway for your use case? Low latency generally matters most. > > > > Gr. Stefan > > > > > > -- > > | BIT BV https://www.bit.nl/ Kamer van Koophandel 09090351 > > | GPG: 0xD14839C6 +31 318 648 688 / i...@bit.nl > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com