One thing to keep in mind is that the blockdb/wal becomes a Single Point Of Failure for all OSDs using it. So if that SSD dies essentially you have to consider all OSDs using it as lost. I think most go with something like 4-8 OSDs per blockdb/wal drive but it really depends how risk-averse you are, what your budget is etc. Given that you only have 5 nodes I'd probably go for fewer OSDs per blockdb device.
On Sat, 6 Jul 2019 at 02:16, Davis Mendoza Paco <davis.men...@gmail.com> wrote: > Hi all, > I have installed ceph luminous, witch 5 nodes(45 OSD), each OSD server > supports up to 16HD and I'm only using 9 > > I wanted to ask for help to improve IOPS performance since I have about > 350 virtual machines of approximately 15 GB in size and I/O processes are > very slow. > You who recommend me? > > In the documentation of ceph recommend using SSD for the journal, my > question is > How many SSD do I have to enable per server so that the journals of the 9 > OSDs can be separated into SSDs? > > I currently use ceph with OpenStack, on 11 servers with SO Debian Stretch: > * 3 controller > * 3 compute > * 5 ceph-osd > network: bond lacp 10GB > RAM: 96GB > HD: 9 disk SATA-3TB (bluestore) > > -- > *Davis Mendoza P.* > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com