adding back in the list :) ---------- Forwarded message --------- From: Luis Periquito <periqu...@gmail.com> Date: Wed, Jun 20, 2018 at 1:54 PM Subject: Re: [ceph-users] Planning all flash cluster To: <nick.b...@gmail.com>
On Wed, Jun 20, 2018 at 1:35 PM Nick A <nick.b...@gmail.com> wrote: > > Thank you, I was under the impression that 4GB RAM per 1TB was quite > generous, or is that not the case with all flash clusters? What's the > recommended RAM per OSD currently? Happy to throw more at it for a > performance boost. The important thing is that I'd like all nodes to be > absolutely identical. I'm doing 8G per OSD, though I use 1.9T SSDs. > > Based on replies so far, it looks like 5 nodes might be a better idea, maybe > each with 14 OSD's (960GB SSD's)? Plenty of 16 slot 2U chassis around to make > it a no brainer if that's what you'd recommend! I tend to add more nodes: 1U with 4-8 SSDs per chassis to start with, and using a single CPU with high frequency. For IOPS/latency cpu frequency is really important. I have started a cluster that only has 2 SSDs (which I share with the OS) for data, but has 8 nodes. Those servers can take up to 10 drives. I'm using the Fujitsu RX1330, believe Dell would be the R330, with a Intel E3-1230v6 cpu and 64G of ram, dual 10G and PSAS (passthrough controller). > > The H710 doesn't do JBOD or passthrough, hence looking for an alternative > HBA. It would be nice to do the boot drives as hardware RAID 1 though, so a > card that can do both at the same time (like the H730 found R630's etc) would > be ideal. > > Regards, > Nick > > On 20 June 2018 at 13:18, Luis Periquito <periqu...@gmail.com> wrote: >> >> Adding more nodes from the beginning would probably be a good idea. >> >> On Wed, Jun 20, 2018 at 12:58 PM Nick A <nick.b...@gmail.com> wrote: >> > >> > Hello Everyone, >> > >> > We're planning a small cluster on a budget, and I'd like to request any >> > feedback or tips. >> > >> > 3x Dell R720XD with: >> > 2x Xeon E5-2680v2 or very similar >> The CPUs look good and sufficiently fast for IOPS. >> >> > 96GB RAM >> 4GB per OSD looks a bit on the short side. Probably 192G would help. >> >> > 2x Samsung SM863 240GB boot/OS drives >> > 4x Samsung SM863 960GB OSD drives >> > Dual 40/56Gbit Infiniband using IPoIB. >> > >> > 3 replica, MON on OSD nodes, RBD only (no object or CephFS). >> > >> > We'll probably add another 2 OSD drives per month per node until full (24 >> > SSD's per node), at which point, more nodes. We've got a few SM863's in >> > production on other system and are seriously impressed with them, so would >> > like to use them for Ceph too. >> > >> > We're hoping this is going to provide a decent amount of IOPS, 20k would >> > be ideal. I'd like to avoid NVMe Journals unless it's going to make a >> > truly massive difference. Same with carving up the SSD's, would rather >> > not, and just keep it as simple as possible. >> I agree: those SSDs shouldn't really require a journal device. Not >> sure about the 20k IOPS specially without any further information. >> Doing 20k IOPS at 1kB block is totally different at 1MB block... >> > >> > Is there anything that obviously stands out as severely unbalanced? The >> > R720XD comes with a H710 - instead of putting them in RAID0, I'm thinking >> > a different HBA might be a better idea, any recommendations please? >> Don't know that HBA. Does it support pass through mode or HBA mode? >> > >> > Regards, >> > Nick >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com