Hi Anthony, Thanks for the reply! We will be using CephFS to access Ceph Storage from clients. So, this will need MDS daemon also.
So, based on your advice, I am thinking of having 4 Dell PowerEdge servers . 3 of them will run 3 Monitor daemons and one of them will run MDS daemon. These Dell Servers will have following hardware :- 1. 4 cores ( 8 threads ) ( Can go for 8 core and 16 threads ) 2. 64G RAM 3. 2x4T Samsung SSD with RA!D 1 to install OS and run monitor and metadata services. OSD nodes will be upgraded to have 32 cores ( 64 threads ). Disk and RAM will remain same ( 128G and 22X8T Samsung SSD ) Actually , I want to use OSD nodes to run OSD damons and not any other demons and which is why I am thinking of having 4 additional Dell servers as mentioned above. Please advise if this plan will be better. Thanks, Gagan On Wed, Apr 9, 2025 at 8:12 PM Anthony D'Atri <anthony.da...@gmail.com> wrote: > > > > > We would start deploying Ceph with 4 hosts ( HP Proliant servers ) each > > running RockyLinux 9. > > > > One of the hosts called ceph-adm will be smaller one and will have > > following hardware :- > > > > 2x4T SSD with raid 1 to install OS on. > > > > 8 Core with 3600MHz freq. > > > > 64G RAM > > > > We are planning to run all Ceph daemons except OSD daemon like monitor , > > metadata ,etc on this host. > > 8 core == 16 threads? Are you provisioning this node because you have it > laying around idle? > > Note that you will want *at least* 3 Monitor (monitors) daemons, which > must be on different nodes. 5 is better, but at least 3. You’ll also have > Grafana, Prometheus, MDS (if you’re going to CephFS vs using S3 object > storage or RBD block) > > 8c is likely on the light side for all of that. You would also benefit > from not having that node be a single point of failure. I would suggest if > you can raising this node to the spec of the planned 3x OSD nodes so you > have 4x equivalent nodes, and spread that non-OSD daemons across them. > > Note also that your OSD nodes will also have node_exporter, crash, and > other boilerplate daemons. > > > > We will have 3 hosts to run OSD which will store actual data. > > > > Each OSD host will have following hardware > > > > 2x4T SSD with raid 1 to install OS on. > > > > 22X8T SSD to store data ( OSDs ) ( without partition ). We will use > entire > > disk without partitions > > SAS, SATA, or NVMe SSDs? Which specific model? You really want to avoid > client (desktop) models for Ceph, but you likely do not need to pay for > higher endurance mixed-use SKUs. > > > Each OSD host will have 128G RAM ( No swap space ) > > Thank you for skipping swap. Some people are really stuck in the past in > that regard. > > > Each OSD host will have 16 cores. > > So 32 threads total? That is very light for 22 OSDs + other daemons. For > HDD OSDs a common rule of thumb is at minimum 2x threads per, for SAS/SATA > SSDs, 4, for NVMe SSDs 6. Plus margin for the OS and other processes. > > > All 4 hosts will connect to each via 10G nic. > > Two ports with bonding? Redundant switches? > > > The 500T data > > The specs you list above include 528 TB of *raw* space. Be advised that > with three OSD nodes, you will necessarily be doing replication. For > safety replication with size=3. Taking into consideration TB vs TiB and > headroom, you’re looking at 133TiB of usable space. You could go with > size=2 to get 300TB of usable space, but at increased risk of data > unavailability or loss when drives/hosts fail or reboot. > > With at least 4 OSD nodes - even if they aren’t fully populated with > capacity drives — you could do EC for a more favorable raw:usable ratio, at > the expense of slower writes and recovery. With 4 nodes you could in > theory do 2,2 EC for 200 TiB of usable space, with 5 you could do 3,2 for > 240 TiB usable, etc. > > > will be accessed by the clients. We need to have > > read performance as fast as possible. > > Hope your SSDs are enterprise NVMe. > > > We can't afford data loss and downtime. > > Then no size=2 for you. > > > So, we want to have a Ceph > > deployment which serves our purpose. > > > > So, please advise me if the plan that I have designed will serve our > > purpose. > > Or is there a better way , please advise that. > > > > Thanks, > > Gagan > > > > > > > > > > > > > > We have a HP storage server with 12 SDD of 5T each and have set-up > hardware > > RAID6 on these disks. > > > > HP storage server has 64G RAM and 18 cores. > > > > So, please advise how I should go about setting up Ceph on it to have > best > > read performance. We need fastest read performance. > > > > > > Thanks, > > Gagan > > _______________________________________________ > > ceph-users mailing list -- ceph-users@ceph.io > > To unsubscribe send an email to ceph-users-le...@ceph.io > > _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io