Hi Guys,
                I have 7 OSD nodes with 10X15T NVME disk on each OSD node.

To start with , I want to use only 8X15T disk on each osd node and keep
2X15 disk spare in case of any disk failure and recovery event.

I am going to use the 4X2 EC  CephFS data pool to store data.

So, with the above set-up, what will be the optimal number of placement
groups per OSD.

As per the PG calculator :-

( 8X7X100 ) / 6 = 933.33  nearest power of 2 is 1024.

With 1024 placement groups distributed across 56 OSDs, that evaluates to
approximately 18 placement groups per OSD.

I don't think its optimal as Ceph doc recommends 50-100 PG per OSD.

So, am I doing something wrong? or missing something white calculating the
number of PG per OSD.

Also, will it be best practice to keep 2X15T disk spare on each OSD or
should I use all of them.

Also, I am going to deploy 7 OSD nodes across 4 Racks and will be using the
failure domain as "rack" to have Ceph handle the entire rack failure.

I hope this will provide more protection.

Please advise.


Thanks,
Gagan




Thanks,
Gagan.
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to