Hi,

if you value your data you shouldn't set it to k (in your case 6). The docs [1] are pretty clear about that:

 min_size
Sets the minimum number of replicas required for I/O. See Set the Number of Object Replicas for further details. In the case of Erasure Coded pools this should be set to a value greater than ‘k’ since if we allow IO at the value ‘k’ there is no redundancy and data will be lost in the event of a permanent OSD failure.

Regards,
Eugen

[1] https://docs.ceph.com/en/latest/rados/operations/pools/#set-pool-values

Zitat von Christopher Durham <caduceu...@aol.com>:

Hello,
I have an ec pool set as 6+2. I have noticed that when rebooting servers during system upgrades, I get pgs set to inactive while the osds are down. I then discovered that my min_size for the pool is set to 7, which makes sense that if I reboot two servers that host a pg with OSDs on both servers, then only 6 of the OSDs are available during the reboot cycle, and with a min_size of 7, they go inactive. Is is ok for me to set min_size on the pool to 6, to avoid the inactive problem? I know I could do my reboots sequentially to eliminate multiple server downtime, but wanted to be sure the min_size 6 is ok. I know this may increase other risks, but wanted to know if this min_size change is an option, albeit more risky. Thanks.

-Chris
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io



_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to