Hi,

Your cluster was healthy before purging osd?
How much time did you wait between stoping osd and purging them?

Étienne

> On 17 Jan 2022, at 15:24, Rafael Diaz Maurin 
> <rafael.diazmau...@univ-rennes1.fr> wrote:
>
> Hello,
>
> All my pools on the cluster are replicated (x3).
>
> I purged some OSD (after I stopped them) and remove the disks from the 
> servers, and now I have 4 PGs in stale+undersized+degraded+peered.
>
> Reduced data availability: 4 pgs inactive, 4 pgs stale
>
> pg 1.561 is stuck stale for 39m, current state 
> stale+undersized+degraded+peered, last acting [64]
> pg 1.af2 is stuck stale for 39m, current state 
> stale+undersized+degraded+peered, last acting [63]
> pg 3.3 is stuck stale for 39m, current state 
> stale+undersized+degraded+peered, last acting [48]
> pg 9.5ca is stuck stale for 38m, current state 
> stale+undersized+degraded+peered, last acting [49]
>
>
> Those 4 OSDs have been purged, so ther aren't anymore in the crushmap.
>
> I tried a pg repair :
> ceph pg repair 1.561
> ceph pg repair 1.af2
> ceph pg repair 3.3
> ceph pg repair 9.5ca
>
>
> The PGs are remapped but non of the degraded objects have been repaired
>
> ceph pg map 9.5ca
> osdmap e355782 pg 9.5ca (9.5ca) -> up [54,75,82] acting [54,75,82]
> ceph pg map 3.3
> osdmap e355782 pg 3.3 (3.3) -> up [179,180,107] acting [179,180,107]
> ceph pg map 1.561
> osdmap e355785 pg 1.561 (1.561) -> up [70,188,87] acting [70,188,87]
> ceph pg map 1.af2
> osdmap e355789 pg 1.af2 (1.af2) -> up [189,74,184] acting [189,74,184]
>
> How can I succeed in reparing my 4 PGs ?
>
> This affect the cephfs-metadata pool, and the filesystem is degraded because 
> the rank0 mds node stuck in rejoin state.
>
>
> Thank you.
>
> Rafael
>
>
> --
> Rafael Diaz Maurin
> DSI de l'Université de Rennes 1
> Pôle Infrastructures, équipe Systèmes
> 02 23 23 71 57
>
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to