On Wed, Feb 18, 2015 at 7:53 PM, Brian Rak <b...@gameservers.com> wrote: > We're running ceph version 0.87 (c51c8f9d80fa4e0168aa52685b8de40e42758578), > and seeing this: > > HEALTH_WARN 1 pgs degraded; 1 pgs stuck degraded; 1 pgs stuck unclean; 1 pgs > stuck undersized; 1 pgs undersized > pg 4.2af is stuck unclean for 77192.522960, current state > active+undersized+degraded, last acting [50,42] > pg 4.2af is stuck undersized for 980.617479, current state > active+undersized+degraded, last acting [50,42] > pg 4.2af is stuck degraded for 980.617902, current state > active+undersized+degraded, last acting [50,42] > pg 4.2af is active+undersized+degraded, acting [50,42] > > > However, ceph pg query doesn't really show any issues: > https://gist.githubusercontent.com/devicenull/9d911362e4de83c02e40/raw/565fe18163e261c8105e5493a4e90cc3c461ed9d/gistfile1.txt > (too long to post here) > > I've also tried: > > # ceph pg 4.2af mark_unfound_lost revert > pg has no unfound objects > > How can I get Ceph to rebuild here? The replica count is 3, but I can't > seem to figure out what's going on here. Enabling various debug logs > doesn't reveal anything obvious to me. > > I've tried restarting both OSDs, which did nothing.
What does your crushmap look like (ceph osd getcrushmap -o /tmp/crushmap; crushtool -d /tmp/crushmap)? Does your placement logic prevent Ceph from selecting an OSD for the third replica? Cheers, Florian _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com