On Thu, Oct 12, 2017 at 10:52 AM Florian Haas <flor...@hastexo.com> wrote:

> On Thu, Oct 12, 2017 at 7:22 PM, Gregory Farnum <gfar...@redhat.com>
> wrote:
> >
> >
> > On Thu, Oct 12, 2017 at 3:50 AM Florian Haas <flor...@hastexo.com>
> wrote:
> >>
> >> On Mon, Sep 11, 2017 at 8:13 PM, Andreas Herrmann <andr...@mx20.org>
> >> wrote:
> >> > Hi,
> >> >
> >> > how could this happen:
> >> >
> >> >         pgs: 197528/1524 objects degraded (12961.155%)
> >> >
> >> > I did some heavy failover tests, but a value higher than 100% looks
> >> > strange
> >> > (ceph version 12.2.0). Recovery is quite slow.
> >> >
> >> >   cluster:
> >> >     health: HEALTH_WARN
> >> >             3/1524 objects misplaced (0.197%)
> >> >             Degraded data redundancy: 197528/1524 objects degraded
> >> > (12961.155%), 1057 pgs unclean, 1055 pgs degraded, 3 pgs undersized
> >> >
> >> >   data:
> >> >     pools:   1 pools, 2048 pgs
> >> >     objects: 508 objects, 1467 MB
> >> >     usage:   127 GB used, 35639 GB / 35766 GB avail
> >> >     pgs:     197528/1524 objects degraded (12961.155%)
> >> >              3/1524 objects misplaced (0.197%)
> >> >              1042 active+recovery_wait+degraded
> >> >              991  active+clean
> >> >              8    active+recovering+degraded
> >> >              3    active+undersized+degraded+remapped+backfill_wait
> >> >              2    active+recovery_wait+degraded+remapped
> >> >              2    active+remapped+backfill_wait
> >> >
> >> >   io:
> >> >     recovery: 340 kB/s, 80 objects/s
> >>
> >> Did you ever get to the bottom of this? I'm seeing something very
> >> similar on a 12.2.1 reference system:
> >>
> >> https://gist.github.com/fghaas/f547243b0f7ebb78ce2b8e80b936e42c
> >>
> >> I'm also seeing an unusual MISSING_ON_PRIMARY count in "rados df":
> >> https://gist.github.com/fghaas/59cd2c234d529db236c14fb7d46dfc85
> >>
> >> The odd thing in there is that the "bench" pool was empty when the
> >> recovery started (that pool had been wiped with "rados cleanup"), so
> >> the number of objects deemed to be missing from the primary really
> >> ought to be zero.
> >>
> >> It seems like it's considering these deleted objects to still require
> >> replication, but that sounds rather far fetched to be honest.
> >
> >
> > Actually, that makes some sense. This cluster had an OSD down while (some
> > of) the deletes were happening?
>
> I thought of exactly that too, but no it didn't. That's the problem.
>

Okay, in that case I've no idea. What was the timeline for the recovery
versus the rados bench and cleanup versus the degraded object counts, then?


>
> > I haven't dug through the code but I bet it is considering those as
> degraded
> > objects because the out-of-date OSD knows it doesn't have the latest
> > versions on them! :)
>
> Yeah I bet against that. :)
>
> Another tidbit: these objects were not deleted with rados rm, they
> were cleaned up after rados bench. In the case quoted above, this was
> an explicit "rados cleanup" after "rados bench --no-cleanup"; in
> another, I saw the same behavior after a regular "rados bench" that
> included the automatic cleanup.
>
> So there are two hypotheses here:
> (1) The deletion in rados bench is neglecting to do something that a
> regular object deletion does do. Given the fact that at least one
> other thing is fishy in rados bench
> (http://tracker.ceph.com/issues/21375), this may be due to some simple
> oversight in the Luminous cycle, and thus would constitute a fairly
> minor (if irritating) issue.
> (2) Regular object deletion is buggy in some previously unknown
> fashion. That would would be a rather major problem.
>

These both seem exceedingly unlikely. *shrug*


>
> By the way, *deleting the pool* altogether makes the degraded object
> count drop to expected levels immediately. Probably no surprise there,
> though.
>
> Cheers,
> Florian
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to