Can you post your whole crushmap?

ceph osd getcrushmap -o crushmap
crushtool -d crushmap -o crushmap.txt


Paul


2018-06-14 22:39 GMT+02:00 Oliver Schulz <oliver.sch...@tu-dortmund.de>:

> Thanks, Greg!!
>
> I reset all the OSD weights to 1.00, and I think I'm in a much
> better state now. The only trouble left in "ceph health detail" is
>
> PG_DEGRADED Degraded data redundancy: 4/404985012 objects degraded
> (0.000%), 3 pgs degraded
>     pg 2.47 is active+recovery_wait+degraded+remapped, acting [177,68,187]
>     pg 2.1fd is active+recovery_wait+degraded+remapped, acting [36,83,185]
>     pg 2.748 is active+recovery_wait+degraded, acting [31,8,149]
>
> (There's a lot of misplaced PGs now, obviously). The interesting
> thing is that my "lost" PG is back, too, with three acting OSDs.
>
> Maybe I dodged the bullet - what do you think?
>
> One question: Is there a way to give recovery of the three
> degraded PGs priority over backfilling the misplaced ones?
> I tried "ceph pg force-recovery" but it didn't seem to have
> any effect, they were still on "recovery_wait", after.
>
>
> Cheers,
>
> Oliver
>
>
> On 14.06.2018 22:09, Gregory Farnum wrote:
>
>> On Thu, Jun 14, 2018 at 4:07 PM Oliver Schulz <
>> oliver.sch...@tu-dortmund.de <mailto:oliver.sch...@tu-dortmund.de>>
>> wrote:
>>
>>     Hi Greg,
>>
>>     I increased the hard limit and rebooted everything. The
>>     PG without acting OSDs still has none, but I also have
>>     quite a few PGs with that look like this, now:
>>
>>           pg 1.79c is stuck undersized for 470.640254, current state
>>     active+undersized+degraded, last acting [179,154]
>>
>>     I had that problem before (only two acting OSDs on a few PGs),
>>     I always solved it by setting the primary OSD to out and then
>>     back in a few seconds later (resulting in a very quick recovery,
>>     then all was fine again). But maybe that's not the ideal solution?
>>
>>     Here's "ceph pg map" for one of them:
>>
>>           osdmap e526060 pg 1.79c (1.79c) -> up [179,154] acting [179,154]
>>
>>     I also have two PG's that have only one acting OSD, now:
>>
>>           osdmap e526060 pg 0.58a (0.58a) -> up [174] acting [174]
>>           osdmap e526060 pg 2.139 (2.139) -> up [61] acting [61]
>>
>>     How can I make Ceph assign three OSD's to all of these weird PGs?
>>     Before the reboot, they all did have three OSDs assigned (except for
>>     the one that has none), and they were not shown as degraded.
>>
>>
>>       > If it's the second, then fixing the remapping problem will
>>     resolve it.
>>       > That's probably/hopefully just by undoing the remap-by-utilization
>>       > changes.
>>
>>     How do I do that, best? Just set all the weights back to 1.00?
>>
>>
>> Yeah. This is probably the best way to fix up the other undersized PGs —
>> at least, assuming it doesn't result in an over-full PG!
>>
>> I don't work with overflowing OSDs/clusters often, but my suspicion is
>> you're better off with something like CERN's reweight scripts than using
>> reweight-by-utilization. Unless it's improved without my noticing, that
>> algorithm just isn't very good. :/
>> -Greg
>>
>>
>>
>>     Cheers,
>>
>>     Oliver
>>
>>
>>     P.S.: Thanks so much for helping!
>>
>>
>>
>>     On 14.06.2018 21:37, Gregory Farnum wrote:
>>      > On Thu, Jun 14, 2018 at 3:26 PM Oliver Schulz
>>      > <oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>> wrote:
>>      >
>>      >     But the contents of the remapped PGs should still be
>>      >     Ok, right? What confuses me is that they don't
>>      >     backfill - why don't the "move" where they belong?
>>      >
>>      >     As for the PG hard limit, yes, I ran into this. Our
>>      >     cluster had been very (very) full, but I wanted the
>>      >     new OSD nodes to use bluestore, so I updated to
>>      >     Luminous before I added the additional storage. I
>>      >     temporarily increased the pg hard limit and after
>>      >     a while (and after adding the new OSDs) the cluster
>>      >     seemed to be in a decent state again. Afterwards,
>>      >     I set the PG hard limit back to normal.
>>      >
>>      >     I don't have a "too many PGs per OSD" health warning,
>>      >     currently - should I still increase the PG hard limit?
>>      >
>>      >
>>      > Well, it's either the hard limit getting hit, or the fact that
>>     the PG
>>      > isn't getting mapped to any OSD and there not being an existing
>>     primary
>>      > to take responsibility for remapping it.
>>      >
>>      > If it's the second, then fixing the remapping problem will
>>     resolve it.
>>      > That's probably/hopefully just by undoing the
>>     remap-by-utilization changes.
>>      >
>>      >
>>      >     On 14.06.2018 20:58, Gregory Farnum wrote:
>>      >      > Okay, I can’t tell you what happened to that one pg, but
>>     you’ve got
>>      >      > another 445 remapped pgs and that’s not a good state to be
>>     in. It
>>      >     was
>>      >      > probably your use of the rewritten-by-utilization. :/ I am
>>     pretty
>>      >     sure
>>      >      > the missing PG and remapped ones have the same root cause,
>>     and it’s
>>      >      > possible but by no means certain fixing one will fix the
>>     others.
>>      >      >
>>      >      >
>>      >      > ...oh, actually, the most likely cause just came up in an
>>     unrelated
>>      >      > conversation. You’ve probably run into the pg overdose
>>     protection
>>      >     that
>>      >      > was added in luminous. Check the list archives for the exact
>>      >     name, but
>>      >      > you’ll want to increase the pg hard limit and restart the
>>     osds that
>>      >      > exceeded the previous/current setting.
>>      >      > -Greg
>>      >      > On Thu, Jun 14, 2018 at 2:33 PM Oliver Schulz
>>      >      > <oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>> wrote:
>>      >      >
>>      >      >     I'm not running the balancer, but I did
>>     reweight-by-utilization
>>      >      >     a few times recently.
>>      >      >
>>      >      >     "ceph osd tree" and "ceph -s" say:
>>      >      >
>>      >      >
>>     https://gist.github.com/oschulz/36d92af84851ec42e09ce1f3cacbc110
>>      >      >
>>      >      >
>>      >      >
>>      >      >     On 14.06.2018 20:23, Gregory Farnum wrote:
>>      >      >      > Well, if this pg maps to no osds, something has
>>     certainly
>>      >     gone wrong
>>      >      >      > with your crush map. What’s the crush rule it’s
>>     using, and
>>      >     what’s
>>      >      >     the
>>      >      >      > output of “ceph osd tree”?
>>      >      >      > Are you running the manager’s balancer module or
>>     something
>>      >     that
>>      >      >     might be
>>      >      >      > putting explicit mappings into the osd map and
>>     broken it?
>>      >      >      >
>>      >      >      > I’m not certain off-hand about the pg reporting, but
>> I
>>      >     believe if
>>      >      >     it’s
>>      >      >      > reporting the state as unknown that means *no*
>>     running osd
>>      >     which
>>      >      >      > contains any copy of that pg. That’s not something
>>     which ceph
>>      >      >     could do
>>      >      >      > on its own without failures of osds. What’s the
>>     output of
>>      >     “ceph -s”?
>>      >      >      > On Thu, Jun 14, 2018 at 2:15 PM Oliver Schulz
>>      >      >      > <oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>>> wrote:
>>      >      >      >
>>      >      >      >     Dear Greg,
>>      >      >      >
>>      >      >      >     no, it's a very old cluster (continuous operation
>>      >     since 2013,
>>      >      >      >     with multiple extensions). It's a production
>>     cluster and
>>      >      >      >     there's about 300TB of valuable data on it.
>>      >      >      >
>>      >      >      >     We recently updated to luminous and added more
>>     OSDs (a
>>      >     month
>>      >      >      >     ago or so), but everything seemed Ok since then.
>> We
>>      >     didn't have
>>      >      >      >     any disk failures, but we had trouble with the
>>     MDS daemons
>>      >      >      >     in the last days, so there were a few reboots.
>>      >      >      >
>>      >      >      >     Is it somehow possible to find this "lost" PG
>>     again? Since
>>      >      >      >     it's in the metadata pool, large parts of our
>>     CephFS
>>      >     directory
>>      >      >      >     tree are currently unavailable. I turned the MDS
>>      >     daemons off
>>      >      >      >     for now ...
>>      >      >      >
>>      >      >      >
>>      >      >      >     Cheers
>>      >      >      >
>>      >      >      >     Oliver
>>      >      >      >
>>      >      >      >     On 14.06.2018 19:59, Gregory Farnum wrote:
>>      >      >      >      > Is this a new cluster? Or did the crush map
>>     change
>>      >     somehow
>>      >      >      >     recently? One
>>      >      >      >      > way this might happen is if CRUSH just failed
>>      >     entirely to
>>      >      >     map a pg,
>>      >      >      >      > although I think if the pg exists anywhere it
>>      >     should still be
>>      >      >      >     getting
>>      >      >      >      > reported as inactive.
>>      >      >      >      > On Thu, Jun 14, 2018 at 8:40 AM Oliver Schulz
>>      >      >      >      > <oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>
>>      >      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>>
>>      >      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>
>>      >      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>
>>      >      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>
>>      >     <mailto:oliver.sch...@tu-dortmund.de
>>     <mailto:oliver.sch...@tu-dortmund.de>>>>>> wrote:
>>      >      >      >      >
>>      >      >      >      >     Dear all,
>>      >      >      >      >
>>      >      >      >      >     I have a serious problem with our Ceph
>>     cluster:
>>      >     One of our
>>      >      >      >     PGs somehow
>>      >      >      >      >     ended up in this state (reported by "ceph
>>      >     health detail":
>>      >      >      >      >
>>      >      >      >      >           pg 1.XXX is stuck inactive for
>>     ..., current
>>      >      >     state unknown,
>>      >      >      >      >     last acting []
>>      >      >      >      >
>>      >      >      >      >     Also, "ceph pg map 1.xxx" reports:
>>      >      >      >      >
>>      >      >      >      >           osdmap e525812 pg 1.721 (1.721) ->
>>     up []
>>      >     acting []
>>      >      >      >      >
>>      >      >      >      >     I can't use "ceph pg 1.XXX query", it just
>>      >     hangs with
>>      >      >     no output.
>>      >      >      >      >
>>      >      >      >      >     All OSDs are up and in, I have MON
>>     quorum, all
>>      >     other
>>      >      >     PGs seem
>>      >      >      >     to be
>>      >      >      >      >     fine.
>>      >      >      >      >
>>      >      >      >      >     How can diagnose/fix this?
>>     Unfortunately, the PG in
>>      >      >     question
>>      >      >      >     is part
>>      >      >      >      >     of the CephFS metadata pool ...
>>      >      >      >      >
>>      >      >      >      >     Any help would be very, very much
>>     appreciated!
>>      >      >      >      >
>>      >      >      >      >
>>      >      >      >      >     Cheers,
>>      >      >      >      >
>>      >      >      >      >     Oliver
>>      >      >      >      >         _____________________________
>> __________________
>>      >      >      >      >     ceph-users mailing list
>>      >      >      >      > ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>
>>      >      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com> <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>
>>      >      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>>>
>>      >      >      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>
>>      >      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com> <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>
>>      >      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>
>>      >     <mailto:ceph-users@lists.ceph.com
>>     <mailto:ceph-users@lists.ceph.com>>>>>
>>      >      >      >      >
>>     http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>      >      >      >      >
>>      >      >      >
>>      >      >
>>      >
>>
>> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>



-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to