ceph-mds --undump-journal <rank> <journal-file>
Looks like it accidentally (or on purpose? you can break things with
it) got left out of the help text.

On Tue, Oct 14, 2014 at 8:19 AM, Jasper Siero
<jasper.si...@target-holding.nl> wrote:
> Hello Greg,
>
> I dumped the journal successful to a file:
>
> journal is 9483323613~134215459
> read 134213311 bytes at offset 9483323613
> wrote 134213311 bytes at offset 9483323613 to journaldumptgho
> NOTE: this is a _sparse_ file; you can
>         $ tar cSzf journaldumptgho.tgz journaldumptgho
>       to efficiently compress it while preserving sparseness.
>
> I see the option for resetting the mds journal but I can't find the option 
> for undumping /importing the journal:
>
>  usage: ceph-mds -i name [flags] [[--journal_check 
> rank]|[--hot-standby][rank]]
>   -m monitorip:port
>         connect to monitor at given address
>   --debug_mds n
>         debug MDS level (e.g. 10)
>   --dump-journal rank filename
>         dump the MDS journal (binary) for rank.
>   --dump-journal-entries rank filename
>         dump the MDS journal (JSON) for rank.
>   --journal-check rank
>         replay the journal for rank, then exit
>   --hot-standby rank
>         start up as a hot standby for rank
>   --reset-journal rank
>         discard the MDS journal for rank, and replace it with a single
>         event that updates/resets inotable and sessionmap on replay.
>
> Do you know how to "undump" the journal back into ceph?
>
> Jasper
>
> ________________________________________
> Van: Gregory Farnum [g...@inktank.com]
> Verzonden: vrijdag 10 oktober 2014 23:45
> Aan: Jasper Siero
> CC: ceph-users
> Onderwerp: Re: [ceph-users] mds isn't working anymore after osd's running full
>
> Ugh, "debug journaler", not "debug journaled."
>
> That said, the filer output tells me that you're missing an object out
> of the MDS log. (200.000008f5) I think this issue should be resolved
> if you "dump" the journal to a file, "reset" it, and then "undump" it.
> (These are commands you can invoke from ceph-mds.)
> I haven't done this myself in a long time, so there may be some hard
> edges around it. In particular, I'm not sure if the dumped journal
> file will stop when the data stops, or if it will be a little too
> long. If so, we can fix that by truncating the dumped file to the
> proper length and resetting and undumping again.
> (And just to harp on it, this journal manipulation is a lot simpler in
> Giant... ;) )
> -Greg
> Software Engineer #42 @ http://inktank.com | http://ceph.com
>
> On Wed, Oct 8, 2014 at 7:11 AM, Jasper Siero
> <jasper.si...@target-holding.nl> wrote:
>> Hello Greg,
>>
>> No problem thanks for looking into the log. I attached the log to this email.
>> I'm looking forward for the new release because it would be nice to have 
>> more possibilities to diagnose problems.
>>
>> Kind regards,
>>
>> Jasper Siero
>> ________________________________________
>> Van: Gregory Farnum [g...@inktank.com]
>> Verzonden: dinsdag 7 oktober 2014 19:45
>> Aan: Jasper Siero
>> CC: ceph-users
>> Onderwerp: Re: [ceph-users] mds isn't working anymore after osd's running 
>> full
>>
>> Sorry; I guess this fell off my radar.
>>
>> The issue here is not that it's waiting for an osdmap; it got the
>> requested map and went into replay mode almost immediately. In fact
>> the log looks good except that it seems to finish replaying the log
>> and then simply fail to transition into active. Generate a new one,
>> adding in "debug journaled = 20" and "debug filer = 20", and we can
>> probably figure out how to fix it.
>> (This diagnosis is much easier in the upcoming Giant!)
>> -Greg
>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>
>>
>> On Tue, Oct 7, 2014 at 7:55 AM, Jasper Siero
>> <jasper.si...@target-holding.nl> wrote:
>>> Hello Gregory,
>>>
>>> We still have the same problems with our test ceph cluster and didn't 
>>> receive a reply from you after I send you the requested log files. Do you 
>>> know if it's possible to get our cephfs filesystem working again or is it 
>>> better to give up the files on cephfs and start over again?
>>>
>>> We restarted the cluster serveral times but it's still degraded:
>>> [root@th1-mon001 ~]# ceph -w
>>>     cluster c78209f5-55ea-4c70-8968-2231d2b05560
>>>      health HEALTH_WARN mds cluster is degraded
>>>      monmap e3: 3 mons at 
>>> {th1-mon001=10.1.2.21:6789/0,th1-mon002=10.1.2.22:6789/0,th1-mon003=10.1.2.23:6789/0},
>>>  election epoch 432, quorum 0,1,2 th1-mon001,th1-mon002,th1-mon003
>>>      mdsmap e190: 1/1/1 up {0=th1-mon001=up:replay}, 1 up:standby
>>>      osdmap e2248: 12 osds: 12 up, 12 in
>>>       pgmap v197548: 492 pgs, 4 pools, 60297 MB data, 470 kobjects
>>>             124 GB used, 175 GB / 299 GB avail
>>>                  491 active+clean
>>>                    1 active+clean+scrubbing+deep
>>>
>>> One placement group stays in the deep scrubbing fase.
>>>
>>> Kind regards,
>>>
>>> Jasper Siero
>>>
>>>
>>> ________________________________________
>>> Van: Jasper Siero
>>> Verzonden: donderdag 21 augustus 2014 16:43
>>> Aan: Gregory Farnum
>>> Onderwerp: RE: [ceph-users] mds isn't working anymore after osd's running 
>>> full
>>>
>>> I did restart it but you are right about the epoch number which has changed 
>>> but the situation looks the same.
>>> 2014-08-21 16:33:06.032366 7f9b5f3cd700  1 mds.0.27  need osdmap epoch 
>>> 1994, have 1993
>>> 2014-08-21 16:33:06.032368 7f9b5f3cd700  1 mds.0.27  waiting for osdmap 
>>> 1994 (which blacklists
>>> prior instance)
>>> I started the mds with the debug options and attached the log.
>>>
>>> Thanks,
>>>
>>> Jasper
>>> ________________________________________
>>> Van: Gregory Farnum [g...@inktank.com]
>>> Verzonden: woensdag 20 augustus 2014 18:38
>>> Aan: Jasper Siero
>>> CC: ceph-users@lists.ceph.com
>>> Onderwerp: Re: [ceph-users] mds isn't working anymore after osd's running 
>>> full
>>>
>>> After restarting your MDS, it still says it has epoch 1832 and needs
>>> epoch 1833? I think you didn't really restart it.
>>> If the epoch numbers have changed, can you restart it with "debug mds
>>> = 20", "debug objecter = 20", "debug ms = 1" in the ceph.conf and post
>>> the resulting log file somewhere?
>>> -Greg
>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
>>>
>>>
>>> On Wed, Aug 20, 2014 at 12:49 AM, Jasper Siero
>>> <jasper.si...@target-holding.nl> wrote:
>>>> Unfortunately that doesn't help. I restarted both the active and standby 
>>>> mds but that doesn't change the state of the mds. Is there a way to force 
>>>> the mds to look at the 1832 epoch (or earlier) instead of 1833 (need 
>>>> osdmap epoch 1833, have 1832)?
>>>>
>>>> Thanks,
>>>>
>>>> Jasper
>>>> ________________________________________
>>>> Van: Gregory Farnum [g...@inktank.com]
>>>> Verzonden: dinsdag 19 augustus 2014 19:49
>>>> Aan: Jasper Siero
>>>> CC: ceph-users@lists.ceph.com
>>>> Onderwerp: Re: [ceph-users] mds isn't working anymore after osd's running 
>>>> full
>>>>
>>>> On Mon, Aug 18, 2014 at 6:56 AM, Jasper Siero
>>>> <jasper.si...@target-holding.nl> wrote:
>>>>> Hi all,
>>>>>
>>>>> We have a small ceph cluster running version 0.80.1 with cephfs on five
>>>>> nodes.
>>>>> Last week some osd's were full and shut itself down. To help de osd's 
>>>>> start
>>>>> again I added some extra osd's and moved some placement group directories 
>>>>> on
>>>>> the full osd's (which has a copy on another osd) to another place on the
>>>>> node (as mentioned in
>>>>> http://ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/)
>>>>> After clearing some space on the full osd's I started them again. After a
>>>>> lot of deep scrubbing and two pg inconsistencies which needed to be 
>>>>> repaired
>>>>> everything looked fine except the mds which still is in the replay state 
>>>>> and
>>>>> it stays that way.
>>>>> The log below says that mds need osdmap epoch 1833 and have 1832.
>>>>>
>>>>> 2014-08-18 12:29:22.268248 7fa786182700  1 mds.-1.0 handle_mds_map standby
>>>>> 2014-08-18 12:29:22.273995 7fa786182700  1 mds.0.25 handle_mds_map i am 
>>>>> now
>>>>> mds.0.25
>>>>> 2014-08-18 12:29:22.273998 7fa786182700  1 mds.0.25 handle_mds_map state
>>>>> change up:standby --> up:replay
>>>>> 2014-08-18 12:29:22.274000 7fa786182700  1 mds.0.25 replay_start
>>>>> 2014-08-18 12:29:22.274014 7fa786182700  1 mds.0.25  recovery set is
>>>>> 2014-08-18 12:29:22.274016 7fa786182700  1 mds.0.25  need osdmap epoch 
>>>>> 1833,
>>>>> have 1832
>>>>> 2014-08-18 12:29:22.274017 7fa786182700  1 mds.0.25  waiting for osdmap 
>>>>> 1833
>>>>> (which blacklists prior instance)
>>>>>
>>>>>  # ceph status
>>>>>     cluster c78209f5-55ea-4c70-8968-2231d2b05560
>>>>>      health HEALTH_WARN mds cluster is degraded
>>>>>      monmap e3: 3 mons at
>>>>> {th1-mon001=10.1.2.21:6789/0,th1-mon002=10.1.2.22:6789/0,th1-mon003=10.1.2.23:6789/0},
>>>>> election epoch 362, quorum 0,1,2 th1-mon001,th1-mon002,th1-mon003
>>>>>      mdsmap e154: 1/1/1 up {0=th1-mon001=up:replay}, 1 up:standby
>>>>>      osdmap e1951: 12 osds: 12 up, 12 in
>>>>>       pgmap v193685: 492 pgs, 4 pools, 60297 MB data, 470 kobjects
>>>>>             124 GB used, 175 GB / 299 GB avail
>>>>>                  492 active+clean
>>>>>
>>>>> # ceph osd tree
>>>>> # id    weight    type name    up/down    reweight
>>>>> -1    0.2399    root default
>>>>> -2    0.05997        host th1-osd001
>>>>> 0    0.01999            osd.0    up    1
>>>>> 1    0.01999            osd.1    up    1
>>>>> 2    0.01999            osd.2    up    1
>>>>> -3    0.05997        host th1-osd002
>>>>> 3    0.01999            osd.3    up    1
>>>>> 4    0.01999            osd.4    up    1
>>>>> 5    0.01999            osd.5    up    1
>>>>> -4    0.05997        host th1-mon003
>>>>> 6    0.01999            osd.6    up    1
>>>>> 7    0.01999            osd.7    up    1
>>>>> 8    0.01999            osd.8    up    1
>>>>> -5    0.05997        host th1-mon002
>>>>> 9    0.01999            osd.9    up    1
>>>>> 10    0.01999            osd.10    up    1
>>>>> 11    0.01999            osd.11    up    1
>>>>>
>>>>> What is the way to get the mds up and running again?
>>>>>
>>>>> I still have all the placement group directories which I moved from the 
>>>>> full
>>>>> osds which where down to create disk space.
>>>>
>>>> Try just restarting the MDS daemon. This sounds a little familiar so I
>>>> think it's a known bug which may be fixed in a later dev or point
>>>> release on the MDS, but it's a soft-state rather than a disk state
>>>> issue.
>>>> -Greg
>>>> Software Engineer #42 @ http://inktank.com | http://ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to