Sorry, a step-by-step guide through something like that
is beyond the scope of what we can do on a mailing list.

But what I would do here is carefully asses the situation/
the damage. My wild guess would be to reset and rebuild
the inode table but that might be incorrect and unsafe
without further looking into it.

I don't want to solicit our services here, but we do Ceph
recoveries regularly; reach out to us if you are looking
for a consultant.


Paul


2018-08-10 18:05 GMT+02:00 Amit Handa <amit.ha...@gmail.com>:

> Thanks alot, Paul.
> we did (hopefully) follow through with the disaster recovery.
> however, please guide me in how to get the cluster back up !
>
> Thanks,
>
>
> On Fri, Aug 10, 2018 at 9:32 PM Paul Emmerich <paul.emmer...@croit.io>
> wrote:
>
>> Looks like you got some duplicate inodes due to corrupted metadata, you
>> likely tried to a disaster recovery and didn't follow through it
>> completely or
>> you hit some bug in Ceph.
>>
>> The solution here is probably to do a full recovery of the metadata/full
>> backwards scan after resetting the inodes. I've recovered a cluster from
>> something similar just a few weeks ago. Annoying but recoverable.
>>
>> Paul
>>
>> 2018-08-10 13:26 GMT+02:00 Amit Handa <amit.ha...@gmail.com>:
>>
>>> We are facing constant crash from ceph mds. We have installed mimic
>>> (v13.2.1).
>>>
>>> mds: cephfs-1/1/1 up {0=node2=up:active(laggy or crashed)}
>>>
>>> *mds logs: https://pastebin.com/AWGMLRm0 <https://pastebin.com/AWGMLRm0>*
>>>
>>> we have followed the DR steps listed at
>>>
>>> *http://docs.ceph.com/docs/mimic/cephfs/disaster-recovery/
>>> <http://docs.ceph.com/docs/mimic/cephfs/disaster-recovery/> *
>>>
>>> please help in resolving the errors :(
>>>
>>> mds crash stacktrace
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> * ceph version 13.2.1 (5533ecdc0fda920179d7ad84e0aa65a127b20d77) mimic
>>> (stable) 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
>>> const*)+0xff) [0x7f984fc3ee1f] 2: (()+0x284fe7) [0x7f984fc3efe7] 3:
>>> (()+0x2087fe) [0x5563e88537fe] 4:
>>> (Server::prepare_new_inode(boost::intrusive_ptr<MDRequestImpl>&, CDir*,
>>> inodeno_t, unsigned int, file_layout_t*)+0xf37) [0x5563e87ce777] 5:
>>> (Server::handle_client_openc(boost::intrusive_ptr<MDRequestImpl>&)+0xdb0)
>>> [0x5563e87d0bd0] 6: (Server::handle_client_request(MClientRequest*)+0x49e)
>>> [0x5563e87d3c0e] 7: (Server::dispatch(Message*)+0x2db) [0x5563e87d789b] 8:
>>> (MDSRank::handle_deferrable_message(Message*)+0x434) [0x5563e87514b4] 9:
>>> (MDSRank::_dispatch(Message*, bool)+0x63b) [0x5563e875db5b] 10:
>>> (MDSRank::retry_dispatch(Message*)+0x12) [0x5563e875e302] 11:
>>> (MDSInternalContextBase::complete(int)+0x67) [0x5563e89afb57] 12:
>>> (MDSRank::_advance_queues()+0xd1) [0x5563e875cd51] 13:
>>> (MDSRank::ProgressThread::entry()+0x43) [0x5563e875d3e3] 14: (()+0x7e25)
>>> [0x7f984d869e25] 15: (clone()+0x6d) [0x7f984c949bad] NOTE: a copy of the
>>> executable, or `objdump -rdS <executable>` is needed to interpret this.*
>>>
>>> --
>>> Loading ...
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> --
>> Paul Emmerich
>>
>> Looking for help with your Ceph cluster? Contact us at https://croit.io
>>
>> croit GmbH
>> Freseniusstr. 31h
>> <https://maps.google.com/?q=Freseniusstr.+31h+81247+M%C3%BCnchen&entry=gmail&source=g>
>> 81247 München
>> <https://maps.google.com/?q=Freseniusstr.+31h+81247+M%C3%BCnchen&entry=gmail&source=g>
>> www.croit.io
>> Tel: +49 89 1896585 90
>>
>
>
> --
> Loading ...
>



-- 
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to