Thanks for the tip.

I will stay of the safe side and wait until it will be merged into master)

Many thanks for all your help.

-Mykola

On 19 November 2015 at 11:10, John Spray <jsp...@redhat.com> wrote:

> On Thu, Nov 19, 2015 at 10:07 AM, Mykola Dvornik
> <mykola.dvor...@gmail.com> wrote:
> > I'm guessing in this context that "write data" possibly means creating
> > a file (as opposed to writing to an existing file).
> >
> > Indeed. Sorry for the confusion.
> >
> > You've pretty much hit the limits of what the disaster recovery tools
> > are currently capable of.  What I'd recommend you do at this stage is
> > mount your filesystem read-only, back it up, and then create a new
> > filesystem and restore from backup.
> >
> > Ok. Is it somehow possible to have multiple FSs on the same ceph cluster?
>
> No, we want to do this but it's not there yet.  Your scenario is one
> of the motivations :-)
>
> (for the record multi-fs branch is
> https://github.com/jcsp/ceph/commits/wip-multi-filesystems, which
> works, but we'll probably go back and re-do the mon side of it before
> finishing)
>
> John
>
> >
> >
> > On 19 November 2015 at 10:43, John Spray <jsp...@redhat.com> wrote:
> >>
> >> On Wed, Nov 18, 2015 at 9:21 AM, Mykola Dvornik
> >> <mykola.dvor...@gmail.com> wrote:
> >> > Hi John,
> >> >
> >> > It turned out that mds triggers an assertion
> >> >
> >> > mds/MDCache.cc: 269: FAILED assert(inode_map.count(in->vino()) == 0)
> >> >
> >> > on any attempt to write data to the filesystem mounted via fuse.
> >>
> >> I'm guessing in this context that "write data" possibly means creating
> >> a file (as opposed to writing to an existing file).
> >>
> >> Currently, cephfs-data-scan injects inodes well enough that you can
> >> read them, but it's not updating the inode table to reflect that the
> >> recovered inodes are in use.  As a result, when new files are created
> >> they are probably trying to take inode numbers that are already in use
> >> (by the recovered files), and as a result you're hitting this
> >> assertion.  The ticket for updating the inotable after injection of
> >> recovered inodes is http://tracker.ceph.com/issues/12131
> >>
> >> > Deleting data is still OK.
> >> >
> >> > I cannot really follow why duplicated inodes appear.
> >> >
> >> > Are there any ways to flush/reset the MDS cache?
> >>
> >> You've pretty much hit the limits of what the disaster recovery tools
> >> are currently capable of.  What I'd recommend you do at this stage is
> >> mount your filesystem read-only, back it up, and then create a new
> >> filesystem and restore from backup.
> >>
> >> I'm writing a patch to handle the particular case where someone needs
> >> to update their inode table to mark all inodes as used up to some
> >> maximum, but the chances are that after that you'll still run into
> >> some other issue, until we've finished the tools to make it all the
> >> way through this path.
> >>
> >> John
> >>
> >> >
> >> >
> >> >
> >> > On 17 November 2015 at 13:26, John Spray <jsp...@redhat.com> wrote:
> >> >>
> >> >> On Tue, Nov 17, 2015 at 12:17 PM, Mykola Dvornik
> >> >> <mykola.dvor...@gmail.com> wrote:
> >> >> > Dear John,
> >> >> >
> >> >> > Thanks for such a prompt reply!
> >> >> >
> >> >> > Seems like something happens on the mon side, since there are no
> >> >> > mount-specific requests logged on the mds side (see below).
> >> >> > FYI, some hours ago I've disabled auth completely, but it didn't
> >> >> > help.
> >> >> >
> >> >> > The serialized metadata pool is 9.7G. I can try to compress it with
> >> >> > 7z,
> >> >> > then
> >> >> > setup rssh account for you to scp/rsync it.
> >> >> >
> >> >> > debug mds = 20
> >> >> > debug mon = 20
> >> >>
> >> >> Don't worry about the mon logs.  That MDS log snippet appears to be
> >> >> from several minutes earlier than the client's attempt to mount.
> >> >>
> >> >> In these cases it's generally simpler if you truncate all the logs,
> >> >> then attempt the mount, then send all the logs in full rather than
> >> >> snippets, so that we can be sure nothing is missing.
> >> >>
> >> >> Please also get the client log (use the fuse client with
> >> >> --debug-client=20).
> >> >>
> >> >> John
> >> >
> >> >
> >> >
> >> >
> >> > --
> >> >  Mykola
> >
> >
> >
> >
> > --
> >  Mykola
>



-- 
 Mykola
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to