Thanks for the tip. I will stay of the safe side and wait until it will be merged into master)
Many thanks for all your help. -Mykola On 19 November 2015 at 11:10, John Spray <jsp...@redhat.com> wrote: > On Thu, Nov 19, 2015 at 10:07 AM, Mykola Dvornik > <mykola.dvor...@gmail.com> wrote: > > I'm guessing in this context that "write data" possibly means creating > > a file (as opposed to writing to an existing file). > > > > Indeed. Sorry for the confusion. > > > > You've pretty much hit the limits of what the disaster recovery tools > > are currently capable of. What I'd recommend you do at this stage is > > mount your filesystem read-only, back it up, and then create a new > > filesystem and restore from backup. > > > > Ok. Is it somehow possible to have multiple FSs on the same ceph cluster? > > No, we want to do this but it's not there yet. Your scenario is one > of the motivations :-) > > (for the record multi-fs branch is > https://github.com/jcsp/ceph/commits/wip-multi-filesystems, which > works, but we'll probably go back and re-do the mon side of it before > finishing) > > John > > > > > > > On 19 November 2015 at 10:43, John Spray <jsp...@redhat.com> wrote: > >> > >> On Wed, Nov 18, 2015 at 9:21 AM, Mykola Dvornik > >> <mykola.dvor...@gmail.com> wrote: > >> > Hi John, > >> > > >> > It turned out that mds triggers an assertion > >> > > >> > mds/MDCache.cc: 269: FAILED assert(inode_map.count(in->vino()) == 0) > >> > > >> > on any attempt to write data to the filesystem mounted via fuse. > >> > >> I'm guessing in this context that "write data" possibly means creating > >> a file (as opposed to writing to an existing file). > >> > >> Currently, cephfs-data-scan injects inodes well enough that you can > >> read them, but it's not updating the inode table to reflect that the > >> recovered inodes are in use. As a result, when new files are created > >> they are probably trying to take inode numbers that are already in use > >> (by the recovered files), and as a result you're hitting this > >> assertion. The ticket for updating the inotable after injection of > >> recovered inodes is http://tracker.ceph.com/issues/12131 > >> > >> > Deleting data is still OK. > >> > > >> > I cannot really follow why duplicated inodes appear. > >> > > >> > Are there any ways to flush/reset the MDS cache? > >> > >> You've pretty much hit the limits of what the disaster recovery tools > >> are currently capable of. What I'd recommend you do at this stage is > >> mount your filesystem read-only, back it up, and then create a new > >> filesystem and restore from backup. > >> > >> I'm writing a patch to handle the particular case where someone needs > >> to update their inode table to mark all inodes as used up to some > >> maximum, but the chances are that after that you'll still run into > >> some other issue, until we've finished the tools to make it all the > >> way through this path. > >> > >> John > >> > >> > > >> > > >> > > >> > On 17 November 2015 at 13:26, John Spray <jsp...@redhat.com> wrote: > >> >> > >> >> On Tue, Nov 17, 2015 at 12:17 PM, Mykola Dvornik > >> >> <mykola.dvor...@gmail.com> wrote: > >> >> > Dear John, > >> >> > > >> >> > Thanks for such a prompt reply! > >> >> > > >> >> > Seems like something happens on the mon side, since there are no > >> >> > mount-specific requests logged on the mds side (see below). > >> >> > FYI, some hours ago I've disabled auth completely, but it didn't > >> >> > help. > >> >> > > >> >> > The serialized metadata pool is 9.7G. I can try to compress it with > >> >> > 7z, > >> >> > then > >> >> > setup rssh account for you to scp/rsync it. > >> >> > > >> >> > debug mds = 20 > >> >> > debug mon = 20 > >> >> > >> >> Don't worry about the mon logs. That MDS log snippet appears to be > >> >> from several minutes earlier than the client's attempt to mount. > >> >> > >> >> In these cases it's generally simpler if you truncate all the logs, > >> >> then attempt the mount, then send all the logs in full rather than > >> >> snippets, so that we can be sure nothing is missing. > >> >> > >> >> Please also get the client log (use the fuse client with > >> >> --debug-client=20). > >> >> > >> >> John > >> > > >> > > >> > > >> > > >> > -- > >> > Mykola > > > > > > > > > > -- > > Mykola > -- Mykola
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com