Hi Eugen, >From my experiences, to truely delete and recreate the Ceph FS *cephfs* file system I've done the following:
1. Remove the file system: ceph fs rm cephfs --yes-i-really-mean-it ceph fs rm_data_pool cephfs_data ceph fs rm_data_pool cephfs cephfs_data 2. Remove the associated pools: ceph osd pool delete cephfs_data cephfs_data --yes-i-really-really-mean-it ceph osd pool delete cephfs_metadata cephfs_metadata --yes-i-really-really-mean-it 3. Create a new default ceph file system: ceph osd pool create cephfs_data <int[0-]> {<int[0-]>} {<ruleset>} ceph osd pool create cephfs_metadata <int[0-]> {<int[0-]>} {<ruleset>} ceph fs new cephfs cephfs_metadata cephfs_data ceph fs set_default cephfs Not sure if this helps, as you may need to repeat the whole process from the start. Regards, Tom On Mon, Jan 8, 2018 at 2:19 PM, Eugen Block <ebl...@nde.ag> wrote: > Hi list, > > all this is on Ceph 12.2.2. > > An existing cephFS (named "cephfs") was backed up as a tar ball, then > "removed" ("ceph fs rm cephfs --yes-i-really-mean-it"), a new one created > ("ceph fs new cephfs cephfs-metadata cephfs-data") and the content restored > from the tar ball. According to the output of "ceph fs rm", the old cephFS > has only been deactivated, not deleted. Looking at the Ceph manager's web > interface, it now lists two entries "cephfs", one with id 0 (the "old" FS) > and id "1" (the currently active FS). > > When we try to run "ceph fs status", we get an error with a traceback: > > ---cut here--- > ceph3:~ # ceph fs status > Error EINVAL: Traceback (most recent call last): > File "/usr/lib64/ceph/mgr/status/module.py", line 301, in handle_command > return self.handle_fs_status(cmd) > File "/usr/lib64/ceph/mgr/status/module.py", line 219, in > handle_fs_status > stats = pool_stats[pool_id] > KeyError: (29L,) > ---cut here--- > > while this works: > > ---cut here--- > ceph3:~ # ceph fs ls > name: cephfs, metadata pool: cephfs-metadata, data pools: [cephfs-data ] > ---cut here--- > > We see the new id 1 when we run > > ---cut here--- > ceph3:~ # ceph fs get cephfs > Filesystem 'cephfs' (1) > fs_name cephfs > [...] > data_pools [35] > metadata_pool 36 > inline_data disabled > balancer > standby_count_wanted 1 > [...] > ---cut here--- > > The new FS seems to work properly and can be mounted from the clients, > just like before removing and rebuilding it. I'm not sure which other > commands would fail with this traceback, for now "ceph fs status" is the > only one. > > So it seems that having one deactivated cephFS has an impact on some of > the functions/commands. Is there any way to remove it properly? Most of the > commands work with the name, not the id of the FS, so it's difficult to > access the data from the old FS. Has anyone some insights on how to clean > this up? > > Regards, > Eugen > > -- > Eugen Block voice : +49-40-559 51 75 > NDE Netzdesign und -entwicklung AG fax : +49-40-559 51 77 > Postfach 61 03 15 > D-22423 Hamburg e-mail : ebl...@nde.ag > > Vorsitzende des Aufsichtsrates: Angelika Mozdzen > Sitz und Registergericht: Hamburg, HRB 90934 > Vorstand: Jens-U. Mozdzen > USt-IdNr. DE 814 013 983 > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- Thomas Bennett SKA South Africa Science Processing Team Office: +27 21 5067341 Mobile: +27 79 5237105
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com