On my test cluster I ran into the problem when creating the 2nd or third Ceph 
FS, that the actual mounting and adding to the PVE storage config failed with 
the following in the task log:

-----
creating data pool 'foobar_data'...
pool foobar_data: applying application = cephfs
pool foobar_data: applying pg_num = 32
creating metadata pool 'foobar_metadata'...
pool foobar_metadata: applying pg_num = 8
configuring new CephFS 'foobar'
Successfully create CephFS 'foobar'
Adding 'foobar' to storage configuration...
TASK ERROR: adding storage for CephFS 'foobar' failed, check log and add manually! create 
storage failed: mount error: Job failed. See "journalctl -xe" for details.
------

The matching syslog:
------
Oct 20 15:20:04 cephtest1 systemd[1]: Mounting /mnt/pve/foobar...
Oct 20 15:20:04 cephtest1 mount[45484]: mount error: no mds server is up or the 
cluster is laggy
Oct 20 15:20:04 cephtest1 systemd[1]: mnt-pve-foobar.mount: Mount process 
exited, code=exited, status=32/n/a
Oct 20 15:20:04 cephtest1 systemd[1]: mnt-pve-foobar.mount: Failed with result 
'exit-code'.
Oct 20 15:20:04 cephtest1 systemd[1]: Failed to mount /mnt/pve/foobar.
------

Adding the storage manually right after this worked fine. Seems like the MDS 
might not be fast enough all the time.



Regarding the removal of a Ceph FS we had an off list discussion which resulted 
in the following (I hope I am not forgetting something):

The process needs a few manual steps that are hard to automate:
 - disable storage (so pvestatd does not auto mount it again)
 - unmount on all nodes
 - stop standby and active (for this storage) MDS
        At this point, any still existing mount will be hanging
 - remove storage cfg and pools

Since at least some of those need to be done manually on the CLI, it might not even be 
worth it to have a "remove button" in the GUI but rather a well documented 
procedure in the manual and the actual removal as part of `pveceph`.


On 10/19/21 11:33, Dominik Csapak wrote:
this series support for multiple cephfs. no single patch fixes the bug,
so it's in no commit subject... (feel free to change the commit subject
when applying if you find one patch most appropriate?)

a user already can create multiple cephfs via 'pveceph' (or manually
with the ceph tools), but the ui does not support it properly

storage patch can be applied independently, it only adds a new parameter
that does nothing if not set.

manager:

patches 1,2 enables basic gui support for showing correct info
for multiple cephfs

patches 3,4,5 are mostly preparation for the following patches
(though 4 enables some additional checks that should not hurt either way)

patch 6 enables additional gui support for multiple fs

patch 7,8 depend on the storage patch

patch 9,10,11 are for actually creating multiple cephfs via the gui
so those can be left out if we do not want to support that

---
so if we only want to support basic display functionality, we could only apply
manager 1,2 & maybe 5+6

for being able to configure multiple cephfs on a ceph cluster, we'd need
storage 1/1 and manager 7,8

sorry that it's so complicated, if wanted, i can ofc reorder the patches
or send it in multiple series

pve-storage:

Dominik Csapak (1):
   cephfs: add support for multiple ceph filesystems

  PVE/Storage/CephFSPlugin.pm | 8 ++++++++
  1 file changed, 8 insertions(+)

pve-manager:

Dominik Csapak (11):
   api: ceph-mds: get mds state when multple ceph filesystems exist
   ui: ceph: catch missing version for service list
   api: cephfs: refactor {ls,create}_fs
   api: cephfs: more checks on fs create
   ui: ceph/ServiceList: refactor controller out
   ui: ceph/fs: show fs for active mds
   api: cephfs: add 'fs-name' for cephfs storage
   ui: storage/cephfs: make ceph fs selectable
   ui: ceph/fs: allow creating multiple cephfs
   api: cephfs: add destroy cephfs api call
   ui: ceph/fs: allow destroying cephfs

  PVE/API2/Ceph/FS.pm                      | 148 +++++++++--
  PVE/Ceph/Services.pm                     |  16 +-
  PVE/Ceph/Tools.pm                        |  51 ++++
  www/manager6/Makefile                    |   2 +
  www/manager6/Utils.js                    |   1 +
  www/manager6/ceph/FS.js                  |  52 +++-
  www/manager6/ceph/ServiceList.js         | 313 ++++++++++++-----------
  www/manager6/form/CephFSSelector.js      |  42 +++
  www/manager6/storage/CephFSEdit.js       |  25 ++
  www/manager6/window/SafeDestroyCephFS.js |  22 ++
  10 files changed, 492 insertions(+), 180 deletions(-)
  create mode 100644 www/manager6/form/CephFSSelector.js
  create mode 100644 www/manager6/window/SafeDestroyCephFS.js



_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel

Reply via email to