# mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/volumes/_nogroup/test /mnt/cephfs -o name=proxmoxvolfs,fs=van2-dev,secret=AQBw8X9o8b3GHRAAw58xzwqCP6Zt+WZZbjdb/A== mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized
No luck Regards Dev > On Jul 22, 2025, at 1:19 PM, Devender Singh <deven...@netskrt.io> wrote: > > Thanks Patrick > > But still the same issue > > # mount -t ceph > 10.101.1.246,10.101.1.247,10.101.1.248:/volumes/_nogroup/test/e393e142-977f-4d4a-817d-41bb721cb09c > /mnt/cephfs -o name=proxmoxvolfs,fs=van2-dev > > mount error: no mds (Metadata Server) is up. The cluster might be laggy, or > you may not be authorized > > Whereas ceph-fule is working… > > Regards > Dev > > > >> On Jul 22, 2025, at 2:37 AM, Patrick Begou >> <patrick.be...@univ-grenoble-alpes.fr> wrote: >> >> Le 22/07/2025 à 11:15, Patrick Begou a écrit : >>> I Devender, >>> >>> I went also in some mount troubles with Squid 19.2.2 after adding a second >>> cephfs but I'm new to Ceph. >>> I was getting the same message as yours: >>> >>> mount error: no mds (Metadata Server) is up. The cluster might be laggy, or >>> you may not be authorized >>> >>> I had to add the cephfs volume to avoid mismatch between clients and volume. >>> This is the command I use on the client with my 5 node ceph cluster to >>> mount the Cephfs storage by hand for testing: >>> >>> sudo mount -t ceph >>> 172.30.1.97,172.30.1.98,172.30.1.99,172.30.1.100,172.30.1.101:/ /mnt -o >>> name=*sebfolder*,fs=seb_vol -> my keefile is >>> /etc/ceph/ceph.client.sebfolder.keyring -> my cephfs volume is seb_vol >> >> Should be read: >> >> sudo mount -t ceph >> 172.30.1.97,172.30.1.98,172.30.1.99,172.30.1.100,172.30.1.101:/ /mnt -o >> name=sebfolder,fs=seb_vol >> >> * my keefile is /etc/ceph/ceph.client.sebfolder.keyring >> * my cephfs volume is seb_vol >> >> >>> >>> It works >>> >>> Patrick >>> >>> Le 21/07/2025 à 23:32, Devender Singh a écrit : >>>> Hello Dario >>>> >>>> Thanks for your reply. >>>> No luck using those too.. >>>> >>>> Also, I want to mount cephfs to proxmox, there also getting issue >>>> mounting.. >>>> >>>> # mount -t cephadmin@.van2-dev=/ /cephfs/ -o >>>> noatime,ms_mode=prefer-crc,_netdev >>>> >>>> mount error: no mds (Metadata Server) is up. The cluster might be laggy, >>>> or you may not be authorized >>>> >>>> >>>> I am using ceph version 19.2.2 and client is also 19.2.2. >>>> >>>> Regards >>>> Dev >>>> >>>>> On Jul 18, 2025, at 8:11 AM, Dario Graña<dgr...@pic.es> wrote: >>>>> >>>>> Hi Devender, >>>>> >>>>> I think the problem is that you are using the old syntax, try with the new >>>>> one [1] >>>>> >>>>> mount -t cephadmin@.van2-dev=/ /cephfs/ -o >>>>> noatime,ms_mode=prefer-crc,_netdev >>>>> >>>>> Regards. >>>>> >>>>> [1]https://www.google.com/url?q=https://www.google.com/url?q%3Dhttps://docs.ceph.com/en/squid/man/8/mount.ceph/%26source%3Dgmail-imap%26ust%3D1753457087000000%26usg%3DAOvVaw3RPvx2e7p7nZypeiYOOf9c&source=gmail-imap&ust=1753782129000000&usg=AOvVaw1blQFMK6YVh4Twc1kfXp_- >>>>> >>>>> On Wed, Jul 16, 2025 at 12:11 AM Devender Singh<deven...@netskrt.io >>>>> <mailto:deven...@netskrt.io>> wrote: >>>>> >>>>>> Tried below too but nothing works, any suggestion? >>>>>> >>>>>> # mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs -o >>>>>> name=admin,fs=van2.dev,noatime,_netdev,ms_mode=prefer-crc >>>>>> >>>>>> mount error: no mds (Metadata Server) is up. The cluster might be laggy, >>>>>> or you may not be authorized >>>>>> >>>>>> # mount -t ceph :/ /cephfs/ -o >>>>>> noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev >>>>>> mount error: no mds (Metadata Server) is up. The cluster might be laggy, >>>>>> or you may not be authorized >>>>>> >>>>>> Regards >>>>>> Dev >>>>>> >>>>>>> On Jul 15, 2025, at 2:17 PM, Devender Singh<deven...@netskrt.io >>>>>>> <mailto:deven...@netskrt.io>> wrote: >>>>>>> >>>>>>> Hello all >>>>>>> >>>>>>> I have cephfs working fine with running ads but still not able to mount, >>>>>> getting error mount error: no mds (Metadata Server) is up >>>>>>> Below are details >>>>>>> >>>>>>> >>>>>>> # ceph osd dump | grep min_compat_client >>>>>>> require_min_compat_client reef >>>>>>> min_compat_client reef >>>>>>> >>>>>>> # ceph balancer status >>>>>>> { >>>>>>> "active": true, >>>>>>> "last_optimize_duration": "0:00:00.002882", >>>>>>> "last_optimize_started": "Tue Jul 15 21:16:41 2025", >>>>>>> "mode": "upmap-read", >>>>>>> "no_optimization_needed": true, >>>>>>> "optimize_result": "Unable to find further optimization, or pool(s) >>>>>> pg_num is decreasing, or distribution is already perfect", >>>>>>> "plans": [] >>>>>>> } >>>>>>> >>>>>>> root@van2-host06n:~# dpkg -l ceph-common >>>>>>> Desired=Unknown/Install/Remove/Purge/Hold >>>>>>> | >>>>>> Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend >>>>>> >>>>>>> |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad) >>>>>>> ||/ Name Version Architecture Description >>>>>>> >>>>>> +++-==============-=============-============-================================================================== >>>>>> >>>>>>> ii ceph-common 19.2.2-1jammy amd64 common utilities to mount >>>>>> and interact with a ceph storage cluster >>>>>>> root@van2-host06n:~# ceph fs status >>>>>>> van2.dev - 0 clients >>>>>>> ======== >>>>>>> RANK STATE MDS ACTIVITY DNS >>>>>> INOS DIRS CAPS >>>>>>> 0 active van2.dev.van2-host06n.wqttsi Reqs: 0 /s 10 13 >>>>>> 12 0 >>>>>>> 1 active van2.dev.van2-host07n.fngvlt Reqs: 0 /s 10 13 >>>>>> 11 0 >>>>>>> POOL TYPE USED AVAIL >>>>>>> cephfs.van2.dev.meta metadata 230k 53.0T >>>>>>> cephfs.van2.dev.data data 0 53.0T >>>>>>> STANDBY MDS >>>>>>> van2-dev.van2-host07n.jnacmh >>>>>>> van2-dev.van2-host06n.wpzshe >>>>>>> van2-dev.van2-host08n.slrfej >>>>>>> MDS version: ceph version 19.2.2 >>>>>> (0eceb0defba60152a8182f7bd87d164b639885b8) squid (stable) >>>>>>> >>>>>>> SYstemd file. >>>>>>> =========== >>>>>>> >>>>>>> # cat cephfs.mount >>>>>>> # >>>>>>> # this mounts on /cephfs >>>>>>> >>>>>>> [Unit] >>>>>>> Description=mount /cephfs >>>>>>> After=network-online.target >>>>>>> Requires=network-online.target >>>>>>> >>>>>>> [Mount] >>>>>>> Where=/cephfs >>>>>>> What=10.101.1.246:6789,10.101.1.247:6789,10.101.1.248:6789:/ >>>>>>> Type=ceph >>>>>>> Options=fs=van2.dev,name=admin,noatime,_netdev,x-systemd.automount >>>>>>> >>>>>>> [Install] >>>>>>> WantedBy=multi-user.target >>>>>>> >>>>>>> >>>>>>> root@van2-host06n:~# journalctl -u cephfs.mount >>>>>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount >>>>>> error: no mds (Metadata Server) is up. The clu> >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=> >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount >>>>>> error: no mds (Metadata Server) is up. The clu> >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=> >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount >>>>>> error: no mds (Metadata Server) is up. The clu> >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=> >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount >>>>>> error: no mds (Metadata Server) is up. The clu> >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=> >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=killed, status=> >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'signal'. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Unit process 1817555 (mount.ceph) remains > >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted >>>>>> mount /cephfs. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Found left-over process 1817555 (mount.cep> >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually >>>>>> indicates unclean termination of a previous> >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount >>>>>> error: no mds (Metadata Server) is up. The clu> >>>>>>> lines 1-29...skipping... >>>>>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=killed, status=15/TERM >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'signal'. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Unit process 1817555 (mount.ceph) remains running after unit stopped. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted >>>>>> mount /cephfs. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Found left-over process 1817555 (mount.ceph) in control group while >>>>>> starting uni> >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually >>>>>> indicates unclean termination of a previous run, or service >>>>>> implementation >>>>>> defici> >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be> >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> lines 1-38...skipping... >>>>>>> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=killed, status=15/TERM >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'signal'. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Unit process 1817555 (mount.ceph) remains running after unit stopped. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted >>>>>> mount /cephfs. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Found left-over process 1817555 (mount.ceph) in control group while >>>>>> starting unit. Ignoring. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually >>>>>> indicates unclean termination of a previous run, or service >>>>>> implementation >>>>>> deficiencies. >>>>>>> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 19:58:50 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 19:59:51 van2-host06n.van2.example.com mount[1818287]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Mount process exited, code=exited, status=32/n/a >>>>>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount: >>>>>> Failed with result 'exit-code'. >>>>>>> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: Failed to >>>>>> mount mount /cephfs. >>>>>>> Jul 15 20:05:43 van2-host06n.van2.example.com systemd[1]: Mounting >>>>>> mount /cephfs... >>>>>>> Jul 15 20:06:44 van2-host06n.van2.example.com mount[1818912]: mount >>>>>> error: no mds (Metadata Server) is up. The cluster might be laggy, or you >>>>>> may not be authorized >>>>>>> lines 1-45 >>>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list --ceph-users@ceph.io >>>>>> To unsubscribe send an email toceph-users-le...@ceph.io >>>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list --ceph-users@ceph.io >>>>> To unsubscribe send an email toceph-users-le...@ceph.io >>>> _______________________________________________ >>>> ceph-users mailing list --ceph-users@ceph.io >>>> To unsubscribe send an email toceph-users-le...@ceph.io >>> >>> _______________________________________________ >>> ceph-users mailing list -- ceph-users@ceph.io >>> To unsubscribe send an email to ceph-users-le...@ceph.io >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@ceph.io <mailto:ceph-users@ceph.io> >> To unsubscribe send an email to ceph-users-le...@ceph.io >> <mailto:ceph-users-le...@ceph.io> _______________________________________________ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io