Tried below too but nothing works, any suggestion?

# mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs   -o 
name=admin,fs=van2.dev,noatime,_netdev,ms_mode=prefer-crc

mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you 
may not be authorized

# mount -t ceph :/ /cephfs/ -o noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev
mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you 
may not be authorized

Regards
Dev

> On Jul 15, 2025, at 2:17 PM, Devender Singh <deven...@netskrt.io> wrote:
> 
> Hello all 
> 
> I have cephfs working fine with running ads but still not able to mount, 
> getting error mount error: no mds (Metadata Server) is up
> 
> Below are details 
> 
> 
> # ceph osd dump | grep min_compat_client
> require_min_compat_client reef
> min_compat_client reef
> 
> # ceph balancer status
> {
>     "active": true,
>     "last_optimize_duration": "0:00:00.002882",
>     "last_optimize_started": "Tue Jul 15 21:16:41 2025",
>     "mode": "upmap-read",
>     "no_optimization_needed": true,
>     "optimize_result": "Unable to find further optimization, or pool(s) 
> pg_num is decreasing, or distribution is already perfect",
>     "plans": []
> }
> 
> root@van2-host06n:~# dpkg -l ceph-common
> Desired=Unknown/Install/Remove/Purge/Hold
> | Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
> |/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
> ||/ Name           Version       Architecture Description
> +++-==============-=============-============-==================================================================
> ii  ceph-common    19.2.2-1jammy amd64        common utilities to mount and 
> interact with a ceph storage cluster
> 
> root@van2-host06n:~# ceph fs status
> van2.dev - 0 clients
> ========
> RANK  STATE                  MDS                    ACTIVITY     DNS    INOS  
>  DIRS   CAPS
>  0    active  van2.dev.van2-host06n.wqttsi  Reqs:    0 /s    10     13     12 
>      0
>  1    active  van2.dev.van2-host07n.fngvlt  Reqs:    0 /s    10     13     11 
>      0
>         POOL            TYPE     USED  AVAIL
> cephfs.van2.dev.meta  metadata   230k  53.0T
> cephfs.van2.dev.data    data       0   53.0T
>            STANDBY MDS
> van2-dev.van2-host07n.jnacmh
> van2-dev.van2-host06n.wpzshe
> van2-dev.van2-host08n.slrfej
> MDS version: ceph version 19.2.2 (0eceb0defba60152a8182f7bd87d164b639885b8) 
> squid (stable)
> 
> 
> SYstemd file. 
> ===========
> 
> # cat cephfs.mount
> #
> # this mounts on /cephfs
> 
> [Unit]
> Description=mount /cephfs
> After=network-online.target
> Requires=network-online.target
> 
> [Mount]
> Where=/cephfs
> What=10.101.1.246:6789,10.101.1.247:6789,10.101.1.248:6789:/
> Type=ceph
> Options=fs=van2.dev,name=admin,noatime,_netdev,x-systemd.automount
> 
> [Install]
> WantedBy=multi-user.target
> 
> 
> root@van2-host06n:~# journalctl -u cephfs.mount
> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount error: no 
> mds (Metadata Server) is up. The clu>
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=>
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount error: no 
> mds (Metadata Server) is up. The clu>
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=>
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount error: no 
> mds (Metadata Server) is up. The clu>
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=>
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount error: no 
> mds (Metadata Server) is up. The clu>
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=>
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=killed, status=>
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'signal'.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Unit 
> process 1817555 (mount.ceph) remains >
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted mount 
> /cephfs.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Found 
> left-over process 1817555 (mount.cep>
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually 
> indicates unclean termination of a previous>
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount error: no 
> mds (Metadata Server) is up. The clu>
> lines 1-29...skipping...
> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=killed, status=15/TERM
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'signal'.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Unit 
> process 1817555 (mount.ceph) remains running after unit stopped.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted mount 
> /cephfs.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Found 
> left-over process 1817555 (mount.ceph) in control group while starting uni>
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually 
> indicates unclean termination of a previous run, or service implementation 
> defici>
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be>
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> lines 1-38...skipping...
> Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=killed, status=15/TERM
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'signal'.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Unit 
> process 1817555 (mount.ceph) remains running after unit stopped.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted mount 
> /cephfs.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Found 
> left-over process 1817555 (mount.ceph) in control group while starting unit. 
> Ignoring.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually 
> indicates unclean termination of a previous run, or service implementation 
> deficiencies.
> Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 19:58:50 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 19:59:51 van2-host06n.van2.example.com mount[1818287]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount: Mount 
> process exited, code=exited, status=32/n/a
> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount: 
> Failed with result 'exit-code'.
> Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: Failed to mount 
> mount /cephfs.
> Jul 15 20:05:43 van2-host06n.van2.example.com systemd[1]: Mounting mount 
> /cephfs...
> Jul 15 20:06:44 van2-host06n.van2.example.com mount[1818912]: mount error: no 
> mds (Metadata Server) is up. The cluster might be laggy, or you may not be 
> authorized
> lines 1-45
> 

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to