Could it be a mismatch with the mount reference ?
What about trying:

mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/   /mnt/cephfs   -o name=proxmoxvolfs,fs=van2-dev,secret=AQBw8X9o8b3GHRAAw58xzwqCP6Zt+WZZbjdb/A==

Did you check you key file content (I've done this mistake with an empty key file generated by a wrong command 🙁)

Patrick

Le 22/07/2025 à 22:27, Devender Singh a écrit :
# mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/volumes/_nogroup/test /mnt/cephfs   -o name=proxmoxvolfs,fs=van2-dev,secret=AQBw8X9o8b3GHRAAw58xzwqCP6Zt+WZZbjdb/A== mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized

No luck

Regards
Dev

On Jul 22, 2025, at 1:19 PM, Devender Singh <deven...@netskrt.io> wrote:

Thanks Patrick

But still the same issue

# mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/volumes/_nogroup/test/e393e142-977f-4d4a-817d-41bb721cb09c   /mnt/cephfs   -o name=proxmoxvolfs,fs=van2-dev

mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized

Whereas ceph-fule is working…

Regards
Dev



On Jul 22, 2025, at 2:37 AM, Patrick Begou <patrick.be...@univ-grenoble-alpes.fr> wrote:

Le 22/07/2025 à 11:15, Patrick Begou a écrit :
I Devender,

I went also in some mount troubles with Squid 19.2.2 after adding a second cephfs but I'm new to Ceph.
I was getting the same message as yours:

mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized

I had to add the cephfs volume to avoid mismatch between clients and volume. This is the command I use on the client  with my 5 node ceph cluster to mount  the Cephfs storage by hand for testing:

sudo mount -t ceph 172.30.1.97,172.30.1.98,172.30.1.99,172.30.1.100,172.30.1.101:/ /mnt -o name=*sebfolder*,fs=seb_vol -> my keefile is /etc/ceph/ceph.client.sebfolder.keyring -> my cephfs volume is seb_vol

Should be read:

sudo mount -t ceph 172.30.1.97,172.30.1.98,172.30.1.99,172.30.1.100,172.30.1.101:/ /mnt -o name=sebfolder,fs=seb_vol

* my keefile is /etc/ceph/ceph.client.sebfolder.keyring
* my cephfs volume is seb_vol



It works

Patrick

Le 21/07/2025 à 23:32, Devender Singh a écrit :
Hello Dario

Thanks for your reply.
No luck using those too..

Also, I want to mount cephfs to proxmox, there also getting issue mounting..

# mount -t cephadmin@.van2-dev=/ /cephfs/ -o noatime,ms_mode=prefer-crc,_netdev

mount error: no mds (Metadata Server) is up. The cluster might be laggy, or you may not be authorized


I am using ceph version 19.2.2 and client is also 19.2.2.

Regards
Dev

On Jul 18, 2025, at 8:11 AM, Dario Graña<dgr...@pic.es> wrote:

Hi Devender,

I think the problem is that you are using the old syntax, try with the new
one [1]

mount -t cephadmin@.van2-dev=/ /cephfs/ -o
noatime,ms_mode=prefer-crc,_netdev

Regards.

[1]https://www.google.com/url?q=https://www.google.com/url?q%3Dhttps://docs.ceph.com/en/squid/man/8/mount.ceph/%26source%3Dgmail-imap%26ust%3D1753457087000000%26usg%3DAOvVaw3RPvx2e7p7nZypeiYOOf9c&source=gmail-imap&ust=1753782129000000&usg=AOvVaw1blQFMK6YVh4Twc1kfXp_- <https://www.google.com/url?q=https://www.google.com/url?q%3Dhttps://docs.ceph.com/en/squid/man/8/mount.ceph/%26source%3Dgmail-imap%26ust%3D1753457087000000%26usg%3DAOvVaw3RPvx2e7p7nZypeiYOOf9c&source=gmail-imap&ust=1753782129000000&usg=AOvVaw1blQFMK6YVh4Twc1kfXp_->

On Wed, Jul 16, 2025 at 12:11 AM Devender Singh<deven...@netskrt.io> wrote:

Tried below too but nothing works, any suggestion?

# mount -t ceph 10.101.1.246,10.101.1.247,10.101.1.248:/ /mnt/cephfs   -o
name=admin,fs=van2.dev,noatime,_netdev,ms_mode=prefer-crc

mount error: no mds (Metadata Server) is up. The cluster might be laggy,
or you may not be authorized

# mount -t ceph :/ /cephfs/ -o
noatime,ms_mode=prefer-crc,_netdev,fs=van2-dev
mount error: no mds (Metadata Server) is up. The cluster might be laggy,
or you may not be authorized

Regards
Dev

On Jul 15, 2025, at 2:17 PM, Devender Singh<deven...@netskrt.io> wrote:

Hello all

I have cephfs working fine with running ads but still not able to mount,
getting error mount error: no mds (Metadata Server) is up
Below are details


# ceph osd dump | grep min_compat_client
require_min_compat_client reef
min_compat_client reef

# ceph balancer status
{
    "active": true,
    "last_optimize_duration": "0:00:00.002882",
    "last_optimize_started": "Tue Jul 15 21:16:41 2025",
    "mode": "upmap-read",
    "no_optimization_needed": true,
    "optimize_result": "Unable to find further optimization, or pool(s)
pg_num is decreasing, or distribution is already perfect",
    "plans": []
}

root@van2-host06n:~# dpkg -l ceph-common
Desired=Unknown/Install/Remove/Purge/Hold
|
Status=Not/Inst/Conf-files/Unpacked/halF-conf/Half-inst/trig-aWait/Trig-pend
|/ Err?=(none)/Reinst-required (Status,Err: uppercase=bad)
||/ Name           Version Architecture Description

+++-==============-=============-============-==================================================================
ii ceph-common    19.2.2-1jammy amd64        common utilities to mount
and interact with a ceph storage cluster
root@van2-host06n:~# ceph fs status
van2.dev - 0 clients
========
RANK  STATE                  MDS ACTIVITY     DNS
INOS   DIRS   CAPS
0    active van2.dev.van2-host06n.wqttsi  Reqs: 0 /s    10     13
12 0
1    active van2.dev.van2-host07n.fngvlt  Reqs: 0 /s    10     13
11 0
POOL TYPE USED  AVAIL
cephfs.van2.dev.meta  metadata   230k 53.0T
cephfs.van2.dev.data    data       0 53.0T
STANDBY MDS
van2-dev.van2-host07n.jnacmh
van2-dev.van2-host06n.wpzshe
van2-dev.van2-host08n.slrfej
MDS version: ceph version 19.2.2
(0eceb0defba60152a8182f7bd87d164b639885b8) squid (stable)

SYstemd file.
===========

# cat cephfs.mount
#
# this mounts on /cephfs

[Unit]
Description=mount /cephfs
After=network-online.target
Requires=network-online.target

[Mount]
Where=/cephfs
What=10.101.1.246:6789,10.101.1.247:6789,10.101.1.248:6789:/
Type=ceph
Options=fs=van2.dev,name=admin,noatime,_netdev,x-systemd.automount

[Install]
WantedBy=multi-user.target


root@van2-host06n:~# journalctl -u cephfs.mount
Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
error: no mds (Metadata Server) is up. The clu>
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=>
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
error: no mds (Metadata Server) is up. The clu>
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=>
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
error: no mds (Metadata Server) is up. The clu>
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=>
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
error: no mds (Metadata Server) is up. The clu>
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=>
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=killed, status=>
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'signal'.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Unit process 1817555 (mount.ceph) remains >
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
mount /cephfs.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Found left-over process 1817555 (mount.cep>
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
indicates unclean termination of a previous>
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
error: no mds (Metadata Server) is up. The clu>
lines 1-29...skipping...
Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=killed, status=15/TERM
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'signal'.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Unit process 1817555 (mount.ceph) remains running after unit stopped.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
mount /cephfs.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Found left-over process 1817555 (mount.ceph) in control group while
starting uni>
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
defici>
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be>
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
lines 1-38...skipping...
Jul 15 19:31:10 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:32:12 van2-host06n.van2.example.com mount[1809096]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:32:12 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:34:59 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:35:59 van2-host06n.van2.example.com mount[1810584]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:35:59 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:36:16 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:37:17 van2-host06n.van2.example.com mount[1811498]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:37:17 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:49:38 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:50:40 van2-host06n.van2.example.com mount[1817486]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:50:40 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:51:58 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=killed, status=15/TERM
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'signal'.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Unit process 1817555 (mount.ceph) remains running after unit stopped.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Unmounted
mount /cephfs.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Found left-over process 1817555 (mount.ceph) in control group while
starting unit. Ignoring.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: This usually
indicates unclean termination of a previous run, or service implementation
deficiencies.
Jul 15 19:52:24 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:52:59 van2-host06n.van2.example.com mount[1817555]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:53:26 van2-host06n.van2.example.com mount[1817572]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:53:26 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:56:29 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:57:29 van2-host06n.van2.example.com mount[1817655]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:57:29 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 19:58:50 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 19:59:51 van2-host06n.van2.example.com mount[1818287]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Mount process exited, code=exited, status=32/n/a
Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: cephfs.mount:
Failed with result 'exit-code'.
Jul 15 19:59:51 van2-host06n.van2.example.com systemd[1]: Failed to
mount mount /cephfs.
Jul 15 20:05:43 van2-host06n.van2.example.com systemd[1]: Mounting
mount /cephfs...
Jul 15 20:06:44 van2-host06n.van2.example.com mount[1818912]: mount
error: no mds (Metadata Server) is up. The cluster might be laggy, or you
may not be authorized
lines 1-45

_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

_______________________________________________
ceph-users mailing list --ceph-users@ceph.io
To unsubscribe send an email toceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to