[ceph-users] multiple active MDS servers is OK for production Ceph clusters OR Not

2021-11-13 Thread huxia...@horebdata.cn
Dear Cephers, On reading a technical blog from Croit: https://croit.io/blog/ceph-performance-test-and-optimization It says the following: "It should be noted that it is still debated whether a configuration with multiple active MDS servers is OK for production Ceph clusters." Just wonder whet

[ceph-users] Re: OSDs not starting up

2021-11-13 Thread Stephen J. Thompson
Hello all, I am still digging into this. I disabled this osd from starting at boot, then rebooted the node. I then tried doing the following: sudo cephadm shell And the following was the result. To me it seems to indicate that the OSD drive is ok and can be decrypted but does still not run.

[ceph-users] Re: OSDs not starting up

2021-11-13 Thread 胡 玮文
Hi Stephen, I think this output you posted is pretty normal, there is no systemd in the container, thus the error. You still need to find the logs. You may try “sudo cephadm logs --name osd.0”. If that still fails, you should try to run the ceph-osd daemon manually. Weiwen Hu 发件人: Stephen J.

[ceph-users] Re: [Pacific] OSD Spec problem?

2021-11-13 Thread Eugen Block
Hi, There is no down OSD: yeah, I see the same in my lab, it also shows more daemons than I have for that custom spec. I think it's somehow showing also some failed attempts to deploy OSDs. In my case it's 4/8 but I only have 8 OSDs in total, but I changed the spec a couple of times in o

[ceph-users] Re: Handling node failures.

2021-11-13 Thread Сергей Процун
OSD will probably not sart if wal device is lost. You can give a try by removing the corresponding link to the block device from /var/lib/ceph/osd/ceph-ID/block.wal. Or it will use block.db for wal in that case. IOPS should be counted as well. I would go 1:3 way if we are considering IOPS. But its

[ceph-users] Re: Recursive delete hangs on cephfs

2021-11-13 Thread Sasha Litvak
I continued looking into the issue and have no idea what hinders the performance yet. However: 1. A client operating with kernel 5.3.0-42 (ubuntu 18.04) has no such problems. I delete a directory with hashed subdirs (00 - ff) and total space taken by files ~707MB spread across those 256 in 3.25 s

[ceph-users] Ceph Dashboard

2021-11-13 Thread Innocent Onwukanjo
Hi! While trying to set a domain name for my company's ceph cluster, I used Nginx on another server to reverse proxy the public IP address of the dashboard and the port 8443. The domain name is from CloudFlare. The dashboard came up for HTTP only but threw error for HTTPS and I could not log in. S