[ceph-users] recurring stat mismatch on PG

2022-10-08 Thread Frank Schilder
Hi all, I seem to observe something strange on an octopus(latest) cluster. We have a PG with a stat mismatch: 2022-10-08T10:06:22.206+0200 7fa3c48c7700 0 log_channel(cluster) log [DBG] : 19.1fff deep-scrub starts 2022-10-08T10:22:33.049+0200 7fa3c48c7700 -1 log_channel(cluster) log [ERR] : 19

[ceph-users] Re: recurring stat mismatch on PG

2022-10-08 Thread Dan van der Ster
Hi, Is that 15.2.17? It reminds me of this bug - https://tracker.ceph.com/issues/52705 - where an object with a particular name would hash to and cause a stat mismatch during scrub. But 15.2.17 should have the fix for that. Can you find the relevant osd log for more info? .. Dan On

[ceph-users] Re: recurring stat mismatch on PG

2022-10-08 Thread Frank Schilder
Hi Dan, yes, 15.2.17. I remember that case and was expecting it to be fixed. Here a relevant extract from the log: 2022-10-08T10:06:22.206+0200 7fa3c48c7700 0 log_channel(cluster) log [DBG] : 19.1fff deep-scrub starts 2022-10-08T10:22:33.049+0200 7fa3c48c7700 -1 log_channel(cluster) log [ERR]

[ceph-users] Re: recurring stat mismatch on PG

2022-10-08 Thread Dan van der Ster
Is that the log from the primary OSD? About the restart, you should probably just deep-scrub again to see the current state. .. Dan On Sat, Oct 8, 2022, 11:14 Frank Schilder wrote: > Hi Dan, > > yes, 15.2.17. I remember that case and was expecting it to be fixed. Here > a relevant extract f

[ceph-users] Re: iscsi deprecation

2022-10-08 Thread Lucian Petrut
Hi, As Ilya mentioned, RBD is natively supported on Windows since Pacific. Furthermore, we’re about to add Persistent Reservations support, which is going to enable Microsoft Failover Cluster and CSV support. Regards, Lucian From: Maged Mokhtar Sent: Friday, Octobe

[ceph-users] Re: recurring stat mismatch on PG

2022-10-08 Thread Frank Schilder
Yes, primary OSD. Extracted with grep -e scrub -e repair -e 19.1fff /var/log/ceph/ceph-osd.338.log and then only relevant lines copied. Yes, according to the case I should just run a deep-scrub and should see. I guess if this error was cleared on an aborted repair, this would be a new bug? I wi

[ceph-users] Re: recurring stat mismatch on PG

2022-10-08 Thread Dan van der Ster
It's not necessarily a bug... Running deep scrub again will just tell you the current state of the PG. That's safe any time. If it's comes back inconsistent again, I'd repair the PG again, let it finish completely, then scrub once again to double check that the repair worked. Thinking back, I've

[ceph-users] How to check which directory has ephemeral pinning set?

2022-10-08 Thread Frank Schilder
Hi all, I believe I enabled ephemeral pinning on a home dir, but I can't figure out how to check that its working. Here is my attempt: Set the flag: # setfattr -n ceph.dir.pin.distributed -v 1 /mnt/admin/cephfs/hpc/home Try to read it: # getfattr -n ceph.dir.pin.distributed /mnt/admin/cephfs/h

[ceph-users] Invalid crush class

2022-10-08 Thread Michael Thomas
In 15.2.7, how can I remove an invalid crush class? I'm surprised that I was able to create it in the first place: [root@ceph1 bin]# ceph osd crush class ls [ "ssd", "JBOD.hdd", "nvme", "hdd" ] [root@ceph1 bin]# ceph osd crush class ls-osd JBOD.hdd Invalid command: invalid cha

[ceph-users] LVM osds loose connection to disk

2022-10-08 Thread Frank Schilder
Hi all, we are facing a very annoying and disruptive problem. This happens only on a single type of disk: Vendor: TOSHIBA Product: PX05SMB040Y Revision: AS10 Compliance: SPC-4 User Capacity:400,088,457,216 bytes [400 GB] schedulers: mq-de