[ceph-users] Re: Re-install host OS on Ceph OSD node

2022-10-19 Thread Eugen Block
Hi, there's no need to drain the OSDs before adding them back. Since ceph-volume is in place you could scan and activate existing OSDs on a reinstalled node. In cephadm there's a similar way [1]: ceph cephadm osd activate ... If you want to prevent the OSDs from being marked out you can ru

[ceph-users] Re: How to remove remaining bucket index shard objects

2022-10-19 Thread Konstantin Shalygin
This strange stats, at least one object should be exists for this OMAP's. Try to deep-scrub this PG, try to list objects in this PG `rados ls --pgid 6.2` k Sent from my iPhone > On 18 Oct 2022, at 03:39, Yuji Ito wrote: > > Thank you for your reply. > >> the object need only for OMAP data,

[ceph-users] Re: Re-install host OS on Ceph OSD node

2022-10-19 Thread Michel Jouvin
Hi, Eugen is right, no need to drain and readd your OSDs, this will be very long and lead to a lot of unnecessary load balancing. You should decouple replacing the failed OSD (where you should try to drain it, after setting its primary-affinity to 0 to limit the I/O redirected to it, and read

[ceph-users] Status of Quincy 17.2.5 ?

2022-10-19 Thread Chris Palmer
Hi I've noticed that packages for Quincy 17.2.5 appeared in the debian 11 repo a few days ago. However I haven't seen any mention of it anywhere, can't find any release notes, and the documentation still shows 17.2.4 as the latest version. Is 17.2.5 documented and ready for use yet? It's a b

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-19 Thread Christian Rohmann
On 19/10/2022 11:26, Chris Palmer wrote: I've noticed that packages for Quincy 17.2.5 appeared in the debian 11 repo a few days ago. However I haven't seen any mention of it anywhere, can't find any release notes, and the documentation still shows 17.2.4 as the latest version. Is 17.2.5 docum

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-19 Thread Christophe BAILLON
Hello, I just bootstrap a new cluster with cephadm on ubuntu 20.04 LTS up to date, and got some errors on gui ceph -v ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable) When I try to create a new erasure coded pool, with the gui, when I edit the ec profil, the popu

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-19 Thread Dan van der Ster
There was a mail on d...@ceph.io that 17.2.4 missed a few backports, so I presume 17.2.5 is a hotfix -- it's what 17.2.4 was supposed to be. (And clearly the announcement is pending) https://github.com/ceph/ceph/commits/v17.2.5 -- dan On Wed, Oct 19, 2022 at 11:46 AM Christian Rohmann wrote: >

[ceph-users] Temporary shutdown of subcluster and cephfs

2022-10-19 Thread Frank Schilder
Hi all, we need to prepare for temporary shut-downs of a part of our ceph cluster. I have 2 questions: 1) What is the recommended procedure to temporarily shut down a ceph fs quickly? 2) How to avoid MON store log spam overflow (on octopus 15.2.17)? To 1: Currently, I'm thinking about: - fs fa

[ceph-users] Re: Temporary shutdown of subcluster and cephfs

2022-10-19 Thread Dan van der Ster
Hi Frank, fs fail isn't ideal -- there's an 'fs down' command for this. Here's a procedure we used, last used in the nautilus days: 1. If possible, umount fs from all the clients, so that all dirty pages are flushed. 2. Prepare the ceph cluster: ceph osd set noout/noin 3. Wait until there is zer

[ceph-users] Recommended procedure in case of OSD_SCRUB_ERRORS / PG_DAMAGED

2022-10-19 Thread E Taka
(17.2.4, 3 replicated, Container install) Hello, since many of the information found in the WWW or books is outdated, I want to ask which procedure is recommended to repair damaged PG with status active+clean+inconsistent for Ceph Quincy. IMHO, the best process for a pool with 3 replicas it woul

[ceph-users] Re: Recommended procedure in case of OSD_SCRUB_ERRORS / PG_DAMAGED

2022-10-19 Thread Eugen Block
Hi, you don't need to stop the OSDs, just query the inconsistent object, here's a recent example (form an older cluster though): ---snip--- health: HEALTH_ERR 1 scrub errors Possible data damage: 1 pg inconsistent admin:~ # ceph health detail HEALTH_ERR 1 scrub er

[ceph-users] Re: Temporary shutdown of subcluster and cephfs

2022-10-19 Thread Frank Schilder
Hi Dan, I know that "fs fail ..." is not ideal, but we will not have time for a clean "fs down true" and wait for journal flush procedure to complete (on our cluster this takes at least 20 minutes, which is way too long). My question is more along the lines 'Is an "fs fail" destructive?', that

[ceph-users] Re: Recommended procedure in case of OSD_SCRUB_ERRORS / PG_DAMAGED

2022-10-19 Thread E Taka
Thanks, I will try this the next time! Am Mi., 19. Okt. 2022 um 13:50 Uhr schrieb Eugen Block : > Hi, > > you don't need to stop the OSDs, just query the inconsistent object, > here's a recent example (form an older cluster though): > > ---snip--- > health: HEALTH_ERR > 1 scrub

[ceph-users] Re: Noob install: "rbd pool init" stuck

2022-10-19 Thread Renato Callado Borges
Hi Eugen! How are you? Thank you for your help! # ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -1 174.62640 root default -3 174.62640 host darkside2 0hdd 14.55220 osd.0 up 1.0 1.0 1hdd 14.55

[ceph-users] Re: encrypt OSDs after creation

2022-10-19 Thread Ali Akil
Thanks a lot  Alexander! I was going to take this approach, but wanted to check if there's an alternative way. Bzw setting norebalance flag will prevent destroying the OSD, as the placements groups will not migrate to another OSD. On 12.10.22 04:08, Alexander E. Patrakov wrote: ср, 12 окт. 2022

[ceph-users] Re: Noob install: "rbd pool init" stuck

2022-10-19 Thread Eugen Block
Hi, you mentioned three servers, you'll need those before your crush rule can be applied because the default size is 3 and each PG has to be placed on three different hosts (which you currently don't have). There are a couple of ways to let the pool creation finish, but I recommend to add

[ceph-users] Ceph Leadership Team Meeting Minutes - 2022 Oct 19

2022-10-19 Thread Casey Bodley
only one agenda item discussed today: * 17.2.5 is almost ready, Upgrade testing has been completed in upstream gibba and LRC clusters! ___ ceph-users mailing list -- ceph-users@ceph.io To unsubscribe send an email to ceph-users-le...@ceph.io

[ceph-users] Re: Status of Quincy 17.2.5 ?

2022-10-19 Thread Laura Flores
Hello everyone, Dan is correct that 17.2.5 is a hotfix release. There was a flaw in the release process for 17.2.4 in which five commits were not included in the release. The users mailing list will hear an official announcement about this hotfix release later this week. Thanks, Laura On Wed, Oc

[ceph-users] MDS_CLIENT_LATE_RELEASE after setting up scheduled CephFS snapshots

2022-10-19 Thread Edward R Huyer
I recently set up scheduled snapshots on my CephFS filesystem, and ever since the cluster has been intermittently going into HEALTH_WARN with an MDS_CLIENT_LATE_RELEASE notification. Specifically: [WARN] MDS_CLIENT_LATE_RELEASE: 1 clients failing to respond to capability release mds.[

[ceph-users] Re: Getting started with cephfs-top, how to install

2022-10-19 Thread Zach Heise (SSCC)
Thank you, Xiubo - yes, checking my ceph.repo file as specified at https://docs.ceph.com/en/pacific/install/get-packages/#rhel reminded me that I had set the ceph-noarch repo to disabled, because we didn't want ceph trying to update itself outside of using cephadm fo

[ceph-users] Re: Getting started with cephfs-top, how to install

2022-10-19 Thread Neeraj Pratap Singh
Hi Zach, Seeing the `fs perf stats` output , it looks like you are not using the latest build. Lots of enhancements are being done in cephfs-top recently. Will suggest to use latest build for better results. And regarding your `use_default_colors()` error, it looks like there some issue with the un

[ceph-users] Re: Getting started with cephfs-top, how to install

2022-10-19 Thread Zach Heise (SSCC)
Thank you for the reply, Neeraj - solved! I was just going off of the document at https://docs.ceph.com/en/quincy/cephfs/cephfs-top/ - I did not see a specified terminal emulator listed, but yes, after switching my Secure Shell program to XTerm, that seems to have

[ceph-users] Grafana without presenting data from the first Host

2022-10-19 Thread Murilo Morais
Good evening everyone. I'm experiencing something strange on a cluster regarding monitoring. In Grafana I can't see any data referring to the first Host, I've already tried to redeploy Grafana and Prometheus, but the first Host never appears, if I go to Dashboard -> Hosts -> Performance Detail the