Hi,
there's no need to drain the OSDs before adding them back. Since
ceph-volume is in place you could scan and activate existing OSDs on a
reinstalled node. In cephadm there's a similar way [1]:
ceph cephadm osd activate ...
If you want to prevent the OSDs from being marked out you can ru
This strange stats, at least one object should be exists for this OMAP's. Try
to deep-scrub this PG, try to list objects in this PG `rados ls --pgid 6.2`
k
Sent from my iPhone
> On 18 Oct 2022, at 03:39, Yuji Ito wrote:
>
> Thank you for your reply.
>
>> the object need only for OMAP data,
Hi,
Eugen is right, no need to drain and readd your OSDs, this will be very
long and lead to a lot of unnecessary load balancing. You should
decouple replacing the failed OSD (where you should try to drain it,
after setting its primary-affinity to 0 to limit the I/O redirected to
it, and read
Hi
I've noticed that packages for Quincy 17.2.5 appeared in the debian 11
repo a few days ago. However I haven't seen any mention of it anywhere,
can't find any release notes, and the documentation still shows 17.2.4
as the latest version.
Is 17.2.5 documented and ready for use yet? It's a b
On 19/10/2022 11:26, Chris Palmer wrote:
I've noticed that packages for Quincy 17.2.5 appeared in the debian 11
repo a few days ago. However I haven't seen any mention of it
anywhere, can't find any release notes, and the documentation still
shows 17.2.4 as the latest version.
Is 17.2.5 docum
Hello,
I just bootstrap a new cluster with cephadm on ubuntu 20.04 LTS up to date, and
got some errors on gui
ceph -v
ceph version 17.2.5 (98318ae89f1a893a6ded3a640405cdbb33e08757) quincy (stable)
When I try to create a new erasure coded pool, with the gui, when I edit the ec
profil, the popu
There was a mail on d...@ceph.io that 17.2.4 missed a few backports, so
I presume 17.2.5 is a hotfix -- it's what 17.2.4 was supposed to be.
(And clearly the announcement is pending)
https://github.com/ceph/ceph/commits/v17.2.5
-- dan
On Wed, Oct 19, 2022 at 11:46 AM Christian Rohmann
wrote:
>
Hi all,
we need to prepare for temporary shut-downs of a part of our ceph cluster. I
have 2 questions:
1) What is the recommended procedure to temporarily shut down a ceph fs quickly?
2) How to avoid MON store log spam overflow (on octopus 15.2.17)?
To 1: Currently, I'm thinking about:
- fs fa
Hi Frank,
fs fail isn't ideal -- there's an 'fs down' command for this.
Here's a procedure we used, last used in the nautilus days:
1. If possible, umount fs from all the clients, so that all dirty
pages are flushed.
2. Prepare the ceph cluster: ceph osd set noout/noin
3. Wait until there is zer
(17.2.4, 3 replicated, Container install)
Hello,
since many of the information found in the WWW or books is outdated, I want
to ask which procedure is recommended to repair damaged PG with status
active+clean+inconsistent for Ceph Quincy.
IMHO, the best process for a pool with 3 replicas it woul
Hi,
you don't need to stop the OSDs, just query the inconsistent object,
here's a recent example (form an older cluster though):
---snip---
health: HEALTH_ERR
1 scrub errors
Possible data damage: 1 pg inconsistent
admin:~ # ceph health detail
HEALTH_ERR 1 scrub er
Hi Dan,
I know that "fs fail ..." is not ideal, but we will not have time for a clean
"fs down true" and wait for journal flush procedure to complete (on our cluster
this takes at least 20 minutes, which is way too long). My question is more
along the lines 'Is an "fs fail" destructive?', that
Thanks, I will try this the next time!
Am Mi., 19. Okt. 2022 um 13:50 Uhr schrieb Eugen Block :
> Hi,
>
> you don't need to stop the OSDs, just query the inconsistent object,
> here's a recent example (form an older cluster though):
>
> ---snip---
> health: HEALTH_ERR
> 1 scrub
Hi Eugen!
How are you?
Thank you for your help!
# ceph osd tree
ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
-1 174.62640 root default
-3 174.62640 host darkside2
0hdd 14.55220 osd.0 up 1.0 1.0
1hdd 14.55
Thanks a lot Alexander! I was going to take this approach, but wanted
to check if there's an alternative way.
Bzw setting norebalance flag will prevent destroying the OSD, as the
placements groups will not migrate to another OSD.
On 12.10.22 04:08, Alexander E. Patrakov wrote:
ср, 12 окт. 2022
Hi,
you mentioned three servers, you'll need those before your crush rule
can be applied because the default size is 3 and each PG has to be
placed on three different hosts (which you currently don't have).
There are a couple of ways to let the pool creation finish, but I
recommend to add
only one agenda item discussed today:
* 17.2.5 is almost ready, Upgrade testing has been completed in
upstream gibba and LRC clusters!
___
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io
Hello everyone,
Dan is correct that 17.2.5 is a hotfix release. There was a flaw in the
release process for 17.2.4 in which five commits were not included in the
release. The users mailing list will hear an official announcement about
this hotfix release later this week.
Thanks,
Laura
On Wed, Oc
I recently set up scheduled snapshots on my CephFS filesystem, and ever since
the cluster has been intermittently going into HEALTH_WARN with an
MDS_CLIENT_LATE_RELEASE notification.
Specifically:
[WARN] MDS_CLIENT_LATE_RELEASE: 1 clients failing to respond to capability
release
mds.[
Thank you, Xiubo - yes, checking my ceph.repo file as specified
at https://docs.ceph.com/en/pacific/install/get-packages/#rhel
reminded me that I had set the ceph-noarch repo to disabled,
because we didn't want ceph trying to update itself outside of
using cephadm fo
Hi Zach,
Seeing the `fs perf stats` output , it looks like you are not using the
latest build. Lots of enhancements are being done in cephfs-top recently.
Will suggest to use latest build for better results.
And regarding your `use_default_colors()` error, it looks like there some
issue with the un
Thank you for the reply, Neeraj - solved!
I was just going off of the document at
https://docs.ceph.com/en/quincy/cephfs/cephfs-top/ - I did not see
a specified terminal emulator listed, but yes, after switching my
Secure Shell program to XTerm, that seems to have
Good evening everyone.
I'm experiencing something strange on a cluster regarding monitoring. In
Grafana I can't see any data referring to the first Host, I've already
tried to redeploy Grafana and Prometheus, but the first Host never appears,
if I go to Dashboard -> Hosts -> Performance Detail the
23 matches
Mail list logo