[ceph-users] Re: Replace OSD while cluster is recovering?

2025-03-03 Thread Gregory Orange
On 1/3/25 04:12, Gustavo Garcia Rondina wrote: > Thank you for the suggestion. I started `ceph pg repair {pgid}` inconsistent > PGs but so far, no effect to be seen. Is it possible to monitor the progress > of the repairs? With `ceph progress` I can't see it, and for some reason > `ceph -w` is h

[ceph-users] Deleting a pool with data

2025-03-03 Thread Richard Bade
Hi Everyone, We're reducing back down from multisite to a single rgw zone. This will mean that some pools will be unused so I'd like to delete them. However there are some objects and data remaining in the pool even though the buckets are all deleted. It's just shadow objects. All the actual data h

[ceph-users] March 3rd Ceph Steering Committee Meeting Notes

2025-03-03 Thread Laura Flores
Hi all, Here is a summary of what was discussed at today's CSC Meeting: - Deciding on the best way for a community vote on the name for U release - Tentacle Kickoff PR merged - https://tracker.ceph.com/issues/69953 being investigated by fs team - v18.2.5 status: Working on including a few outstan

[ceph-users] Re: Severe Latency Issues in Ceph Cluster

2025-03-03 Thread Brett Niver
I think using the dashboard you can check for incorrect MTU settings, that is sometimes an issue. Brett On Mon, Mar 3, 2025 at 12:42 PM Ramin Najjarbashi < ramin.najarba...@gmail.com> wrote: > The Ceph version is 17.2.7. > > > • OSDs are a mix of SSD and HDD, with DB/WAL colocated on the same OS

[ceph-users] Re: Severe Latency Issues in Ceph Cluster

2025-03-03 Thread Ramin Najjarbashi
The Ceph version is 17.2.7. • OSDs are a mix of SSD and HDD, with DB/WAL colocated on the same OSDs. • SSDs are used for metadata and index pools with replication 3. • HDDs store the data pool using EC 4+2. Interestingly, the same issue has appeared on another cluster where DB/WAL is placed o

[ceph-users] Re: Replace OSD while cluster is recovering?

2025-03-03 Thread Frédéric Nass
Hi Gustavo, In your situation, I would run a 'ceph pg repair {pgid}' on each one of these inconsistent PGs reported by 'ceph health detail' so they eventually get active+clean ASAP. And I would leave scrubbing enabled and set osd_scrub_auto_repair to true with a 'ceph config set osd osd_scrub_

[ceph-users] Re: Severe Latency Issues in Ceph Cluster

2025-03-03 Thread Stefan Kooman
On 01-03-2025 15:10, Ramin Najjarbashi wrote: Hi We are currently facing severe latency issues in our Ceph cluster, particularly affecting read and write operations. At times, write operations completely stall, leading to significant service degradation. Below is a detailed breakdown of the issue