[ceph-users] Re: Remove ... something

2025-03-17 Thread Eugen Block
Hi, rbd list --pool erasure42 return a empty list. that's because the rbd metadata is stored in a replicated pool. You need to look into your replicated pool to delete the test image. If you don't find it with: rbd list --pool you can inspect the rbd_data prefix of the EC chunks and

[ceph-users] Re: My new osd is not normally ?

2025-03-17 Thread Eugen Block
Hi, is the balancer on? And which mode is enabled? ceph balancer status You definitely should split PGs, aim at 100 - 150 PGs per OSD at first. I would inspect the PG sizes of the new OSDs: ceph pg ls-by-osd 288 (column BYTES) and compare them with older OSDs. If you have very large PG siz

[ceph-users] Re: ceph-osd/bluestore using page cache

2025-03-17 Thread Joshua Baergen
Hey Brian, The setting you're looking for is bluefs_buffered_io. This is very much a YMMV setting, so it's best to test with both modes, but I usually recommend turning it off for all but omap-intensive workloads (e.g. RGW index) due to it causing writes to tend to be split up into smaller pieces.

[ceph-users] Re: Is it safe to set multiple OSD out across multiple failure domain?

2025-03-17 Thread Eugen Block
Before I replied, I wanted to renew my confidence and do a small test in a lab environment. I also created a k4m2 pool with host as failure-domain, started to write data chunks into it in a while loop and then marked three of the OSDs "out" simultaneously. After a few seconds of repeering b

[ceph-users] Re: Remove ... something

2025-03-17 Thread Albert Shih
Le 17/03/2025 à 13:46:58+, Eugen Block a écrit Hi, > > that's because the rbd metadata is stored in a replicated pool. You need to > look into your replicated pool to delete the test image. If you don't find > it with: Ok. I know that. But I didn't know I should remove the image metadata fr

[ceph-users] Re: Experience with 100G Ceph in Proxmox

2025-03-17 Thread Martin Konold
600 MB/s is rather slow. With 10 GBit/s I regularly measure 1,28 GB/s bandwidth even with a single connection.The issue is latency not bandwidth!The latenc is bound by the CPU serving the osds when using decent  NVMe storage is used.In an optimal world the network latency would be the limiting fact

[ceph-users] Re: ceph-osd/bluestore using page cache

2025-03-17 Thread Janne Johansson
Den mån 17 mars 2025 kl 14:48 skrev Joshua Baergen : > Hey Brian, > > The setting you're looking for is bluefs_buffered_io. This is very > much a YMMV setting, so it's best to test with both modes, but I > usually recommend turning it off for all but omap-intensive workloads > (e.g. RGW index) due

[ceph-users] My new osd is not normally ?

2025-03-17 Thread Yunus Emre Sarıpınar
I have 6 ssd sata and 12 osd per server in a 24 server cluster. This environment was created when it was in the natilus version. I switched this environment to the Octopus version 6 months ago. The cluster is working healthily. I added 8 new servers, I created 6 ssd sata and 12 osd on these serve

[ceph-users] Adding OSD nodes

2025-03-17 Thread Sinan Polat
Hello, I am currently managing a Ceph cluster that consists of 3 racks, each with 4 OSD nodes. Each node contains 24 OSDs. I plan to add three new nodes, one to each rack, to help alleviate the high OSD utilization. The current highest OSD utilization is 85%. I am concerned about the possibility

[ceph-users] Re: ceph-ansible LARGE OMAP in RGW pool

2025-03-17 Thread Frédéric Nass
Hi Danish, Have you tried restarting all RGWs or at least the one running the sync thread? Unless the synchronization thread generates more errors than you have time to clean up, you should be able to trim shard 31's error log entirely by looping through all IDs (--marker). Also, have you tr

[ceph-users] Ceph Tentacle release - dev freeze timeline

2025-03-17 Thread Yaarit Hatuka
Hi everyone, In previous discussions, the Ceph Steering Committee tentatively agreed on a Tentacle dev freeze around the end of March or mid-April. We would like to revisit this and check in with all the tech leads to assess the readiness level and ensure we're aligned on the timeline. Please pro

[ceph-users] Remove ... something

2025-03-17 Thread Albert Shih
Hi, I got two pool one replica 3 and another with EC 4+2 for rbd. During my test to access from a KVM hyperviseur I create a image with meta data in the replica 3 pool and the data in the EC 4+2. After I end the test (where everything works) I can't see anything in erasure42 pool. rbd lis