[ceph-users] osd_memory_target=level0 ?

2021-09-30 Thread Szabo, Istvan (Agoda)
Hi, Still suffering with the spilledover disks and stability issue in 3 of my cluster after uploaded 6-900 millions objects to the cluster. (Octopus 15.2.10). I’ve set memory target around 31-32GB so could that be that the spilledover issue is coming from here? So have mem target 31GB, next le

[ceph-users] Re: osd_memory_target=level0 ?

2021-09-30 Thread Frédéric Nass
Hi, As Christian said, osd_memory_target has nothing to do with rocksdb levels and will certainly not decide when overspilling occurs. With that said, I doubt any of us here ever gave 32GB of RAM to any OSD, so in case you're not sure that OSDs can handle that much memory correctly, I would a

[ceph-users] Re: osd_memory_target=level0 ?

2021-09-30 Thread Szabo, Istvan (Agoda)
Hi Christian, Yes, I very clearly know what is spillover, read that github leveled document in the last couple of days every day multiple time. (Answers for your questions are after the cluster background information). About the cluster: - users are doing continuously put/head/delete operations

[ceph-users] Migrating CEPH OS looking for suggestions

2021-09-30 Thread Drew Weaver
Hi, I am going to migrate our ceph cluster to a new OS and I am trying to choose the right one so that I won't have to replace it again when python4 becomes a requirement mid-cycle [or whatever]. Has anyone seen any recommendations from the devs as to what distro they are targeting for lets sa

[ceph-users] Re: RGW performance as a Veeam capacity tier

2021-09-30 Thread Drew Weaver
Just an update for anyone that sees this it looks like Veeam doesn't index it's content real well and as such when it offloads it, it is random IO which means that the IOPS and throughput is not great and you really need to overbuild your volumes (RAID) on your Veeam server to get any kind of pe

[ceph-users] Re: osd marked down

2021-09-30 Thread Abdelillah Asraoui
the /var/lib/ceph/osd/ceph-3/keyring is missing here .. is there way to generate a keyring for osd.3 ? thanks! On Thu, Sep 30, 2021 at 1:18 AM Eugen Block wrote: > Is the content of OSD.3 still available in the filesystem? If the > answer is yes you can get the OSD's keyring from > > /var/lib/

[ceph-users] bucket_index_max_shards vs. no resharding in multisite? How to brace RADOS for huge buckets

2021-09-30 Thread Christian Rohmann
Hello Ceph-Users, I just switched from a single to a multi-site setup with all sorts of bucket sizes and large differences in the number of stored objects. Usually resharding is handled by RADOSGW automagically whenever a certain object count per shard is reached, 100k per default. The funct

[ceph-users] Trying to understand what overlapped roots means in pg_autoscale's scale-down mode

2021-09-30 Thread Andrew Gunnerson
Hello, I'm trying to figure out what overlapping roots entails with the default scale-down autoscaling profile in Ceph Pacific. My test setup involves a CRUSH map that looks like: ID=-1 | root=default ID=-58 | rack=rack1 ID=-70 | host=ssd-1 | I

[ceph-users] Re: bucket_index_max_shards vs. no resharding in multisite? How to brace RADOS for huge buckets

2021-09-30 Thread Christian Rohmann
On 30/09/2021 17:02, Christian Rohmann wrote: Looking at my zones I can see that the master zone (converted from previously single-site setup) has  bucket_index_max_shards=0 while the other, secondary zone has  bucket_index_max_shards=11 Should I align this and use "11" as the default s

[ceph-users] Re: osd_memory_target=level0 ?

2021-09-30 Thread Szabo, Istvan (Agoda)
Hi, Thank you your input, had a look couple of osds, settings for these values the following: #ceph daemon osd.8 config get bluestore_min_alloc_size { "bluestore_min_alloc_size": "0" } # ceph daemon osd.8 config get bluestore_min_alloc_size_ssd { "bluestore_min_alloc_size_ssd": "4096" }

[ceph-users] Re: [EXTERNAL] RE: OSDs flapping with "_open_alloc loaded 132 GiB in 2930776 extents available 113 GiB"

2021-09-30 Thread Igor Fedotov
On 9/30/2021 6:28 PM, Dave Piper wrote: Thanks so much Igor, this is making a lot of sense. First of all you're using custom 4K min_alloc_size which wasn't adapted before Pacific, aren't you? We've set bluestore_min_alloc_size = 4096 because we write a lot of small objects. Various sources

[ceph-users] Re: Migrating CEPH OS looking for suggestions

2021-09-30 Thread Martin Verges
Just PXE boot whatever OS you like at the time. If you need to switch to another, a reboot is enough to switch OS. It's even possible without containers, so absolute no problem at all. -- Martin Verges Managing director Mobile: +49 174 9335695 | Chat: https://t.me/MartinVerges croit GmbH, Frese

[ceph-users] Rbd mirror

2021-09-30 Thread Ignazio Cassano
Hello All, Please I would like to know if it is possibile two clusters can mirror rbd to a third cluster. In other words I have 3 separated ceph cluster : A B C. I would like cluster A and cluster B can mirror some pools on cluster C. Is it possible ? Thanks Ignazio ___

[ceph-users] Re: is it possible to remove the db+wal from an external device (nvme)

2021-09-30 Thread Szabo, Istvan (Agoda)
Wow, it works like a charm 😊 Thank you very much, I've tried in my lab, however I need to update the cluster to 15.2.14, because in this version is available the migrate. In error state not sure I can update though. Very smooth: num=14;ceph-volume lvm migrate --osd-id $num --osd-fsid `cat /var/

[ceph-users] Re: Migrating CEPH OS looking for suggestions

2021-09-30 Thread Stefan Kooman
On 9/30/21 14:48, Drew Weaver wrote: Hi, I am going to migrate our ceph cluster to a new OS and I am trying to choose the right one so that I won't have to replace it again when python4 becomes a requirement mid-cycle [or whatever]. Has anyone seen any recommendations from the devs as to what

[ceph-users] Re: [EXTERNAL] RE: OSDs flapping with "_open_alloc loaded 132 GiB in 2930776 extents available 113 GiB"

2021-09-30 Thread Dave Piper
Thanks so much Igor, this is making a lot of sense. > First of all you're using custom 4K min_alloc_size which wasn't adapted > before Pacific, aren't you? We've set bluestore_min_alloc_size = 4096 because we write a lot of small objects. Various sources recommended this as a solution to not ov

[ceph-users] ceph rebalance behavior

2021-09-30 Thread Chu, Vincent
So we were forced out of our datacenter and had to move all our osd nodes to new racks. Accordingly, we changed the crush map to reflect our OSD nodes' new rack positions and that triggered a huge rebalance. We're now getting OSD nearfull warnings on OSDs across all the racks. Started off with

[ceph-users] dealing with unfound pg in 4:2 ec pool

2021-09-30 Thread Szabo, Istvan (Agoda)
Hi, If I set the min size of the pool to 4, will this pg be recovered? Or how I can take out the cluster from health error like this? Mark as lost seems risky based on some maillist experience, even if marked lost after you still have issue, so curious what is the way to take the cluster out fr

[ceph-users] Re: osd_memory_target=level0 ?

2021-09-30 Thread Christian Wuerdig
Ok, so I guess there are several things coming together that end up making your life a bit miserable at the moment: - PG scaling causing increase IO - Ingesting large number of objects into RGW causing lots of IOPs - Usual client traffic - Your NVME that's being used for WAL/DB has only half the li

[ceph-users] Re: [EXTERNAL] RE: OSDs flapping with "_open_alloc loaded 132 GiB in 2930776 extents available 113 GiB"

2021-09-30 Thread Stefan Kooman
Hi, On 9/30/21 18:02, Igor Fedotov wrote: Using non-default min_alloc_size is generally not recommended. Primarily due to perfomance penalties. Some side effects (like your ones) can be observed as well. That's simple - non-default parameters generally mean much worse QA coverage devs and les

[ceph-users] Re: osd_memory_target=level0 ?

2021-09-30 Thread Christian Wuerdig
That is - one thing you could do is to rate limit PUT requests on your haproxy down to a level that your cluster is stable. At least that gives you a chance to finish the PG scaling without OSDs dying on you constantly On Fri, 1 Oct 2021 at 11:56, Christian Wuerdig wrote: > > Ok, so I guess there

[ceph-users] Re: is it possible to remove the db+wal from an external device (nvme)

2021-09-30 Thread Victor Hooi
Hi, I'm curious - how did you tell that the separate WAL+DB volume was slowing things down? I assume you did some benchmarking - is there any chance you'd be willing to share results? (Or anybody else that's been in a similar situation). What sorts of devices are you using for the WAL+DB, versus

[ceph-users] MDS: corrupted header/values: decode past end of struct encoding: Malformed input

2021-09-30 Thread von Hoesslin, Volker
hi! my cephfs is broken and i can not recover the mds-daemons. yesterday i have update my ceph-cluster from v15 to v16 and i thought all working fine. next day (today) some of my services goes down and throw errors, so i dig into the problem and find my cephfs is down, all mds-daemons in standb

[ceph-users] Re: prometheus - figure out which mgr (metrics endpoint) that is active

2021-09-30 Thread Karsten Nielsen
Hi Ernesto, Thank you. That did the trick :) Kind Regards, - Karsten On 29-09-2021 19:59, Ernesto Puerta wrote: Hi Karsten, Endpoints returning no data shouldn't be an issue. If all endpoints are scraped under the same job, they'll only differ on the "instance" label. The "instance" lab

[ceph-users] urgent question about rdb mirror

2021-09-30 Thread Ignazio Cassano
Hello All, Please I would like to know if it is possibile two clusters can mirror rbd to a third cluster. In other words I have 3 separated ceph cluster : A B C. I would like cluster A and cluster B can mirror some pools on cluster C. Is it possible ? Thanks ___