Hi,
Still suffering with the spilledover disks and stability issue in 3 of my
cluster after uploaded 6-900 millions objects to the cluster. (Octopus 15.2.10).
I’ve set memory target around 31-32GB so could that be that the spilledover
issue is coming from here?
So have mem target 31GB, next le
Hi,
As Christian said, osd_memory_target has nothing to do with rocksdb
levels and will certainly not decide when overspilling occurs. With that
said, I doubt any of us here ever gave 32GB of RAM to any OSD, so in
case you're not sure that OSDs can handle that much memory correctly, I
would a
Hi Christian,
Yes, I very clearly know what is spillover, read that github leveled document
in the last couple of days every day multiple time. (Answers for your questions
are after the cluster background information).
About the cluster:
- users are doing continuously put/head/delete operations
Hi,
I am going to migrate our ceph cluster to a new OS and I am trying to choose
the right one so that I won't have to replace it again when python4 becomes a
requirement mid-cycle [or whatever].
Has anyone seen any recommendations from the devs as to what distro they are
targeting for lets sa
Just an update for anyone that sees this it looks like Veeam doesn't index it's
content real well and as such when it offloads it, it is random IO which means
that the IOPS and throughput is not great and you really need to overbuild your
volumes (RAID) on your Veeam server to get any kind of pe
the /var/lib/ceph/osd/ceph-3/keyring is missing here ..
is there way to generate a keyring for osd.3 ?
thanks!
On Thu, Sep 30, 2021 at 1:18 AM Eugen Block wrote:
> Is the content of OSD.3 still available in the filesystem? If the
> answer is yes you can get the OSD's keyring from
>
> /var/lib/
Hello Ceph-Users,
I just switched from a single to a multi-site setup with all sorts of
bucket sizes and large differences in the number of stored objects.
Usually resharding is handled by RADOSGW automagically whenever a
certain object count per shard is reached, 100k per default.
The funct
Hello,
I'm trying to figure out what overlapping roots entails with the default
scale-down autoscaling profile in Ceph Pacific. My test setup involves a CRUSH
map that looks like:
ID=-1 | root=default
ID=-58 | rack=rack1
ID=-70 | host=ssd-1
|
I
On 30/09/2021 17:02, Christian Rohmann wrote:
Looking at my zones I can see that the master zone (converted from
previously single-site setup) has
bucket_index_max_shards=0
while the other, secondary zone has
bucket_index_max_shards=11
Should I align this and use "11" as the default s
Hi,
Thank you your input, had a look couple of osds, settings for these values the
following:
#ceph daemon osd.8 config get bluestore_min_alloc_size
{
"bluestore_min_alloc_size": "0"
}
# ceph daemon osd.8 config get bluestore_min_alloc_size_ssd
{
"bluestore_min_alloc_size_ssd": "4096"
}
On 9/30/2021 6:28 PM, Dave Piper wrote:
Thanks so much Igor, this is making a lot of sense.
First of all you're using custom 4K min_alloc_size which wasn't adapted before
Pacific, aren't you?
We've set bluestore_min_alloc_size = 4096 because we write a lot of small
objects. Various sources
Just PXE boot whatever OS you like at the time. If you need to switch to
another, a reboot is enough to switch OS. It's even possible without
containers, so absolute no problem at all.
--
Martin Verges
Managing director
Mobile: +49 174 9335695 | Chat: https://t.me/MartinVerges
croit GmbH, Frese
Hello All,
Please I would like to know if it is possibile two clusters can mirror rbd
to a third cluster.
In other words I have 3 separated ceph cluster : A B C.
I would like cluster A and cluster B can mirror some pools on cluster C.
Is it possible ?
Thanks
Ignazio
___
Wow, it works like a charm 😊 Thank you very much, I've tried in my lab, however
I need to update the cluster to 15.2.14, because in this version is available
the migrate. In error state not sure I can update though.
Very smooth:
num=14;ceph-volume lvm migrate --osd-id $num --osd-fsid `cat
/var/
On 9/30/21 14:48, Drew Weaver wrote:
Hi,
I am going to migrate our ceph cluster to a new OS and I am trying to choose
the right one so that I won't have to replace it again when python4 becomes a
requirement mid-cycle [or whatever].
Has anyone seen any recommendations from the devs as to what
Thanks so much Igor, this is making a lot of sense.
> First of all you're using custom 4K min_alloc_size which wasn't adapted
> before Pacific, aren't you?
We've set bluestore_min_alloc_size = 4096 because we write a lot of small
objects. Various sources recommended this as a solution to not ov
So we were forced out of our datacenter and had to move all our osd nodes to
new racks. Accordingly, we changed the crush map to reflect our OSD nodes' new
rack positions and that triggered a huge rebalance.
We're now getting OSD nearfull warnings on OSDs across all the racks. Started
off with
Hi,
If I set the min size of the pool to 4, will this pg be recovered? Or how I can
take out the cluster from health error like this?
Mark as lost seems risky based on some maillist experience, even if marked lost
after you still have issue, so curious what is the way to take the cluster out
fr
Ok, so I guess there are several things coming together that end up
making your life a bit miserable at the moment:
- PG scaling causing increase IO
- Ingesting large number of objects into RGW causing lots of IOPs
- Usual client traffic
- Your NVME that's being used for WAL/DB has only half the li
Hi,
On 9/30/21 18:02, Igor Fedotov wrote:
Using non-default min_alloc_size is generally not recommended. Primarily
due to perfomance penalties. Some side effects (like your ones) can be
observed as well. That's simple - non-default parameters generally mean
much worse QA coverage devs and les
That is - one thing you could do is to rate limit PUT requests on your
haproxy down to a level that your cluster is stable. At least that
gives you a chance to finish the PG scaling without OSDs dying on you
constantly
On Fri, 1 Oct 2021 at 11:56, Christian Wuerdig
wrote:
>
> Ok, so I guess there
Hi,
I'm curious - how did you tell that the separate WAL+DB volume was slowing
things down? I assume you did some benchmarking - is there any chance you'd
be willing to share results? (Or anybody else that's been in a similar
situation).
What sorts of devices are you using for the WAL+DB, versus
hi!
my cephfs is broken and i can not recover the mds-daemons. yesterday i have
update my ceph-cluster from v15 to v16 and i thought all working fine. next day
(today) some of my services goes down and throw errors, so i dig into the
problem and find my cephfs is down, all mds-daemons in standb
Hi Ernesto,
Thank you. That did the trick :)
Kind Regards,
- Karsten
On 29-09-2021 19:59, Ernesto Puerta wrote:
Hi Karsten,
Endpoints returning no data shouldn't be an issue. If all endpoints are scraped under the same job, they'll only differ on the "instance" label.
The "instance" lab
Hello All,
Please I would like to know if it is possibile two clusters can mirror rbd
to a third cluster.
In other words I have 3 separated ceph cluster : A B C.
I would like cluster A and cluster B can mirror some pools on cluster C.
Is it possible ?
Thanks
___
25 matches
Mail list logo