>
> So the question is, is the extra multiple of 1000 incorrect in the 'OSD
> Overview' dashboard? Or am I not understanding things correctly?
latency_count is integer, returns numbers of samples, latency_sum is sum of
latencies from _count samples in seconds (so you multiply it by 1000 to get
ms)
Looks like I'm not alone in drop off scrub performance after last update? :)
Łukasz Borek
luk...@borek.org.pl
On Wed, 25 Jun 2025 at 11:58, Eugen Block wrote:
> Thanks Frédéric.
> The customer found the sticky flag, too. I must admit, I haven't used
> the mute command too often yet, usually I
Hi MJ,
Not sure if it's the right direction but backup software (like for example
Commvault) has an archiving option with stub support. You can configure the
s3 cloud library with CEPH s3 as a backend. I've never tested it - just a
thought.
On Mon, 5 May 2025 at 00:49, sacawulu wrote:
> Hi all,
>
> For upgrade the OS we have something similar, but exiting maintenance mode
> is broken (with 17.2.7) :(
> I need to check the tracker for similar issues and if I can't find
> anything, I will create a ticket
For 18.2.2 first maint exit command threw an exception for some reason. In
my patching
+1
I wasn't aware that this module is obsolete and was trying to start it a
few weeks ago.
We develop a home-made solution some time ago to monitor smart data from
both HDD (uncorrected errors, grown defect list) and SSD (WLC/TBW). But
keeping it up to date with non-unified disk models is a nigh
I never used fqdn this way, but there is an option for cephadm bootstrap
command
--allow-fqdn-hostname
allow hostname that is fully-qualified (contains
".")
Worth checking. Not sure what's behind.
Thanks
On Wed, 8 Jan 2025 at 12:14, Piotr Pisz wrote:
> Hi,
>
> We ad
I'd start with 3+2, so you have one node left for recovery in case one
fails. 6-node and 90 hdd per node sounds like a long recovery that needs to
be tested for sure.
On Mon, 9 Dec 2024 at 06:10, Phong Tran Thanh
wrote:
> Hi community,
>
> Please help with advice on selecting an erasure coding a
Don't think that the root cause has been found. I disabled versioning as I
have to manually remove expired objects using s3 client.
On Thu, 17 Oct 2024 at 17:50, Reid Guyett wrote:
> Hello,
>
> I am experiencing an issue where it seems all lifecycles are showing either
> PROCESSING or UNINITIAL.
Adding --zap to orch command cleans WALL logical volume :
ceph orch osd rm 37 --replace *--zap*
After replacement, new OSD is correctly created. Tested a few times with
18.2.4.
Thanks.
On Fri, 27 Sept 2024 at 19:31, Igor Fedotov wrote:
> Hi!
>
> I'm not an expert in the Ceph orchestrator but
Hi,
I'm having issue with lifecycle jobs for 18.2.4 cluster with versioning
enabled bucket.
/# radosgw-admin lc list
[
{
"bucket":
":mongobackup-prod:c3e0a369-71df-40f5-a5c0-51e859efe0e0.96754.1",
"shard": "lc.0",
"started": "Thu, 01 Jan 1970 00:00:00 GMT",
"s
>
> You could check if your devices support NVMe namespaces and create more
> than one namespace on the device.
Wow, tricky. Will give it a try.
Thanks!
Łukasz Borek
luk...@borek.org.pl
On Tue, 4 Jun 2024 at 16:26, Robert Sander
wrote:
> Hi,
>
> On 6/4/24 16:15, Anthony D'Atri wrote:
>
>
> I have certainly seen cases where the OMAPS have not stayed within the
> RocksDB/WAL NVME space and have been going down to disk.
How to monitor OMAPS size and if it does not get out of NVME?
The OP's number suggest IIRC like 120GB-ish for WAL+DB, though depending on
> workload spillover coul
Anthony, Darren
Thanks for response.
Answering your questions:
What is the network you have for this cluster?
25GB/s
> Is this a chassis with universal slots, or is that NVMe device maybe M.2
> or rear-cage?
12 * HDD via LSI jbod + 1 PCI NVME. Now it's 1.6TB, for the production plan
is to use
Hi Everyone,
I'm putting together a HDD cluster with an ECC pool dedicated to the backup
environment. Traffic via s3. Version 18.2, 7 OSD nodes, 12 * 12TB HDD +
1NVME each, 4+2 ECC pool.
Wondering if there is some general guidance for startup setup/tuning in
regards to s3 object size. Files are
NG_ON_PRIMARY UNFOUND DEGRADED RD_OPS RD WR_OPS WR
> >> USED COMPR UNDER COMPR
> >> default.rgw.buckets.index 0 B 11 0 33
> >>00 0 208 207 KiB 41 20 KiB 0 B
> >>0 B
> >>
>
7fb0a3df-9553-4a76-938d-d23711e67677.34162.1.2
> default.rgw.buckets.index/.dir.7fb0a3df-9553-4a76-938d-d23711e67677.34162.1.2
> mtime 2022-12-20T07:32:11.00-0500, size 0
>
>
> On Sun, Apr 7, 2024 at 10:06 PM Lukasz Borek wrote:
>
>> Hi!
>>
>> I'm
Hi!
I'm working on a POC cluster setup dedicated to backup app writing objects
via s3 (large objects, up to 1TB transferred via multipart upload process).
Initial setup is 18 storage nodes (12HDDs + 1 NVME card for DB/WALL) + EC
pool. Plan is to use cephadm.
I'd like to follow good practice and
17 matches
Mail list logo