[ceph-users] Re: Where are you running Ceph?

2025-07-02 Thread Hans van den Bogert
Lately I don't have a production cluster anymore. But my homelab is ceph with Rook. I'm running it since the alpha versions. Awesome piece of software. You'll need Kubernetes familiarity though. H On Wed, Jul 2, 2025, 18:17 Anthony Fecarotta wrote: > Hello, I was wondering if there are any stat

[ceph-users] Re: librbd 4k read/write?

2023-08-10 Thread Hans van den Bogert
On Thu, Aug 10, 2023, 17:36 Murilo Morais wrote: > Good afternoon everybody! > > I have the following scenario: > Pool RBD replication x3 > 5 hosts with 12 SAS spinning disks each > > I'm using exactly the following line with FIO to test: > fio -ioengine=libaio -direct=1 -invalidate=1 -name=test

[ceph-users] Re: weird performance issue on ceph

2022-07-26 Thread Hans van den Bogert
Is rook/CSI still not using efficient rbd object maps ? It could be that you issued a new benchmark while ceph was busy (inefficiently) removing the old rbd images. This is quite a stretch but could be worth exploring. On Mon, Jul 25, 2022, 21:42 Mark Nelson wrote: > I don't think so if this i

[ceph-users] Re: Performance in Proof-of-Concept cluster

2022-07-07 Thread Hans van den Bogert
Hi, Run a close to the metal benchmark on the disks first, just to see the theoretical ceiling. Also, rerun your benchmarks with random write, just to get more honest numbers as well. Based on the numbers so far, you seem to be getting 40k client iops @512 threads, due to 3x replication an

[ceph-users] Re: Ceph in kubernetes

2022-03-07 Thread Hans van den Bogert
Just to add to warm fuzzy feeling, although just in a homelab, I'm using rook for many years now, it's awesome. I trust* it with the family photo's on a selfhosted nextcloud. All on K8s/Ceph/RGW. Hans * I have backups though ;) On 3/7/22 09:45, Bo Thorsen wrote: Hi Nico and Janne, Thank y

[ceph-users] Re: Why you might want packages not containers for Ceph deployments

2021-11-17 Thread Hans van den Bogert
On 11/17/21 8:19 PM, Martin Verges wrote: There are still alternative solutions without the need for useless containers and added complexity. Stay away from that crap and you won't have a hard time. 😜 I don't have a problem with the containers *at all*. And with me probably a lot of users. But

[ceph-users] Re: v15-2-14-octopus no docker images on docker hub ceph/ceph ?

2021-08-20 Thread Hans van den Bogert
Strange to change this in a patch version. This could have all kinds of nasty implications during upgrade. Firewalls being a prominent one. Or am I misunderstanding the change? It's also not mentioned prominently in the release notes. Hans On 8/20/21 8:54 AM, Stefan Fleischmann wrote: If you

[ceph-users] Re: we're living in 2005.

2021-07-27 Thread Hans van den Bogert
Try to install a completely new ceph cluster from scratch on fresh installed LTS Ubuntu by this doc https://docs.ceph.com/en/latest/cephadm/install/ . Many interesting discoveries await you. You mean it doesn't work as advertised? If you did encounter bugs, have you made bug reports? -H __

[ceph-users] Re: PG_DAMAGED

2020-12-04 Thread Hans van den Bogert
Interesting, your comment implies that it is a replication issue, which does not stem from a faulty disk. But, couldn't the disk have a bit flip? Or would you argue that would've shown as a disk read error somewhere (because of ECC on the disk.) On 12/4/20 10:51 AM, Dan van der Ster wrote: No

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-16 Thread Hans van den Bogert
All good points (also replying to Frank Schilder) On 11/16/20 2:36 PM, Janne Johansson wrote: Not trying to say you don't understand this, but rather that people who run small ceph clusters tend to start out with R=2 or K+1 EC because the larger faults are easier to imagine. TBH, I think I did

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-16 Thread Hans van den Bogert
I think we're deviating from the original thread quite a bit and I would never argue that in a production environment with plenty OSDs you should go for R=2 or K+1, so my example cluster which happens to be 2+1 is a bit unlucky. However I'm interested in the following On 11/16/20 11:31 AM, Ja

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-16 Thread Hans van den Bogert
> With this profile you can only loose one OSD at a time, which is really > not that redundant. That's rather situation dependent. I don't have really large disks, so the repair time isn't that large. Further, my SLO isn't that high that I need 99.xxx% uptime, if 2 disks break in the same repair wi

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-11 Thread Hans van den Bogert
=default jerasure-per-chunk-alignment=false k=2 m=1 plugin=jerasure technique=reed_sol_van w=8 On 11/11/20 12:46 PM, Hans van den Bogert wrote: Can you show a `ceph osd tree` ? On 11/7/20 1:14 AM, seffyr...@gmail.com wrote: I've inherited a Ceph Octopus cluster that seems like it needs u

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-11 Thread Hans van den Bogert
Can you show a `ceph osd tree` ? On 11/7/20 1:14 AM, seffyr...@gmail.com wrote: I've inherited a Ceph Octopus cluster that seems like it needs urgent maintenance before data loss begins to happen. I'm the guy with the most Ceph experience on hand and that's not saying much. I'm experiencing mo

[ceph-users] Re: (Ceph Octopus) Repairing a neglected Ceph cluster - Degraded Data Reduncancy, all PGs degraded, undersized, not scrubbed in time

2020-11-11 Thread Hans van den Bogert
Hoping to learn from this myself, why will the current setup never work? On 11/11/20 10:29 AM, Robert Sander wrote: Am 07.11.20 um 01:14 schrieb seffyr...@gmail.com: I've inherited a Ceph Octopus cluster that seems like it needs urgent maintenance before data loss begins to happen. I'm the guy

[ceph-users] Re: Low Memory Nodes

2020-11-06 Thread Hans van den Bogert
> I already ordered more ram. Can i turn temporary down the RAM usage of > the OSDs to not get into that vicious cycle and just suffer small but > stable performance? Hi, Look at https://docs.ceph.com/en/latest/rados/configuration/bluestore-config-ref/#bluestore-config-reference and then spec

[ceph-users] Re: radosgw bucket subdomain with tls

2020-10-15 Thread Hans van den Bogert
To my knowledge that's a S3 client setting only. Mostly referred to as 'path-style'. Kickin' in open doors here, but you could opt for a wildcard certificate, that would be more future proof; Path-style will not be supported anymore, at least not at Amazon[1], so client support will fade away i

[ceph-users] Re: Setting up a small experimental CEPH network

2020-09-21 Thread Hans van den Bogert
Perhaps not SBCs, but I have 4x HP 6300s and have been running Kubernetes together with Ceph/Rook for more than 3 years. The HPs can be picked up around 80-120eu. I learned so much in 3 years, last time I had that was when I started using Linux. This was money well spent and still is, it runs n

[ceph-users] Re: Can 16 server grade ssd's be slower then 60 hdds? (no extra journals)

2020-08-31 Thread Hans van den Bogert
Perhaps both clusters have the same bottleneck and you perceive them as equally fast. Can you provide as much details of your clusters as possible? Also please show outputs of the tests that you've run. On 8/31/20 1:02 PM, VELARTIS Philipp Dürhammer wrote: I have a productive 60 osd's cluster

[ceph-users] Re: How to change the pg numbers

2020-08-18 Thread Hans van den Bogert
I don't think it might lead to more client slow requests if you set it to 4096 in one step, since there is a cap on how many recovery/backfill requests there can be per OSD at any given time. I am not sure though, but I am happy to be proved wrong by the senior members in this list :) Hans

[ceph-users] Re: Can you block gmail.com or so!!!

2020-08-06 Thread Hans van den Bogert
Please not a simple gmail block 8) Not everyone wants to use their corporate account or selfhost email or use a marginally better/worse commercial gmail alternative On 8/6/20 12:52 PM, Marc Roos wrote: Can you block gmail.com or so!!! ___ ceph

[ceph-users] Re: about replica size

2020-07-16 Thread Hans van den Bogert
The Ceph team has chosen to default to N+2 redundancy. This is analogous to RAID 6 (NOT RAID 1). I have to fundamentally disagree; Ceph's replica size of 3 behaves more like RAID1 with 3 mirrors, than it would RAID6. Raid 6 would be more in line with Ceph's erasure coding of X+2, e.g., 3+

[ceph-users] Re: rbd-mirror with snapshot, not doing any actaul data sync

2020-06-08 Thread Hans van den Bogert
transitioning to an image based snapshot config has some problems? But that's just guessing. Thanks for the follow-up though! Regards, Hans On Mon, Jun 8, 2020, 13:38 Jason Dillaman wrote: > On Sun, Jun 7, 2020 at 8:06 AM Hans van den Bogert > wrote: > > > > Hi lis

[ceph-users] rbd-mirror with snapshot, not doing any actaul data sync

2020-06-07 Thread Hans van den Bogert
Hi list, I've awaited octopus for a along time to be able to use mirror with snapshotting, since my setup does not allow for journal based mirroring. (K8s/Rook 1.3.x with ceph 15.2.2) However, I seem to be stuck, i've come to the point where on the cluster on which the (non-active) replicas shoul

[ceph-users] Re: Very bad performance on a ceph rbd pool via iSCSI to VMware esx

2020-05-29 Thread Hans van den Bogert
What are the specs of your nodes? And what specific harddisks are you using? On Fri, May 29, 2020, 18:41 Salsa wrote: > I have a 3 hosts, 10 4TB HDDs per host ceph storage set up. I deined a 3 > replica rbd pool and some images and presented them to a Vmware host via > ISCSI, but the write perfo

[ceph-users] Re: CEPH failure domain - power considerations

2020-05-28 Thread Hans van den Bogert
I would second that, there's no winning in this case for your requirements and single PSU nodes. If there were 3 feeds,  then yes; you could make an extra layer in your crushmap much like you would incorporate a rack topology in the crushmap. On 5/28/20 2:42 PM, Chris Palmer wrote: Immediate t

[ceph-users] snapshot-based mirroring explanation in docs

2020-05-28 Thread Hans van den Bogert
Hi list, When reading the documentation for the new way of mirroring [1], some questions arose, especially with the following sentence: > Since this mode is not point-in-time consistent, the full snapshot delta will need to be synced prior to use during a failover scenario. 1) I'm not sure