[ceph-users] Re: Custom CRUSH maps HOWTO?

2023-05-30 Thread Thorne Lawler
Thanks to Anthony D'Atri, Joshua Beaman and Nino Kotur. TL;DR- I need to ditch the spinning rust. As long as all my pools are using all the OSDs (currently necessary) this is not really a tuning problem- just a consequence of adding awful old recycled disks to my shiny NVME. To answer a few

[ceph-users] Ceph client version vs server version inter-operability

2023-05-30 Thread Mark Kirkwood
Hi, We are running a ceph cluster that is currently on Luminous. At this point most of our clients are also Luminous, but as we provision new client hosts we are using client versions that are more recent (e.g Octopus, Pacific and more recently Quincy). Is this safe? Is there a known list of

[ceph-users] Re: reef v18.1.0 QE Validation status

2023-05-30 Thread Travis Nielsen
Rook daily CI is passing against the image quay.io/ceph/daemon-base:latest-main-devel, which means the Reef release is looking good from Rook's perspective: With the Reef release we need to have the tags soon: quay.io/ceph/daemon-base:latest-reef-devel quay.io/ceph/ceph:v18 Guillaume, will th

[ceph-users] Re: BlueStore fragmentation woes

2023-05-30 Thread Fox, Kevin M
Ok, I restarted it May 25th, ~11:30, let it run over the long weekend and just checked on it. Data attached. May 21 11:24:34 cf8 ceph-4e4184f5-7733-453b-b72c-2b43422fd027-osd-183[2282674]: debug 2023-05-21T18:24:34.040+ 7f53603fc700 0 bluestore(/var/lib/ceph/osd/ceph-183) allocation stats

[ceph-users] Re: Custom CRUSH maps HOWTO?

2023-05-30 Thread Nino Kotur
What kind of pool are you using, or do you have different pools for different purposes... Do you have cephfs or rbd only pools etc... describe your setup. It is generally best practice to create new rules and apply them to pools and not to modify existing pools, but that is possible as well. Below

[ceph-users] Re: reef v18.1.0 QE Validation status

2023-05-30 Thread Ilya Dryomov
On Tue, May 30, 2023 at 6:54 PM Yuri Weinstein wrote: > > Details of this release are summarized here: > > https://tracker.ceph.com/issues/61515#note-1 > Release Notes - TBD > > Seeking approvals/reviews for: > > rados - Neha, Radek, Travis, Ernesto, Adam King (we still have to > merge https://git

[ceph-users] Re: slow mds requests with random read test

2023-05-30 Thread Patrick Donnelly
On Tue, May 30, 2023 at 8:42 AM Ben wrote: > > Hi, > > We are performing couple performance tests on CephFS using fio. fio is run > in k8s pod and 3 pods will be up running mounting the same pvc to CephFS > volume. Here is command line for random read: > fio -direct=1 -iodepth=128 -rw=randread -io

[ceph-users] reef v18.1.0 QE Validation status

2023-05-30 Thread Yuri Weinstein
Details of this release are summarized here: https://tracker.ceph.com/issues/61515#note-1 Release Notes - TBD Seeking approvals/reviews for: rados - Neha, Radek, Travis, Ernesto, Adam King (we still have to merge https://github.com/ceph/ceph/pull/51788 for the core) rgw - Casey fs - Venky orch -

[ceph-users] Re: [EXTERNAL] Custom CRUSH maps HOWTO?

2023-05-30 Thread Beaman, Joshua
I’m going to start by assuming your pool(s) are deployed with the default 3 replicas and a min_size of 2. The quickest and safest thing you can do to potentially realize some improvement, is set the primary-affinity for all of your HDD-based OSDs to zero. https://docs.ceph.com/en/quincy/rados/op

[ceph-users] Re: CEPH Version choice

2023-05-30 Thread Frank Schilder
Hi Marc, I uploaded all scripts and a rudimentary readme to https://github.com/frans42/cephfs-bench . I hope it is sufficient to get started. I'm afraid its very much tailored to our deployment and I can't make it fully configurable anytime soon. I hope it serves a purpose though - at least I

[ceph-users] Custom CRUSH maps HOWTO?

2023-05-30 Thread Thorne Lawler
Hi folks! I have a Ceph production 17.2.6 cluster with 6 machines in it - four newer, faster machines with 4x3.84TB NVME drives each, and two with 24x1.68TB SAS disks each. I know I should have done something smart with the CRUSH maps for this up front, but until now I have shied away from C

[ceph-users] RBD imagem mirroring doubt

2023-05-30 Thread Work Ceph
Hello guys, What would happen if we set up an RBD mirroring configuration, and in the target system (the system where the RBD image is mirrored) we create snapshots of this image? Would that cause some problems? Also, what happens if we delete the source RBD image? Would that trigger a deletion in

[ceph-users] Re: Important: RGW multisite bug may silently corrupt encrypted objects on replication

2023-05-30 Thread Casey Bodley
On Tue, May 30, 2023 at 8:22 AM Tobias Urdin wrote: > > Hello Casey, > > Thanks for the information! > > Can you please confirm that this is only an issue when using > “rgw_crypt_default_encryption_key” > config opt that says “testing only” in the documentation [1] to enable > encryption and not

[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-30 Thread Michel Jouvin
+1 Michel Le 30/05/2023 à 11:23, Frank Schilder a écrit : What I'm having in mind is if the command is already in history. A wrong history reference can execute a command with "--yes-i-really-mean-it" even though you really don't mean it. Been there. For an OSD this is maybe tolerable, but fo

[ceph-users] slow mds requests with random read test

2023-05-30 Thread Ben
Hi, We are performing couple performance tests on CephFS using fio. fio is run in k8s pod and 3 pods will be up running mounting the same pvc to CephFS volume. Here is command line for random read: fio -direct=1 -iodepth=128 -rw=randread -ioengine=libaio -bs=4k -size=1G -numjobs=5 -runtime=500 -gr

[ceph-users] Re: Important: RGW multisite bug may silently corrupt encrypted objects on replication

2023-05-30 Thread Tobias Urdin
Hello Casey, Thanks for the information! Can you please confirm that this is only an issue when using “rgw_crypt_default_encryption_key” config opt that says “testing only” in the documentation [1] to enable encryption and not when using Barbican or Vault as KMS or using SSE-C with the S3 API?

[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-30 Thread Frank Schilder
What I'm having in mind is if the command is already in history. A wrong history reference can execute a command with "--yes-i-really-mean-it" even though you really don't mean it. Been there. For an OSD this is maybe tolerable, but for an entire cluster ... not really. Some things need to be h

[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-30 Thread Nico Schottelius
Hey Frank, in regards to destroying a cluster, I'd suggest to reuse the old --yes-i-really-mean-it parameter, as it is already in use by ceph osd destroy [0]. Then it doesn't matter whether it's prod or not, if you really mean it ... ;-) Best regards, Nico [0] https://docs.ceph.com/en/latest/r

[ceph-users] Re: Seeking feedback on Improving cephadm bootstrap process

2023-05-30 Thread Frank Schilder
Hi, I would like to second Nico's comment. What happened to the idea that a deployment tool should be idempotent? The most natural option would be: 1) start install -> something fails 2) fix problem 3) repeat exact same deploy command -> deployment picks up at current state (including cleaning u