[ceph-users] Re: Ceph rbd mirror journal pool

2022-04-04 Thread Eugen Block
Hi samuel, I haven't used dedicated rbd journal pools so I don't have any comment on that. But there's an alternative to journal-based mirroring, you can also mirror based on snapshot [1]. Would this be an alternative for you to look deeper into? Regards, Eugen [1] https://docs.ceph.com/

[ceph-users] Re: can't deploy osd/db on nvme with other db logical volume

2022-04-04 Thread Eugen Block
Hi, this is handled by ceph-volume, do you find anything helpful in /var/log/ceph//ceph-volume.log? Also check the cephadm.log for any hints. Zitat von 彭勇 : we have a running ceph, 16.2.7, with SATA OSD and DB on nvme. and we insert some SATA to host, and the status of new host is AVAILA

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Fulvio Galeazzi
Hi again Dan! Things are improving, all OSDs are up, but still that one PG is down. More info below. On 4/1/22 19:26, Dan van der Ster wrote: Here is the output of "pg 85.12 query": https://pastebin.ubuntu.com/p/ww3JdwDXVd/ and its status (also showing the other 85.XX, for refere

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Dan van der Ster
Hi Fulvio, Yes -- that choose/chooseleaf thing is definitely a problem.. Good catch! I suggest to fix it and inject the new crush map and see how it goes. Next, in your crush map for the storage type, you have an error: # types type 0 osd type 1 host type 2 chassis type 3 rack type 4 row type 5

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Dan van der Ster
Could you share the output of `ceph pg 85.25 query`. Then increase the crush weights of those three osds to 0.1, then check if the PG goes active. (It is possible that the OSDs are not registering as active while they have weight zero). -- dan On Mon, Apr 4, 2022 at 10:01 AM Fulvio Galeazzi wr

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Fulvio Galeazzi
Yesss! Fixing the choose/chooseleaf thing did make the magic. :-) Thanks a lot for your support Dan. Lots of lessons learned from my side, I'm really grateful. All PGs are now active, will let Ceph rebalance. Ciao ciao Fulvio On 4/4/22 10:50, Dan van der Ster

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Dan van der Ster
Excellent news! After everything is back to active+clean, don't forget to set min_size to 4 :) have a nice day On Mon, Apr 4, 2022 at 10:59 AM Fulvio Galeazzi wrote: > > Yesss! Fixing the choose/chooseleaf thing did make the magic. :-) > >Thanks a lot for your support Dan. Lots of lessons l

[ceph-users] Re: PG down, due to 3 OSD failing

2022-04-04 Thread Dan van der Ster
BTW -- i've created https://tracker.ceph.com/issues/55169 to ask that we add some input validation. Injecting such a crush map would ideally not be possible. -- dan On Mon, Apr 4, 2022 at 11:02 AM Dan van der Ster wrote: > > Excellent news! > After everything is back to active+clean, don't forge

[ceph-users] Re: Recovery or recreation of a monitor rocksdb

2022-04-04 Thread Konstantin Shalygin
Hi, The fast way to fix quorum issue is redeploy ceph-mon service k Sent from my iPhone > On 1 Apr 2022, at 14:43, Victor Rodriguez wrote: > > Hello, > > Have a 3 node cluster using Proxmox + ceph version 14.2.22 (nautilus). After > a power failure one of the monitors does not start. The l

[ceph-users] Re: can't deploy osd/db on nvme with other db logical volume

2022-04-04 Thread Eugen Block
I know, that’s why I asked if the logs show why ceph-volume didn’t create the required logical volumes. Zitat von 彭勇 : thanks, we have done by following commands: ceph-volume lvm prepare --no-systemd --bluestore --data /dev/sdh --block.db /dev/nvme0n1 --block.db-size 7301032 we should

[ceph-users] Ceph Bluestore tweaks for Bcache

2022-04-04 Thread Richard Bade
Hi Everyone, I just wanted to share a discovery I made about running bluestore on top of Bcache in case anyone else is doing this or considering it. We've run Bcache under Filestore for a long time with good results but recently rebuilt all the osds on bluestore. This caused some degradation in per

[ceph-users] Re: quincy v17.2.0 QE Validation status

2022-04-04 Thread Josh Durgin
Hi Venky and Ernesto, how are the mount fix and grafana container build looking? Josh On Fri, Apr 1, 2022 at 8:22 AM Venky Shankar wrote: > On Thu, Mar 31, 2022 at 8:51 PM Venky Shankar wrote: > > > > Hi Yuri, > > > > On Wed, Mar 30, 2022 at 11:24 PM Yuri Weinstein > wrote: > > > > > > We me

[ceph-users] loosing one node from a 3-node cluster

2022-04-04 Thread Felix Joussein
Hi Everyone, I run a 3-node proxmox+ceph cluster in my home-lab serving as rdb storage for virtual machines for 2 years now. When I installed it, I did some testing to ensure, that when one node would fail, the remaining 2 nodes would keep the system up while the 3rd node is being replaced. Rec

[ceph-users] Re: loosing one node from a 3-node cluster

2022-04-04 Thread Wolfpaw - Dale Corse
Hi Felix, Where are your monitors located? Do you have one on each node? Dale Corse CEO/CTO Cell: 780-504-1756 24/7 NOC: 888-965-3729 www.wolfpaw.com

[ceph-users] Re: quincy v17.2.0 QE Validation status

2022-04-04 Thread Venky Shankar
Hey Josh, On Tue, Apr 5, 2022 at 4:34 AM Josh Durgin wrote: > > Hi Venky and Ernesto, how are the mount fix and grafana container build > looking? Currently running into various teuthology related issues when testing out the mount fix. We'll want a test run without these failures to be really

[ceph-users] Re: loosing one node from a 3-node cluster

2022-04-04 Thread Felix Joussein
Yes, each node has one monitor, manager and mds running. regards, Felix       Gesendet: Dienstag, 05. April 2022 um 03:00 Uhr Von: "Wolfpaw - Dale Corse" An: "'Felix Joussein'" , ceph-users@ceph.io Betreff: RE: [ceph-users] loosing one node from a 3-node cluster Hi Felix,   Where are yo