Thanks for the link and the details

I have not installed cephadm and respectively ceph-volume on other hosts
but the controller

I am running Ubuntu 24.04.2 and noticed the versions available for both
cephadm and ceph-volume are either 19.2.3 or 19.2.0  - no 19.2.2.??
On the controller, I have 19.2.2

Should I install 19.2.3 or 19.2.0 ?

Steven

On Fri, 20 Feb 2026 at 08:29, Dario Graña via ceph-users <[email protected]>
wrote:

> I used the approach recommended by Eugen a few months ago and it worked
> well.
>
> I made a small change to the lv names, instead of osd.0 in
> # Create Logical Volume
> ceph:~ # lvcreate -L 5G -n ceph-osd0-db ceph-db
>
> I used
> ceph:~ # lvcreate -L 5G -n ceph-db-${OSD_FSID} ceph-db
>
> where ${OSD_FSID} is the following
> # Get the OSD's FSID
> [ceph: root@ceph /]# OSD_FSID=$(ceph-volume lvm list 0 | awk '/osd fsid/
> {print $3}')
> fb69ba54-4d56-4c90-a855-6b350d186df5
>
>
>
> On Thu, Feb 19, 2026 at 8:35 PM Eugen Block via ceph-users <
> [email protected]> wrote:
>
> > Maybe this helps:
> >
> >
> >
> https://heiterbiswolkig.blogs.nde.ag/2025/02/05/cephadm-migrate-db-wal-to-new-device/
> >
> > Note that cephadm shell is always executed locally on the host, not
> > remotely, there’s no ‚--host‘ parameter.
> >
> > Zitat von Steven Vacaroaia via ceph-users <[email protected]>:
> >
> > > Hi,
> > >
> > > I have a 7 nodes cluster deployed using cephadm with a mixt of HDD, SSD
> > and
> > > NVME
> > >
> > > I would like to reconfigure some of the HDD OSDs and put their DB on
> NVME
> > > There are only 6 x18 TB OSDs and there is a dedicate , "empty" 1.6 TB
> > NVME
> > > installed for this purpose
> > >
> > > The cluster is configured as EC 4+2 and  is not under heavy use
> > >
> > > I would appreciate if you could advise what is the best way of
> achieving
> > > that
> > >
> > > As far as I was able to find, there are these 2 options
> > >
> > > 1. remove and redeploy OSDs
> > > 2. use ceph-volume
> > >
> > > Option 1 is pretty straight forward but more time consuming
> > >
> > > For Option2 though I was not able to find any clear instructions
> > >
> > > The first thing I stumbled one was  how to use ceph-volume in a
> > > cephadm managed cluster on a host that does not have it
> > >
> > > ( "cephadm shell --host HOSTNAME -- ceph-volume ls"
> > > is complaining about "unrecognized arguments --host")
> > >
> > > Any help,links,suggestions will be greatly appreciated
> > >
> > > Steven
> > > _______________________________________________
> > > ceph-users mailing list -- [email protected]
> > > To unsubscribe send an email to [email protected]
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
> >
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to