We are running octopus but will be upgrading to reef or squid in the next few 
weeks.  As part of that upgrade I am planning on switching over to using 
cephadm as well.

Part of what I am doing right now is going through and replacing old drives and 
removing some of our oldest nodes and replacing them with new ones…then I will 
convert the rest of the filestore osd over to bluestore so that I can upgrade.


One other question based on your suggestion below…my typical process of 
removing or replacing an osd involves the following:

ceph osd crush reweight osd.id 0.0

ceph osd out osd.id
service ceph stop osd.id
ceph osd crush remove osd.id
ceph auth del osd.id

ceph osd rm id



Does `ceph osd destroy` do something other than the last 3 commands above or am 
I just doing the same thing using multiple commands?  If I need to start 
issuing the destroy command as well I can.



Thank you.

Shain



From: Anthony D'Atri <a...@dreamsnake.net>
Date: Friday, October 18, 2024 at 9:01 AM
To: Shain Miley <smi...@npr.org>
Cc: ceph-users@ceph.io <ceph-users@ceph.io>
Subject: Re: [ceph-users] Influencing the osd.id when creating or replacing an 
osd
!-------------------------------------------------------------------|
  External Email - Use Caution

|-------------------------------------------------------------------!

What release are you running where ceph-deploy still works?

I get what you're saying, but really you should get used to OSD IDs being 
arbitrary.

    - ``ceph osd ls-tree <name>`` will output a list of OSD ids under
      the given CRUSH name (like a host or rack name).  This is useful
      for applying changes to entire subtrees.  For example, ``ceph
      osd down `ceph osd ls-tree rack1```.

This is useful for one-off scripts, where you can e.g. use it to get a list of 
OSDs on a given host.

Normally the OSD ID selected is the lowest-numbered unused one.  Which can 
either be an ID that has never been used before, or one that has been deleted.  
So if you delete an OSD entirely and redeploy, you may or may not get the same 
ID depending on the cluster’s history.

    - ``ceph osd destroy`` will mark an OSD destroyed and remove its
      cephx and lockbox keys.  However, the OSD id and CRUSH map entry
      will remain in place, allowing the id to be reused by a
      replacement device with minimal data rebalancing.

Destroying OSDs and redeploying them can help with what you’re after.

> On Oct 17, 2024, at 9:14 PM, Shain Miley <smi...@npr.org> wrote:
>
> Hello,
> I am still using ceph-deploy to add osd’s to my cluster.  From what I have 
> read ceph-deploy does not allow you to specify the osd.id when creating new 
> osds, however I am wondering if there is a way to influence the number that 
> ceph will assign for the next osd that is created.
>
> I know that it really shouldn’t matter what osd number gets assigned to the 
> disk but as the number of osd increases it is much easier to keep track of 
> where things are if you can control the id when replacing failed disks or 
> adding new nodes.
>
> Thank you,
> Shain
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to