Hi Eugen,

Thanks again for taking the time to help us with this.

Here are answers to your questions:

Nothing stands out from the mgr logs. Even when `ceph orch device ls` stops 
reporting, it still shows a claim on the osd in the logs when I run it:

Sep 27 09:39:24 ceph-mon3 bash[476409]: debug 2024-09-27T13:39:24.731+0000 
7fd4dc6fa700  0 [cephadm INFO root] Found osd claims -> {'ceph-osd31': ['88']}
Sep 27 09:39:24 ceph-mon3 bash[476409]: debug 2024-09-27T13:39:24.731+0000 
7fd4dc6fa700  0 log_channel(cephadm) log [INF] : Found osd claims -> 
{'ceph-osd31': ['88']}
Sep 27 09:39:24 ceph-mon3 bash[476409]: debug 2024-09-27T13:39:24.731+0000 
7fd4dc6fa700  0 [cephadm INFO cephadm.services.osd] Found osd claims for 
drivegroup ceph-osd31 -> {'ceph-osd31': ['88']}
Sep 27 09:39:24 ceph-mon3 bash[476409]: debug 2024-09-27T13:39:24.731+0000 
7fd4dc6fa700  0 log_channel(cephadm) log [INF] : Found osd claims for 
drivegroup ceph-osd31 -> {'ceph-osd31': ['88’]}


Here’s a sample of mgr logs right after a mgr failover (I’ve filtered out some 
noise from pgmap, prometheus, pg_autoscaler, balancer, and progress):

Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.006+0000 
7f8d2f15c700  1 mgr handle_mgr_map Activating!
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.006+0000 
7f8d2f15c700  1 mgr handle_mgr_map I am now activating
Sep 27 09:55:18 ceph-mon3 bash[476409]: [27/Sep/2024:13:55:18] ENGINE HTTP 
Server cherrypy._cpwsgi_server.CPWSGIServer(('::', 9283)) shut down
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.102+0000 
7f8c4baa7700  0 [cephadm DEBUG root] setting log level based on debug_mgr: INFO 
(2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.202+0000 
7f8c4baa7700  1 mgr load Constructed class from module: cephadm
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.206+0000 
7f8c4baa7700  0 [crash DEBUG root] setting log level based on debug_mgr: INFO 
(2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.206+0000 
7f8c4baa7700  1 mgr load Constructed class from module: crash
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.222+0000 
7f8c4baa7700  0 [devicehealth DEBUG root] setting log level based on debug_mgr: 
INFO (2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.222+0000 
7f8c4baa7700  1 mgr load Constructed class from module: devicehealth
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.222+0000 
7f8c3e28c700  0 [devicehealth INFO root] Starting
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.238+0000 
7f8c4baa7700  0 [orchestrator DEBUG root] setting log level based on debug_mgr: 
INFO (2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.242+0000 
7f8c4baa7700  1 mgr load Constructed class from module: orchestrator
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.318+0000 
7f8c4baa7700  0 [rbd_support DEBUG root] setting log level based on debug_mgr: 
INFO (2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: [27/Sep/2024:13:55:18] ENGINE Bus 
STARTING
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.346+0000 
7f8c31272700  0 [rbd_support INFO root] recovery thread starting
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.346+0000 
7f8c31272700  0 [rbd_support INFO root] starting setup
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.354+0000 
7f8c4baa7700  1 mgr load Constructed class from module: rbd_support
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.358+0000 
7f8c31272700  0 [rbd_support INFO root] MirrorSnapshotScheduleHandler: 
load_schedules
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.370+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: rbd, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.374+0000 
7f8c4baa7700  0 [status DEBUG root] setting log level based on debug_mgr: INFO 
(2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.374+0000 
7f8c4baa7700  1 mgr load Constructed class from module: status
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.378+0000 
7f8c4baa7700  0 [telemetry DEBUG root] setting log level based on debug_mgr: 
INFO (2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.378+0000 
7f8c4baa7700  1 mgr load Constructed class from module: telemetry
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.382+0000 
7f8c4baa7700  0 [volumes DEBUG root] setting log level based on debug_mgr: INFO 
(2/5)
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.386+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: images, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.390+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: volumes, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.394+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: vms, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.398+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: backups, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.402+0000 
7f8c21252700  0 [rbd_support INFO root] MirrorSnapshotScheduleHandler: starting
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.402+0000 
7f8c1fa4f700  0 [rbd_support INFO root] PerfHandler: starting
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.422+0000 
7f8c31272700  0 [rbd_support INFO root] load_task_task: rbd, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.430+0000 
7f8c4baa7700  1 mgr load Constructed class from module: volumes
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.434+0000 
7f8c31272700  0 [rbd_support INFO root] load_task_task: images, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.446+0000 
7f8c31272700  0 [rbd_support INFO root] load_task_task: volumes, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.458+0000 
7f8c31272700  0 [rbd_support INFO root] load_task_task: vms, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.870+0000 
7f8c3e28c700  0 [devicehealth INFO root] Check health
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.874+0000 
7f8c31272700  0 [rbd_support INFO root] load_task_task: backups, start_after=
Sep 27 09:55:18 ceph-mon3 bash[476409]: [27/Sep/2024:13:55:18] ENGINE Serving 
on http://:::9283
Sep 27 09:55:18 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:18.914+0000 
7f8c179bf700  0 [rbd_support INFO root] TaskHandler: starting
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.106+0000 
7f8c31272700  0 [rbd_support INFO root] TrashPurgeScheduleHandler: 
load_schedules
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.122+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: rbd, start_after=
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.126+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: images, start_after=
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.134+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: volumes, start_after=
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.138+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: vms, start_after=
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.146+0000 
7f8c31272700  0 [rbd_support INFO root] load_schedules: backups, start_after=
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.150+0000 
7f8c171be700  0 [rbd_support INFO root] TrashPurgeScheduleHandler: starting
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.150+0000 
7f8c31272700  0 [rbd_support INFO root] setup complete
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.638+0000 
7f8c43a97700  0 [cephadm INFO cherrypy.error] [27/Sep/2024:13:55:19] ENGINE Bus 
STARTING
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.638+0000 
7f8c43a97700  0 log_channel(cephadm) log [INF] : [27/Sep/2024:13:55:19] ENGINE 
Bus STARTING
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.770+0000 
7f8c43a97700  0 [cephadm INFO cherrypy.error] [27/Sep/2024:13:55:19] ENGINE 
Serving on https://10.5.74.23:7150
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.770+0000 
7f8c43a97700  0 log_channel(cephadm) log [INF] : [27/Sep/2024:13:55:19] ENGINE 
Serving on https://10.5.74.23:7150
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.770+0000 
7f8c43a97700  0 [cephadm INFO cherrypy.error] [27/Sep/2024:13:55:19] ENGINE Bus 
STARTED
Sep 27 09:55:19 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:19.770+0000 
7f8c43a97700  0 log_channel(cephadm) log [INF] : [27/Sep/2024:13:55:19] ENGINE 
Bus STARTED
Sep 27 09:55:28 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:28.030+0000 
7f8c4b2a6700  2 mgr.server handle_open ignoring open from mgr.ceph-mon1 
10.5.74.21:0/3328700921; not ready for session (expect reconnect)
Sep 27 09:55:29 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:29.030+0000 
7f8c4b2a6700  2 mgr.server handle_open ignoring open from mgr.ceph-mon1 
10.5.74.21:0/3328700921; not ready for session (expect reconnect)
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.410+0000 
7f8c42294700  0 [cephadm INFO root] Found osd claims -> {'ceph-osd31': ['88']}
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.410+0000 
7f8c42294700  0 log_channel(cephadm) log [INF] : Found osd claims -> 
{'ceph-osd31': ['88']}
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.410+0000 
7f8c42294700  0 [cephadm INFO cephadm.services.osd] Found osd claims for 
drivegroup ceph-osd31 -> {'ceph-osd31': ['88']}
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.414+0000 
7f8c42294700  0 log_channel(cephadm) log [INF] : Found osd claims for 
drivegroup ceph-osd31 -> {'ceph-osd31': ['88']}
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.426+0000 
7f8c42294700  0 [cephadm INFO root] Found osd claims -> {'ceph-osd31': ['88']}
Sep 27 09:55:36 ceph-mon3 bash[476409]: debug 2024-09-27T13:55:36.426+0000 
7f8c42294700  0 log_channel(cephadm) log [INF] : Found osd claims -> 
{'ceph-osd31': ['88']}


`ceph orch osd rm status` reports "No OSD remove/replace operations reported”

On the osd node:

`ceph-volume inventory --list-all --with-lsm /dev/sdg` reports:

====== Device report /dev/sdg ======

     path                      /dev/sdg
     ceph device               False
     lsm data                  {}
     available                 False
     rejected reasons          Insufficient space (<5GB)
     device id                 INTEL_SSDSC2KG038T8_PHYG039600UB3P8EGN

vs. a drive which has been successfully converted:

====== Device report /dev/sdf ======

     path                      /dev/sdf
     ceph device               True
     lsm data                  {}
     available                 False
     rejected reasons          Has a FileSystem, LVM detected, Insufficient 
space (<10 extents) on vgs, Insufficient space (<5GB)
     device id                 INTEL_SSDSC2KG038T8_PHYG039600CM3P8EGN
    --- Logical Volume ---
     name                      osd-block-88426db7-2322-4807-ac2e-b49929e170d6
     osd id                    87
     cluster name              ceph
     type                      block
     osd fsid                  88426db7-2322-4807-ac2e-b49929e170d6
     cluster fsid              9b3b3539-59a9-4338-8bab-3badfab6e855
     osdspec affinity          ceph-osd31
     block uuid                LNG2gB-pa0w-gl2v-hVQ3-6qTd-aXsR-Lenri3

We’ve zapped /dev/sdg a few times, initially when we ran the command to fail it 
out (`ceph orch osd rm 88 --replace —zap`), but also from the osd node itself  
with `ceph-volume lvm zap /dev/sdg —destroy`. We’ve also zapped the drive 
manually with:


sgdisk --zap-all /dev/sdg
wipefs --all --force /dev/sdg
dd if=/dev/zero bs=1M count=100 oflag=direct of=/dev/sdg
dd bs=512 if=/dev/zero of=/dev/sdg oflag=direct count=204800 seek=$(($(blockdev 
--getsz /dev/sdg) - 204800))
partprobe /dev/sdg

… based on the suggestion here: https://github.com/rook/rook/issues/11474

We’ve rebooted the osd node, with and without the drive inserted.

We’re unable to zap the drive from the orchestrator, as expected:

ceph orch device zap ceph-osd31 /dev/sdg --force
Error EINVAL: Device path '/dev/sdg' not found on host 'ceph-osd31’

Cheers,
/rjg

On Sep 27, 2024, at 2:07 AM, Eugen Block <ebl...@nde.ag> wrote:

EXTERNAL EMAIL | USE CAUTION

Right, if you need encryption, a rebuild is required. Your procedure
has already worked 4 times, so I'd say nothing seems wrong with that
per se.
Regarding the stuck device list, do you see the mgr logging anything
suspicious? Especially when you say that it only returns output after
a failover. Those two osd specs are not conflicting since the first is
"unmanaged" after adoption.
Is there something in 'ceph orch osd rm status'? Can you run 'cephadm
ceph-volume inventory' locally on that node? Do you see any hints in
the node's syslog? Maybe try a reboot or something?


Zitat von Bob Gibson <r...@oicr.on.ca>:

Thanks for your reply Eugen. I’m fairly new to cephadm so I wasn’t
aware that we could manage the drives without rebuilding them.
However, we thought we’d take advantage of this opportunity to also
encrypt the drives, and that does require a rebuild.

I have a theory on why the orchestrator is confused. I want to
create an osd service for each osd node so I can manage drives on a
per-node basis.

I started by creating a spec for the first node:

service_type: osd
service_id: ceph-osd31
placement:
 hosts:
 - ceph-osd31
spec:
 data_devices:
   rotational: 0
   size: '3TB:'
 encrypted: true
 filter_logic: AND
 objectstore: bluestore

But I also see a default spec, “osd”, which has placement set to “unmanaged”.

`ceph orch ls osd —export` shows the following:

service_type: osd
service_name: osd
unmanaged: true
spec:
 filter_logic: AND
 objectstore: bluestore
---
service_type: osd
service_id: ceph-osd31
service_name: osd.ceph-osd31
placement:
 hosts:
 - ceph-osd31
spec:
 data_devices:
   rotational: 0
   size: '3TB:'
 encrypted: true
 filter_logic: AND
 objectstore: bluestore

`ceph orch ls osd` shows that I was able to convert 4 drives using my spec:

NAME            PORTS  RUNNING  REFRESHED  AGE  PLACEMENT
osd                         95  10m ago    -    <unmanaged>
osd.ceph-osd31               4  10m ago    43m  ceph-osd31

Despite being able to convert 4 drives, I’m wondering if these specs
are conflicting with one another, and that has confused the
orchestrator. If so, how do I safely get from where I am now to
where I want to be? :-)

Cheers,
/rjg

On Sep 26, 2024, at 3:31 PM, Eugen Block <ebl...@nde.ag> wrote:

EXTERNAL EMAIL | USE CAUTION

Hi,

this seems a bit unnecessary to rebuild OSDs just to get them managed.
If you apply a spec file that targets your hosts/OSDs, they will
appear as managed. So when you would need to replace a drive, you
could already utilize the orchestrator to remove and zap the drive.
That works just fine.
How to get out of your current situation is not entirely clear to me
yet. I’ll reread your post tomorrow.

Regards,
Eugen

Zitat von Bob Gibson <r...@oicr.on.ca>:

Hi,

We recently converted a legacy cluster running Quincy v17.2.7 to
cephadm. The conversion went smoothly and left all osds unmanaged by
the orchestrator as expected. We’re now in the process of converting
the osds to be managed by the orchestrator. We successfully
converted a few of them, but then the orchestrator somehow got
confused. `ceph health detail` reports a “stray daemon” for the osd
we’re trying to convert, and the orchestrator is unable to refresh
its device list so it doesn’t see any available devices.

From the perspective of the osd node, the osd has been wiped and is
ready to be reinstalled. We’ve also rebooted the node for good
measure. `ceph osd tree` shows that the osd has been destroyed, but
the orchestrator won’t reinstall it because it thinks the device is
still active. The orchestrator device information is stale, but
we’re unable to refresh it. The usual recommended workaround of
failing over the mgr hasn’t helped. We’ve also tried `ceph orch
device ls —refresh` to no avail. In fact after running that command
subsequent runs of `ceph orch device ls` produce no output until the
mgr is failed over again.

Is there a way to force the orchestrator to refresh its list of
devices when in this state? If not, can anyone offer any suggestions
on how to fix this problem?

Cheers,
/rjg

P.S. Some additional information in case it’s helpful...

We’re using the following command to replace existing devices so
that they’re managed by the orchestrator:

```
ceph orch osd rm <osd> --replace —zap
```

and we’re currently stuck on osd 88.

```
ceph health detail
HEALTH_WARN 1 stray daemon(s) not managed by cephadm
[WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm
  stray daemon osd.88 on host ceph-osd31 not managed by cephadm
```

`ceph osd tree` shows that the osd has been destroyed and is ready
to be replaced:

```
ceph osd tree-from ceph-osd31
ID   CLASS  WEIGHT    TYPE NAME        STATUS     REWEIGHT  PRI-AFF
-46         34.93088  host ceph-osd31
84    ssd   3.49309      osd.84              up   1.00000  1.00000
85    ssd   3.49309      osd.85              up   1.00000  1.00000
86    ssd   3.49309      osd.86              up   1.00000  1.00000
87    ssd   3.49309      osd.87              up   1.00000  1.00000
88    ssd   3.49309      osd.88       destroyed         0  1.00000
89    ssd   3.49309      osd.89              up   1.00000  1.00000
90    ssd   3.49309      osd.90              up   1.00000  1.00000
91    ssd   3.49309      osd.91              up   1.00000  1.00000
92    ssd   3.49309      osd.92              up   1.00000  1.00000
93    ssd   3.49309      osd.93              up   1.00000  1.00000
```

The cephadm log shows a claim on node `ceph-osd31` for that osd:

```
2024-09-25T14:15:45.699348-0400 mgr.ceph-mon3.qzjgws [INF] Found osd
claims -> {'ceph-osd31': ['88']}
2024-09-25T14:15:45.699534-0400 mgr.ceph-mon3.qzjgws [INF] Found osd
claims for drivegroup ceph-osd31 -> {'ceph-osd31': ['88']}
```

`ceph orch device ls` shows that the device list isn’t refreshing:

```
ceph orch device ls ceph-osd31
HOST        PATH      TYPE  DEVICE ID
SIZE  AVAILABLE  REFRESHED  REJECT REASONS
ceph-osd31  /dev/sdc  ssd   INTEL_SSDSC2KG038T8_PHYG039603PE3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdd  ssd   INTEL_SSDSC2KG038T8_PHYG039600AY3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sde  ssd   INTEL_SSDSC2KG038T8_PHYG039600CW3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdf  ssd   INTEL_SSDSC2KG038T8_PHYG039600CM3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdg  ssd   INTEL_SSDSC2KG038T8_PHYG039600UB3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdh  ssd   INTEL_SSDSC2KG038T8_PHYG039603753P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdi  ssd   INTEL_SSDSC2KG038T8_PHYG039603R63P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdj  ssd   INTEL_SSDSC2KG038TZ_PHYJ4011032M3P8DGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdk  ssd   INTEL_SSDSC2KG038TZ_PHYJ3234010J3P8DGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
ceph-osd31  /dev/sdl  ssd   INTEL_SSDSC2KG038T8_PHYG039603NS3P8EGN
3576G  No         22h ago    Insufficient space (<10 extents) on
vgs, LVM detected, locked
```

`ceph node ls` thinks the osd still exists

```
ceph node ls osd | jq -r '."ceph-osd31"'
[
84,
85,
86,
87,
88, <— this shouldn’t exist
89,
90,
91,
92,
93
]
```

Each osd node has 10x 3.8 TB ssd drives for osds. On `ceph-osd31`,
cephadm doesn’t see osd.88 as expected:

```
cephadm ls --no-detail
[
  {
      "style": "cephadm:v1",
      "name": "osd.93",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.93"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.85",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.85"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.90",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.90"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.92",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.92"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.89",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.89"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.87",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.87"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.86",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.86"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.84",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.84"
  },
  {
      "style": "cephadm:v1",
      "name": "osd.91",
      "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
      "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.91"
  }
]
```

`lsblk` shows that `/dev/sdg` has been wiped.

```
NAME
                               MAJ:MIN RM   SIZE RO TYPE  MOUNTPOINT
sda
                                 8:0    0 223.6G  0 disk
|-sda1
                                 8:1    0    94M  0 part
`-sda2
                                 8:2    0 223.5G  0 part
`-md0
                                 9:0    0 223.4G  0 raid1 /
sdb
                                 8:16   0 223.6G  0 disk
|-sdb1
                                 8:17   0    94M  0 part
`-sdb2
                                 8:18   0 223.5G  0 part
`-md0
                                 9:0    0 223.4G  0 raid1 /
sdc
                                 8:32   1   3.5T  0 disk
`-ceph--03782b4c--9faa--49f5--b554--98e7b8515834-osd--block--ba272724--daa6--45f5--9f69--789cc0bda077
 253:3    0   3.5T
lvm
`-keCkP2-o6h8-jKkw-RKiD-UBFf-A8EL-JDJGPR
                               253:9    0   3.5T  0 crypt
sdd
                                 8:48   1   3.5T  0 disk
`-ceph--c07907d8--4a75--4ba3--b5e1--2ebf49ecbdf6-osd--block--58d1d50d--6228--4e6f--9a52--2a305ba00700
 253:7    0   3.5T
lvm
`-WB8Mxn-qCHI-4T01-imiG-hNBR-by60-YuxgfD
                               253:11   0   3.5T  0 crypt
sde
                                 8:64   1   3.5T  0 disk
`-ceph--6f9d4df4--7ce6--44a4--a7b1--62c85af8cfe0-osd--block--aabcb30d--0084--490a--969b--78f7af6e94da
 253:8    0   3.5T
lvm
`-g9qErH-vTXY-JQbs-eh61-W0Mn-TAV8-gof4zy
                               253:12   0   3.5T  0 crypt
sdf
                                 8:80   1   3.5T  0 disk
`-ceph--d6b728f8--e365--46db--b30f--6c00805c752b-osd--block--88426db7--2322--4807--ac2e--b49929e170d6
 253:6    0   3.5T
lvm
`-LNG2gB-pa0w-gl2v-hVQ3-6qTd-aXsR-Lenri3
                               253:10   0   3.5T  0 crypt
sdg
                                 8:96   1   3.5T  0 disk
sdh
                                 8:112  1   3.5T  0 disk
`-ceph--de2cfee6--8e0a--4aa0--9e6b--90c09025768c-osd--block--a3b86251--2799--4243--a857--f218fa90f29a
 253:2    0   3.5T
lvm
sdi
                                 8:128  1   3.5T  0 disk
`-ceph--30dee450--0fdd--46ea--9eec--6a4c7706df9c-osd--block--bfc090db--dde4--47dd--a1c9--1cd838ea43b3
 253:4    0   3.5T
lvm
sdj
                                 8:144  1   3.5T  0 disk
`-ceph--78febcf5--43f4--4820--8dc7--0f6c22816c9f-osd--block--da1e69c7--6427--4562--8290--90bcb9526747
 253:0    0   3.5T
lvm
sdk
                                 8:160  1   3.5T  0 disk
`-ceph--fe210281--b1f5--4d5e--9ab0--2f226612af00-osd--block--6bb9f308--e853--4303--83ea--553c3a3513e1
 253:1    0   3.5T
lvm
sdl
                                 8:176  1   3.5T  0 disk
`-ceph--9f21c916--f211--4d1b--8214--6ad1cecac810-osd--block--572d850c--c201--4af4--ac42--0ed2a6ed73ed
 253:5    0   3.5T
lvm
```

_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io


_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io




_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to