Hello everyone, 

I am running ceph version 15.2.8 on Ubuntu servers. I am using bluestore osds 
with data on hdd and db and wal on ssd drives. Each ssd has been partitioned 
such that it holds 5 dbs and 5 wals. The ssd were were prepared a while back 
probably when I was running ceph 13.x. I have been gradually adding new osd 
drives as needed. Recently, I've tried to add more osds, which have failed to 
my surprise. Previously I've had no issues adding the drives. However, it seems 
that I can no longer do that with version 15.2.x 

Here is what I get: 


root@arh-ibstorage4-ib  /home/andrei  ceph-volume lvm prepare --bluestore 
--data /dev/sds --block.db /dev/ssd3/db5 --block.wal /dev/ssd3/wal5 
Running command: /usr/bin/ceph-authtool --gen-print-key 
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd 
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 
6aeef34b-0724-4d20-a10b-197cab23e24d 
Running command: /usr/sbin/vgcreate --force --yes 
ceph-1c7cef26-327a-4785-96b3-dcb1b97e8e2f /dev/sds 
stderr: WARNING: PV /dev/sdp in VG ceph-bc7587b5-0112-4097-8c9f-4442e8ea5645 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdo in VG ceph-33eda27c-53ed-493e-87a8-39e1862da809 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdn in VG ssd2 is using an old PV header, modify the 
VG to update. 
stderr: WARNING: PV /dev/sdm in VG ssd1 is using an old PV header, modify the 
VG to update. 
stderr: WARNING: PV /dev/sdj in VG ceph-9d8da00c-f6b9-473f-b499-fa60d74b46c5 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdi in VG ceph-1603149e-1e50-4b86-a360-1372f4243603 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdh in VG ceph-a5f4416c-8e69-4a66-a884-1d1229785acb is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sde in VG ceph-aac71121-e308-4e25-ae95-ca51bca7aaff is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdd in VG ceph-1e216580-c01b-42c5-a10f-293674a55c4c is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdc in VG ceph-630f7716-3d05-41bb-92c9-25402e9bb264 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sdb in VG ceph-a549c28d-9b06-46d5-8ba3-3bd99ff54f57 is 
using an old PV header, modify the VG to update. 
stderr: WARNING: PV /dev/sda in VG ceph-70943bd0-de71-4651-a73d-c61bc624755f is 
using an old PV header, modify the VG to update. 
stdout: Physical volume "/dev/sds" successfully created. 
stdout: Volume group "ceph-1c7cef26-327a-4785-96b3-dcb1b97e8e2f" successfully 
created 
Running command: /usr/sbin/lvcreate --yes -l 3814911 -n 
osd-block-6aeef34b-0724-4d20-a10b-197cab23e24d 
ceph-1c7cef26-327a-4785-96b3-dcb1b97e8e2f 
stdout: Logical volume "osd-block-6aeef34b-0724-4d20-a10b-197cab23e24d" 
created. 
--> blkid could not detect a PARTUUID for device: /dev/ssd3/wal5 
--> Was unable to complete a new OSD, will rollback changes 
Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd 
--keyring /var/lib/ceph/bootstrap-osd/ceph.keyring osd purge-new osd.15 
--yes-i-really-mean-it 
stderr: 2021-04-28T20:05:52.290+0100 7f76bbfa9700 -1 auth: unable to find a 
keyring on 
/etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyring,/etc
 
/ceph/keyring.bin,: (2) No such file or directory 
2021-04-28T20:05:52.290+0100 7f76bbfa9700 -1 AuthRegistry(0x7f76b4058e60) no 
keyring found at 
/etc/ceph/ceph.client.bootstrap-osd.keyring,/etc/ceph/ceph.keyring,/etc/ceph/keyrin
 
g,/etc/ceph/keyring.bin,, disabling cephx 
stderr: purged osd.15 
--> RuntimeError: unable to use device 

I have tried to find a solution, but wasn't able to resolve the problem? I am 
sure that I've previously added new volumes using the above command. 

lvdisplay shows: 

--- Logical volume --- 
LV Path /dev/ssd3/wal5 
LV Name wal5 
VG Name ssd3 
LV UUID WPQJs9-olAj-ACbU-qnEM-6ytu-aLMv-hAABYy 
LV Write Access read/write 
LV Creation host, time arh-ibstorage4-ib, 2020-07-29 23:45:17 +0100 
LV Status available 
# open 0 
LV Size 1.00 GiB 
Current LE 256 
Segments 1 
Allocation inherit 
Read ahead sectors auto 
- currently set to 256 
Block device 253:6 


--- Logical volume --- 
LV Path /dev/ssd3/db5 
LV Name db5 
VG Name ssd3 
LV UUID FVT2Mm-a00P-eCoQ-FZAf-AulX-4q9r-PaDTC6 
LV Write Access read/write 
LV Creation host, time arh-ibstorage4-ib, 2020-07-29 23:46:01 +0100 
LV Status available 
# open 0 
LV Size 177.00 GiB 
Current LE 45312 
Segments 1 
Allocation inherit 
Read ahead sectors auto 
- currently set to 256 
Block device 253:11 



How do I resolve the errors and create the new osd? 

Cheers 

Andrei 




_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to