So I disabled ceph-disk and will chalk it up as a red herring to ignore.
On Thu, Jul 20, 2017 at 11:02 AM Roger Brown <rogerpbr...@gmail.com> wrote: > Also I'm just noticing osd1 is my only OSD host that even has an enabled > target for ceph-disk (ceph-disk@dev-sdb2.service). > > roger@osd1:~$ systemctl list-units ceph* > UNIT LOAD ACTIVE SUB DESCRIPTION > ● ceph-disk@dev-sdb2.service loaded failed failed Ceph disk activation: > /dev/sdb2 > ceph-osd@3.service loaded active running Ceph object storage > daemon osd.3 > ceph-mds.target loaded active active ceph target allowing to > start/stop all ceph-mds@.service instances at once > ceph-mgr.target loaded active active ceph target allowing to > start/stop all ceph-mgr@.service instances at once > ceph-mon.target loaded active active ceph target allowing to > start/stop all ceph-mon@.service instances at once > ceph-osd.target loaded active active ceph target allowing to > start/stop all ceph-osd@.service instances at once > ceph-radosgw.target loaded active active ceph target allowing to > start/stop all ceph-radosgw@.service instances at once > ceph.target loaded active active ceph target allowing to > start/stop all ceph*@.service instances at once > > roger@osd2:~$ systemctl list-units ceph* > UNIT LOAD ACTIVE SUB DESCRIPTION > ceph-osd@4.service loaded active running Ceph object storage daemon osd.4 > ceph-mds.target loaded active active ceph target allowing to > start/stop all ceph-mds@.service instances at once > ceph-mgr.target loaded active active ceph target allowing to > start/stop all ceph-mgr@.service instances at once > ceph-mon.target loaded active active ceph target allowing to > start/stop all ceph-mon@.service instances at once > ceph-osd.target loaded active active ceph target allowing to > start/stop all ceph-osd@.service instances at once > ceph-radosgw.target loaded active active ceph target allowing to > start/stop all ceph-radosgw@.service instances at once > ceph.target loaded active active ceph target allowing to > start/stop all ceph*@.service instances at once > > roger@osd3:~$ systemctl list-units ceph* > UNIT LOAD ACTIVE SUB DESCRIPTION > ceph-osd@0.service loaded active running Ceph object storage daemon osd.0 > ceph-mds.target loaded active active ceph target allowing to > start/stop all ceph-mds@.service instances at once > ceph-mgr.target loaded active active ceph target allowing to > start/stop all ceph-mgr@.service instances at once > ceph-mon.target loaded active active ceph target allowing to > start/stop all ceph-mon@.service instances at once > ceph-osd.target loaded active active ceph target allowing to > start/stop all ceph-osd@.service instances at once > ceph-radosgw.target loaded active active ceph target allowing to > start/stop all ceph-radosgw@.service instances at once > ceph.target loaded active active ceph target allowing to > start/stop all ceph*@.service instances at once > > > On Thu, Jul 20, 2017 at 10:23 AM Roger Brown <rogerpbr...@gmail.com> > wrote: > >> I think I need help with some OSD trouble. OSD daemons on two hosts >> started flapping. At length, I rebooted host osd1 (osd.3), but the OSD >> daemon still fails to start. Upon closer >> inspection, ceph-disk@dev-sdb2.service is failing to start due to, >> "Error: /dev/sdb2 is not a block device" >> >> This is the command I see it failing to run: >> >> roger@osd1:~$ sudo /usr/sbin/ceph-disk --verbose activate-block /dev/sdb2 >> Traceback (most recent call last): >> File "/usr/sbin/ceph-disk", line 9, in <module> >> load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')() >> File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5731, >> in run >> main(sys.argv[1:]) >> File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5682, >> in main >> args.func(args) >> File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5438, >> in <lambda> >> func=lambda args: main_activate_space(name, args), >> File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4160, >> in main_activate_space >> osd_uuid = get_space_osd_uuid(name, dev) >> File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4115, >> in get_space_osd_uuid >> raise Error('%s is not a block device' % path) >> ceph_disk.main.Error: Error: /dev/sdb2 is not a block device >> >> osd1 environment: >> $ ceph -v >> ceph version 12.1.1 (f3e663a190bf2ed12c7e3cda288b9a159572c800) luminous >> (rc) >> $ uname -r >> 4.4.0-83-generic >> $ lsb_release -sc >> xenial >> >> Please advise. >> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com