So I disabled ceph-disk and will chalk it up as a red herring to ignore.

On Thu, Jul 20, 2017 at 11:02 AM Roger Brown <rogerpbr...@gmail.com> wrote:

> Also I'm just noticing osd1 is my only OSD host that even has an enabled
> target for ceph-disk (ceph-disk@dev-sdb2.service).
>
> roger@osd1:~$ systemctl list-units ceph*
>   UNIT                       LOAD   ACTIVE SUB     DESCRIPTION
> ● ceph-disk@dev-sdb2.service loaded failed failed  Ceph disk activation:
> /dev/sdb2
>   ceph-osd@3.service         loaded active running Ceph object storage
> daemon osd.3
>   ceph-mds.target            loaded active active  ceph target allowing to
> start/stop all ceph-mds@.service instances at once
>   ceph-mgr.target            loaded active active  ceph target allowing to
> start/stop all ceph-mgr@.service instances at once
>   ceph-mon.target            loaded active active  ceph target allowing to
> start/stop all ceph-mon@.service instances at once
>   ceph-osd.target            loaded active active  ceph target allowing to
> start/stop all ceph-osd@.service instances at once
>   ceph-radosgw.target        loaded active active  ceph target allowing to
> start/stop all ceph-radosgw@.service instances at once
>   ceph.target                loaded active active  ceph target allowing to
> start/stop all ceph*@.service instances at once
>
> roger@osd2:~$ systemctl list-units ceph*
> UNIT                LOAD   ACTIVE SUB     DESCRIPTION
> ceph-osd@4.service  loaded active running Ceph object storage daemon osd.4
> ceph-mds.target     loaded active active  ceph target allowing to
> start/stop all ceph-mds@.service instances at once
> ceph-mgr.target     loaded active active  ceph target allowing to
> start/stop all ceph-mgr@.service instances at once
> ceph-mon.target     loaded active active  ceph target allowing to
> start/stop all ceph-mon@.service instances at once
> ceph-osd.target     loaded active active  ceph target allowing to
> start/stop all ceph-osd@.service instances at once
> ceph-radosgw.target loaded active active  ceph target allowing to
> start/stop all ceph-radosgw@.service instances at once
> ceph.target         loaded active active  ceph target allowing to
> start/stop all ceph*@.service instances at once
>
> roger@osd3:~$ systemctl list-units ceph*
> UNIT                LOAD   ACTIVE SUB     DESCRIPTION
> ceph-osd@0.service  loaded active running Ceph object storage daemon osd.0
> ceph-mds.target     loaded active active  ceph target allowing to
> start/stop all ceph-mds@.service instances at once
> ceph-mgr.target     loaded active active  ceph target allowing to
> start/stop all ceph-mgr@.service instances at once
> ceph-mon.target     loaded active active  ceph target allowing to
> start/stop all ceph-mon@.service instances at once
> ceph-osd.target     loaded active active  ceph target allowing to
> start/stop all ceph-osd@.service instances at once
> ceph-radosgw.target loaded active active  ceph target allowing to
> start/stop all ceph-radosgw@.service instances at once
> ceph.target         loaded active active  ceph target allowing to
> start/stop all ceph*@.service instances at once
>
>
> On Thu, Jul 20, 2017 at 10:23 AM Roger Brown <rogerpbr...@gmail.com>
> wrote:
>
>> I think I need help with some OSD trouble. OSD daemons on two hosts
>> started flapping. At length, I rebooted host osd1 (osd.3), but the OSD
>> daemon still fails to start. Upon closer
>> inspection, ceph-disk@dev-sdb2.service is failing to start due to,
>> "Error: /dev/sdb2 is not a block device"
>>
>> This is the command I see it failing to run:
>>
>> roger@osd1:~$ sudo /usr/sbin/ceph-disk --verbose activate-block /dev/sdb2
>> Traceback (most recent call last):
>>   File "/usr/sbin/ceph-disk", line 9, in <module>
>>     load_entry_point('ceph-disk==1.0.0', 'console_scripts', 'ceph-disk')()
>>   File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5731,
>> in run
>>     main(sys.argv[1:])
>>   File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5682,
>> in main
>>     args.func(args)
>>   File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 5438,
>> in <lambda>
>>     func=lambda args: main_activate_space(name, args),
>>   File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4160,
>> in main_activate_space
>>     osd_uuid = get_space_osd_uuid(name, dev)
>>   File "/usr/lib/python2.7/dist-packages/ceph_disk/main.py", line 4115,
>> in get_space_osd_uuid
>>     raise Error('%s is not a block device' % path)
>> ceph_disk.main.Error: Error: /dev/sdb2 is not a block device
>>
>> osd1 environment:
>> $ ceph -v
>> ceph version 12.1.1 (f3e663a190bf2ed12c7e3cda288b9a159572c800) luminous
>> (rc)
>> $ uname -r
>> 4.4.0-83-generic
>> $ lsb_release -sc
>> xenial
>>
>> Please advise.
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to