I was able to bring the osd's up by looking at my other OSD node which is
the exact same hardware/disks and finding out which disks map.  But I still
cant bring up any of the start ceph-disk@dev-sd* services... When I first
installed the cluster and got the OSD's up, I had to run the following:

# sgdisk -t 1:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb

# sgdisk -t 2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb

# sgdisk -t 3:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb

# sgdisk -t 4:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb

# sgdisk -t 5:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb

# sgdisk -t 1:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc

# sgdisk -t 2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc

# sgdisk -t 3:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc

# sgdisk -t 4:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc

# sgdisk -t 5:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc


Do i need to run that again?


Cheers,

Mike

On Tue, Nov 29, 2016 at 4:13 PM, Sean Redmond <sean.redmo...@gmail.com>
wrote:

> Normally they mount based upon the gpt label, if it's not working you can
> mount the disk under /mnt and then cat the file called whoami to find out
> the osd number
>
> On 29 Nov 2016 23:56, "Mike Jacobacci" <mi...@flowjo.com> wrote:
>
>> OK I am in some trouble now and would love some help!  After updating
>> none of the OSDs on the node will come back up:
>>
>> ● ceph-disk@dev-sdb1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdb1
>> ● ceph-disk@dev-sdb2.service
>>  loaded failed failed    Ceph disk activation: /dev/sdb2
>> ● ceph-disk@dev-sdb3.service
>>  loaded failed failed    Ceph disk activation: /dev/sdb3
>> ● ceph-disk@dev-sdb4.service
>>  loaded failed failed    Ceph disk activation: /dev/sdb4
>> ● ceph-disk@dev-sdb5.service
>>  loaded failed failed    Ceph disk activation: /dev/sdb5
>> ● ceph-disk@dev-sdc1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdc1
>> ● ceph-disk@dev-sdc2.service
>>  loaded failed failed    Ceph disk activation: /dev/sdc2
>> ● ceph-disk@dev-sdc3.service
>>  loaded failed failed    Ceph disk activation: /dev/sdc3
>> ● ceph-disk@dev-sdc4.service
>>  loaded failed failed    Ceph disk activation: /dev/sdc4
>> ● ceph-disk@dev-sdc5.service
>>  loaded failed failed    Ceph disk activation: /dev/sdc5
>> ● ceph-disk@dev-sdd1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdd1
>> ● ceph-disk@dev-sde1.service
>>  loaded failed failed    Ceph disk activation: /dev/sde1
>> ● ceph-disk@dev-sdf1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdf1
>> ● ceph-disk@dev-sdg1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdg1
>> ● ceph-disk@dev-sdh1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdh1
>> ● ceph-disk@dev-sdi1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdi1
>> ● ceph-disk@dev-sdj1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdj1
>> ● ceph-disk@dev-sdk1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdk1
>> ● ceph-disk@dev-sdl1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdl1
>> ● ceph-disk@dev-sdm1.service
>>  loaded failed failed    Ceph disk activation: /dev/sdm1
>> ● ceph-osd@0.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@1.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@2.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@3.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@4.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@5.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@6.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@7.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@8.service
>>  loaded failed failed    Ceph object storage daemon
>> ● ceph-osd@9.service
>>  loaded failed failed    Ceph object storage daemon
>>
>> I did some searching and saw that the issue is that the disks aren't
>> mounting... My question is how can I mount them correctly again (note sdb
>> and sdc are ssd for cache)? I am not sure which disk maps to ceph-osd@0
>> and so on.  Also, can I add them to /etc/fstab to work around?
>>
>> Cheers,
>> Mike
>>
>> On Tue, Nov 29, 2016 at 10:41 AM, Mike Jacobacci <mi...@flowjo.com>
>> wrote:
>>
>>> Hello,
>>>
>>> I would like to install OS updates on the ceph cluster and activate a
>>> second 10gb port on the OSD nodes, so I wanted to verify the correct steps
>>> to perform maintenance on the cluster.  We are only using rbd to back our
>>> xenserver vm's at this point, and our cluster consists of 3 OSD nodes, 3
>>> Mon nodes and 1 admin node...  So would this be the correct steps:
>>>
>>> 1. Shut down VM's?
>>> 2. run "ceph osd set noout" on admin node
>>> 3. install updates on each monitoring node and reboot one at a time.
>>> 4. install updates on OSD nodes and activate second 10gb port, reboot
>>> one OSD node at a time
>>> 5. once all nodes back up, run "ceph osd unset noout"
>>> 6. bring VM's back online
>>>
>>> Does this sound correct?
>>>
>>>
>>> Cheers,
>>> Mike
>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to