Hi Vasu,

Thank you that is good to know!

I am running ceph version 10.2.3 and CentOS 7.2.1511 (Core) minimal.

Cheers,
Mike

On Tue, Nov 29, 2016 at 7:26 PM, Vasu Kulkarni <vakul...@redhat.com> wrote:

> you can ignore that, its a known issue http://tracker.ceph.com/
> issues/15990
>
> regardless waht version of ceph are you running and what are the details
> of os version you updated to ?
>
> On Tue, Nov 29, 2016 at 7:12 PM, Mike Jacobacci <mi...@flowjo.com> wrote:
>
>> Found some more info, but getting weird... All three OSD nodes shows the
>> same unknown cluster message on all the OSD disks.  I don't know where it
>> came from, all the nodes were configured using ceph-deploy on the admin
>> node.  In any case, the OSD's seem to be up and running, the health is ok.
>>
>> no ceph-disk@ services are running on any of the OSD nodes which I
>> didn't notice before and each node was setup the exact same, yet there are
>> different services listed under systemctl:
>>
>> OSD NODE 1:
>> Output in earlier email
>>
>> OSD NODE 2:
>>
>> ● ceph-disk@dev-sdb1.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb1
>>
>> ● ceph-disk@dev-sdb2.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb2
>>
>> ● ceph-disk@dev-sdb5.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb5
>>
>> ● ceph-disk@dev-sdc2.service
>>         loaded failed failed    Ceph disk activation: /dev/sdc2
>>
>> ● ceph-disk@dev-sdc4.service
>>         loaded failed failed    Ceph disk activation: /dev/sdc4
>>
>>
>> OSD NODE 3:
>>
>> ● ceph-disk@dev-sdb1.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb1
>>
>> ● ceph-disk@dev-sdb3.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb3
>>
>> ● ceph-disk@dev-sdb4.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb4
>>
>> ● ceph-disk@dev-sdb5.service
>>         loaded failed failed    Ceph disk activation: /dev/sdb5
>>
>> ● ceph-disk@dev-sdc2.service
>>         loaded failed failed    Ceph disk activation: /dev/sdc2
>>
>> ● ceph-disk@dev-sdc3.service
>>         loaded failed failed    Ceph disk activation: /dev/sdc3
>>
>> ● ceph-disk@dev-sdc4.service
>>         loaded failed failed    Ceph disk activation: /dev/sdc4
>>
>> From my understanding, the disks have already been activated... Should
>> these services even be running or enabled?
>>
>> Mike
>>
>>
>>
>> On Tue, Nov 29, 2016 at 6:33 PM, Mike Jacobacci <mi...@flowjo.com> wrote:
>>
>>> Sorry about that... Here is the output of ceph-disk list:
>>>
>>> ceph-disk list
>>> /dev/dm-0 other, xfs, mounted on /
>>> /dev/dm-1 swap, swap
>>> /dev/dm-2 other, xfs, mounted on /home
>>> /dev/sda :
>>>  /dev/sda2 other, LVM2_member
>>>  /dev/sda1 other, xfs, mounted on /boot
>>> /dev/sdb :
>>>  /dev/sdb1 ceph journal
>>>  /dev/sdb2 ceph journal
>>>  /dev/sdb3 ceph journal
>>>  /dev/sdb4 ceph journal
>>>  /dev/sdb5 ceph journal
>>> /dev/sdc :
>>>  /dev/sdc1 ceph journal
>>>  /dev/sdc2 ceph journal
>>>  /dev/sdc3 ceph journal
>>>  /dev/sdc4 ceph journal
>>>  /dev/sdc5 ceph journal
>>> /dev/sdd :
>>>  /dev/sdd1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.0
>>> /dev/sde :
>>>  /dev/sde1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.1
>>> /dev/sdf :
>>>  /dev/sdf1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.2
>>> /dev/sdg :
>>>  /dev/sdg1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.3
>>> /dev/sdh :
>>>  /dev/sdh1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.4
>>> /dev/sdi :
>>>  /dev/sdi1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.5
>>> /dev/sdj :
>>>  /dev/sdj1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.6
>>> /dev/sdk :
>>>  /dev/sdk1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.7
>>> /dev/sdl :
>>>  /dev/sdl1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.8
>>> /dev/sdm :
>>>  /dev/sdm1 ceph data, active, unknown cluster
>>> e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9, osd.9
>>>
>>>
>>>
>>> On Tue, Nov 29, 2016 at 6:32 PM, Mike Jacobacci <mi...@flowjo.com>
>>> wrote:
>>>
>>>> I forgot to add:
>>>>
>>>>
>>>> On Tue, Nov 29, 2016 at 6:28 PM, Mike Jacobacci <mi...@flowjo.com>
>>>> wrote:
>>>>
>>>>> So it looks like the journal partition is mounted:
>>>>>
>>>>> ls -lah /var/lib/ceph/osd/ceph-0/journal
>>>>> lrwxrwxrwx. 1 ceph ceph 9 Oct 10 16:11 /var/lib/ceph/osd/ceph-0/journal
>>>>> -> /dev/sdb1
>>>>>
>>>>> Here is the output of journalctl -xe when I try to start the
>>>>> ceph-diak@dev-sdb1 service:
>>>>>
>>>>> sh[17481]: mount_activate: Failed to activate
>>>>> sh[17481]: unmount: Unmounting /var/lib/ceph/tmp/mnt.m9ek7W
>>>>> sh[17481]: command_check_call: Running command: /bin/umount --
>>>>> /var/lib/ceph/tmp/mnt.m9ek7W
>>>>> sh[17481]: Traceback (most recent call last):
>>>>> sh[17481]: File "/usr/sbin/ceph-disk", line 9, in <module>
>>>>> sh[17481]: load_entry_point('ceph-disk==1.0.0', 'console_scripts',
>>>>> 'ceph-disk')()
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 5011, in run
>>>>> sh[17481]: main(sys.argv[1:])
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 4962, in main
>>>>> sh[17481]: args.func(args)
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 4720, in <lambda>
>>>>> sh[17481]: func=lambda args: main_activate_space(name, args),
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 3739, in main_activate_space
>>>>> sh[17481]: reactivate=args.reactivate,
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 3073, in mount_activate
>>>>> sh[17481]: (osd_id, cluster) = activate(path, activate_key_template,
>>>>> init)
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 3220, in activate
>>>>> sh[17481]: ' with fsid %s' % ceph_fsid)
>>>>> sh[17481]: ceph_disk.main.Error: Error: No cluster conf found in
>>>>> /etc/ceph with fsid e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9
>>>>> sh[17481]: Traceback (most recent call last):
>>>>> sh[17481]: File "/usr/sbin/ceph-disk", line 9, in <module>
>>>>> sh[17481]: load_entry_point('ceph-disk==1.0.0', 'console_scripts',
>>>>> 'ceph-disk')()
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 5011, in run
>>>>> sh[17481]: main(sys.argv[1:])
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 4962, in main
>>>>> sh[17481]: args.func(args)
>>>>> sh[17481]: File "/usr/lib/python2.7/site-packages/ceph_disk/main.py",
>>>>> line 4399, in main_trigger
>>>>> sh[17481]: raise Error('return code ' + str(ret))
>>>>> sh[17481]: ceph_disk.main.Error: Error: return code 1
>>>>> systemd[1]: ceph-disk@dev-sdb1.service: main process exited,
>>>>> code=exited, status=1/FAILURE
>>>>> systemd[1]: Failed to start Ceph disk activation: /dev/sdb1.
>>>>>
>>>>> I dont understand this error:
>>>>> ceph_disk.main.Error: Error: No cluster conf found in /etc/ceph with
>>>>> fsid e1d7b4ae-2dcd-40ee-bea5-d103fe1fa9c9
>>>>>
>>>>> My fsid in ceph.conf is:
>>>>> fsid = 75d6dba9-2144-47b1-87ef-1fe21d3c58a8
>>>>>
>>>>> I don't know why the fsid would change or be different. I thought I
>>>>> had a basic cluster setup, I don't understand what's going wrong.
>>>>>
>>>>> Mike
>>>>>
>>>>> On Tue, Nov 29, 2016 at 5:15 PM, Mike Jacobacci <mi...@flowjo.com>
>>>>> wrote:
>>>>>
>>>>>> Hi John,
>>>>>>
>>>>>> Thanks I wasn't sure if something happened to the journal partitions
>>>>>> or not.
>>>>>>
>>>>>> Right now, the ceph-osd.0-9 services are back up and the cluster
>>>>>> health is good, but none of the ceph-disk@dev-sd* services are
>>>>>> running.   How can I get the Journal partitions mounted again?
>>>>>>
>>>>>> Cheers,
>>>>>> Mike
>>>>>>
>>>>>> On Tue, Nov 29, 2016 at 4:30 PM, John Petrini <jpetr...@coredial.com>
>>>>>> wrote:
>>>>>>
>>>>>>> Also, don't run sgdisk again; that's just for creating the journal
>>>>>>> partitions. ceph-disk is a service used for prepping disks, only the OSD
>>>>>>> services need to be running as far as I know. Are the ceph-osd@x.
>>>>>>> services running now that you've mounted the disks?
>>>>>>>
>>>>>>> ___
>>>>>>>
>>>>>>> John Petrini
>>>>>>>
>>>>>>> NOC Systems Administrator   //   *CoreDial, LLC*   //   coredial.com
>>>>>>>    //   [image: Twitter] <https://twitter.com/coredial>   [image:
>>>>>>> LinkedIn] <http://www.linkedin.com/company/99631>   [image: Google
>>>>>>> Plus] <https://plus.google.com/104062177220750809525/posts>   [image:
>>>>>>> Blog] <http://success.coredial.com/blog>
>>>>>>> Hillcrest I, 751 Arbor Way, Suite 150, Blue Bell PA, 19422
>>>>>>> *P: *215.297.4400 x232   //   *F: *215.297.4401   //   *E: *
>>>>>>> jpetr...@coredial.com
>>>>>>>
>>>>>>> [image: Exceptional people. Proven Processes. Innovative Technology.
>>>>>>> Discover CoreDial - watch our video]
>>>>>>> <http://cta-redirect.hubspot.com/cta/redirect/210539/4c492538-6e4b-445e-9480-bef676787085>
>>>>>>>
>>>>>>> The information transmitted is intended only for the person or
>>>>>>> entity to which it is addressed and may contain confidential and/or
>>>>>>> privileged material. Any review, retransmission,  dissemination or other
>>>>>>> use of, or taking of any action in reliance upon, this information by
>>>>>>> persons or entities other than the intended recipient is prohibited. If 
>>>>>>> you
>>>>>>> received this in error, please contact the sender and delete the 
>>>>>>> material
>>>>>>> from any computer.
>>>>>>>
>>>>>>> On Tue, Nov 29, 2016 at 7:27 PM, John Petrini <jpetr...@coredial.com
>>>>>>> > wrote:
>>>>>>>
>>>>>>>> What command are you using to start your OSD's?
>>>>>>>>
>>>>>>>> ___
>>>>>>>>
>>>>>>>> John Petrini
>>>>>>>>
>>>>>>>> NOC Systems Administrator   //   *CoreDial, LLC*   //
>>>>>>>> coredial.com   //   [image: Twitter] <https://twitter.com/coredial>
>>>>>>>>    [image: LinkedIn] <http://www.linkedin.com/company/99631>   [image:
>>>>>>>> Google Plus] <https://plus.google.com/104062177220750809525/posts>
>>>>>>>>    [image: Blog] <http://success.coredial.com/blog>
>>>>>>>> Hillcrest I, 751 Arbor Way, Suite 150, Blue Bell PA, 19422
>>>>>>>> *P: *215.297.4400 x232   //   *F: *215.297.4401   //   *E: *
>>>>>>>> jpetr...@coredial.com
>>>>>>>>
>>>>>>>> [image: Exceptional people. Proven Processes. Innovative
>>>>>>>> Technology. Discover CoreDial - watch our video]
>>>>>>>> <http://cta-redirect.hubspot.com/cta/redirect/210539/4c492538-6e4b-445e-9480-bef676787085>
>>>>>>>>
>>>>>>>> The information transmitted is intended only for the person or
>>>>>>>> entity to which it is addressed and may contain confidential and/or
>>>>>>>> privileged material. Any review, retransmission,  dissemination or 
>>>>>>>> other
>>>>>>>> use of, or taking of any action in reliance upon, this information by
>>>>>>>> persons or entities other than the intended recipient is prohibited. 
>>>>>>>> If you
>>>>>>>> received this in error, please contact the sender and delete the 
>>>>>>>> material
>>>>>>>> from any computer.
>>>>>>>>
>>>>>>>> On Tue, Nov 29, 2016 at 7:19 PM, Mike Jacobacci <mi...@flowjo.com>
>>>>>>>> wrote:
>>>>>>>>
>>>>>>>>> I was able to bring the osd's up by looking at my other OSD node
>>>>>>>>> which is the exact same hardware/disks and finding out which disks 
>>>>>>>>> map.
>>>>>>>>> But I still cant bring up any of the start ceph-disk@dev-sd*
>>>>>>>>> services... When I first installed the cluster and got the OSD's up, 
>>>>>>>>> I had
>>>>>>>>> to run the following:
>>>>>>>>>
>>>>>>>>> # sgdisk -t 1:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb
>>>>>>>>>
>>>>>>>>> # sgdisk -t 2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb
>>>>>>>>>
>>>>>>>>> # sgdisk -t 3:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb
>>>>>>>>>
>>>>>>>>> # sgdisk -t 4:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb
>>>>>>>>>
>>>>>>>>> # sgdisk -t 5:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdb
>>>>>>>>>
>>>>>>>>> # sgdisk -t 1:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc
>>>>>>>>>
>>>>>>>>> # sgdisk -t 2:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc
>>>>>>>>>
>>>>>>>>> # sgdisk -t 3:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc
>>>>>>>>>
>>>>>>>>> # sgdisk -t 4:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc
>>>>>>>>>
>>>>>>>>> # sgdisk -t 5:45b0969e-9b03-4f30-b4c6-b4b80ceff106 /dev/sdc
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Do i need to run that again?
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Cheers,
>>>>>>>>>
>>>>>>>>> Mike
>>>>>>>>>
>>>>>>>>> On Tue, Nov 29, 2016 at 4:13 PM, Sean Redmond <
>>>>>>>>> sean.redmo...@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Normally they mount based upon the gpt label, if it's not working
>>>>>>>>>> you can mount the disk under /mnt and then cat the file called 
>>>>>>>>>> whoami to
>>>>>>>>>> find out the osd number
>>>>>>>>>>
>>>>>>>>>> On 29 Nov 2016 23:56, "Mike Jacobacci" <mi...@flowjo.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> OK I am in some trouble now and would love some help!  After
>>>>>>>>>>> updating none of the OSDs on the node will come back up:
>>>>>>>>>>>
>>>>>>>>>>> ● ceph-disk@dev-sdb1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdb1
>>>>>>>>>>> ● ceph-disk@dev-sdb2.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdb2
>>>>>>>>>>> ● ceph-disk@dev-sdb3.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdb3
>>>>>>>>>>> ● ceph-disk@dev-sdb4.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdb4
>>>>>>>>>>> ● ceph-disk@dev-sdb5.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdb5
>>>>>>>>>>> ● ceph-disk@dev-sdc1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdc1
>>>>>>>>>>> ● ceph-disk@dev-sdc2.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdc2
>>>>>>>>>>> ● ceph-disk@dev-sdc3.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdc3
>>>>>>>>>>> ● ceph-disk@dev-sdc4.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdc4
>>>>>>>>>>> ● ceph-disk@dev-sdc5.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdc5
>>>>>>>>>>> ● ceph-disk@dev-sdd1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdd1
>>>>>>>>>>> ● ceph-disk@dev-sde1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sde1
>>>>>>>>>>> ● ceph-disk@dev-sdf1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdf1
>>>>>>>>>>> ● ceph-disk@dev-sdg1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdg1
>>>>>>>>>>> ● ceph-disk@dev-sdh1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdh1
>>>>>>>>>>> ● ceph-disk@dev-sdi1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdi1
>>>>>>>>>>> ● ceph-disk@dev-sdj1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdj1
>>>>>>>>>>> ● ceph-disk@dev-sdk1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdk1
>>>>>>>>>>> ● ceph-disk@dev-sdl1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdl1
>>>>>>>>>>> ● ceph-disk@dev-sdm1.service
>>>>>>>>>>>            loaded failed failed    Ceph disk activation: /dev/sdm1
>>>>>>>>>>> ● ceph-osd@0.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@1.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@2.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@3.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@4.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@5.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@6.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@7.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@8.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>> ● ceph-osd@9.service
>>>>>>>>>>>            loaded failed failed    Ceph object storage daemon
>>>>>>>>>>>
>>>>>>>>>>> I did some searching and saw that the issue is that the disks
>>>>>>>>>>> aren't mounting... My question is how can I mount them correctly 
>>>>>>>>>>> again
>>>>>>>>>>> (note sdb and sdc are ssd for cache)? I am not sure which disk maps 
>>>>>>>>>>> to
>>>>>>>>>>> ceph-osd@0 and so on.  Also, can I add them to /etc/fstab to
>>>>>>>>>>> work around?
>>>>>>>>>>>
>>>>>>>>>>> Cheers,
>>>>>>>>>>> Mike
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Nov 29, 2016 at 10:41 AM, Mike Jacobacci <
>>>>>>>>>>> mi...@flowjo.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hello,
>>>>>>>>>>>>
>>>>>>>>>>>> I would like to install OS updates on the ceph cluster and
>>>>>>>>>>>> activate a second 10gb port on the OSD nodes, so I wanted to 
>>>>>>>>>>>> verify the
>>>>>>>>>>>> correct steps to perform maintenance on the cluster.  We are only 
>>>>>>>>>>>> using rbd
>>>>>>>>>>>> to back our xenserver vm's at this point, and our cluster consists 
>>>>>>>>>>>> of 3 OSD
>>>>>>>>>>>> nodes, 3 Mon nodes and 1 admin node...  So would this be the 
>>>>>>>>>>>> correct steps:
>>>>>>>>>>>>
>>>>>>>>>>>> 1. Shut down VM's?
>>>>>>>>>>>> 2. run "ceph osd set noout" on admin node
>>>>>>>>>>>> 3. install updates on each monitoring node and reboot one at a
>>>>>>>>>>>> time.
>>>>>>>>>>>> 4. install updates on OSD nodes and activate second 10gb port,
>>>>>>>>>>>> reboot one OSD node at a time
>>>>>>>>>>>> 5. once all nodes back up, run "ceph osd unset noout"
>>>>>>>>>>>> 6. bring VM's back online
>>>>>>>>>>>>
>>>>>>>>>>>> Does this sound correct?
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> Cheers,
>>>>>>>>>>>> Mike
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> _______________________________________________
>>>>>>>>>>> ceph-users mailing list
>>>>>>>>>>> ceph-users@lists.ceph.com
>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> _______________________________________________
>>>>>>>>> ceph-users mailing list
>>>>>>>>> ceph-users@lists.ceph.com
>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to