That's right..
But, my question was: when an OSD down, all data will be moved to other
OSDs from downed OSD. - Is this correct?
Now, I change the crushmap as out an OSD, then again data will be moved
across the cluster?

Thanks
Swami

On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote:

> Hi
>
> Because "down" and "out" are different to ceph cluster
>
> Crush map of ceph is depends on how many osds are in ths cluster.
>
> Crush map doesn't change when osds are down. However crush map would chage
> when the osds are absolutelly out.
> Data location also will change, there fore rebalancing starts.
>
> Thanks
> John Haan
>
>
>
> 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이 작성:
>
>> Sure, will try with "*ceph osd crush reweight 0.0" *and update the
>> status.
>>
>> Thanks
>> Swami
>>
>> On Fri, Dec 2, 2016 at 8:15 PM, David Turner <
>> david.tur...@storagecraft.com> wrote:
>>
>>> If you want to reweight only once when you have a failed disk that is
>>> being balanced off of, set the crush weight for that osd to 0.0.  Then when
>>> you fully remove the disk from the cluster it will not do any additional
>>> backfilling.  Any change to the crush map will likely move data around,
>>> even if you're removing an already "removed" osd.
>>>
>>> ------------------------------
>>>
>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | 
>>> StorageCraft
>>> Technology Corporation <https://storagecraft.com>
>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>> Office: 801.871.2760 <(801)%20871-2760> | Mobile: 385.224.2943
>>> <(385)%20224-2943>
>>>
>>> ------------------------------
>>>
>>> If you are not the intended recipient of this message or received it
>>> erroneously, please notify the sender and delete it, together with any
>>> attachments, and be advised that any dissemination or copying of this
>>> message is prohibited.
>>>
>>> ------------------------------
>>>
>>> ------------------------------
>>> *From:* M Ranga Swami Reddy [swamire...@gmail.com]
>>> *Sent:* Thursday, December 01, 2016 11:45 PM
>>> *To:* David Turner
>>> *Cc:* ceph-users
>>> *Subject:* Re: [ceph-users] node and its OSDs down...
>>>
>>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which
>>> started the recovery.
>>> My worries is - why Ceph is doing the recovery, if an OSD is already
>>> down and no more in the cluster. That means, ceph already
>>> maintained down OSDs objects copied to another OSDs.. here is the ceph osd
>>> tree o/p:
>>> ===
>>>
>>> 227     0.91                            osd.227 down    0
>>>
>>> ....
>>>
>>> 250     0.91                            osd.250 down    0
>>>
>>> ===
>>>
>>>
>>> So to avoid the recovery/rebalance , can I set the weight of OSD (which
>>> was in down state). But is this weight setting also lead to rebalance
>>> activity.
>>>
>>>
>>> Thanks
>>>
>>> Swami
>>>
>>>
>>>
>>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner <
>>> david.tur...@storagecraft.com> wrote:
>>>
>>>> I assume you also did ceph osd crush remove osd.<id>.  When you removed
>>>> the osd that was down/out and balanced off of, you changed the weight of
>>>> the host that it was on which triggers additional backfilling to balance
>>>> the crush map.
>>>>
>>>> ------------------------------
>>>>
>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | 
>>>> StorageCraft
>>>> Technology Corporation <https://storagecraft.com>
>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>>> Office: 801.871.2760 <(801)%20871-2760> | Mobile: 385.224.2943
>>>> <(385)%20224-2943>
>>>>
>>>> ------------------------------
>>>>
>>>> If you are not the intended recipient of this message or received it
>>>> erroneously, please notify the sender and delete it, together with any
>>>> attachments, and be advised that any dissemination or copying of this
>>>> message is prohibited.
>>>>
>>>> ------------------------------
>>>>
>>>> ------------------------------
>>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of M
>>>> Ranga Swami Reddy [swamire...@gmail.com]
>>>> *Sent:* Thursday, December 01, 2016 3:03 AM
>>>> *To:* ceph-users
>>>> *Subject:* [ceph-users] node and its OSDs down...
>>>>
>>>> Hello,
>>>> One of my ceph node with 20 OSDs down...After a couple of hours,
>>>> ceph health is in OK state.
>>>>
>>>> Now, I tried to remove those OSDs, which were down state from
>>>> ceph cluster...
>>>> using the "ceh osd remove osd.<id>"
>>>> then ceph clsuter started rebalancing...which is strange ..because
>>>> thsoe OSDs are down for a long time and health also OK..
>>>> my question - why recovery or reblance started when I remove the OSD
>>>> (which was down).
>>>>
>>>> Thanks
>>>> Swami
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to