On Wed, Dec 7, 2016 at 9:11 PM, M Ranga Swami Reddy <swamire...@gmail.com>
wrote:

> That's right..
> But, my question was: when an OSD down, all data will be moved to other
> OSDs from downed OSD. - Is this correct?
>

No, only after it is marked out.

"If an OSD is down and the degraded condition persists, Ceph may mark the
down OSD as out of the cluster and remap the data from the down OSD to
another OSD. The time between being marked down and being marked out is
controlled by mon osd down out interval, which is set to 300 seconds by
default."


> Now, I change the crushmap as out an OSD, then again data will be moved
> across the cluster?
>


>
> Thanks
> Swami
>
> On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote:
>
>> Hi
>>
>> Because "down" and "out" are different to ceph cluster
>>
>> Crush map of ceph is depends on how many osds are in ths cluster.
>>
>> Crush map doesn't change when osds are down. However crush map would
>> chage when the osds are absolutelly out.
>> Data location also will change, there fore rebalancing starts.
>>
>> Thanks
>> John Haan
>>
>>
>>
>> 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이 작성:
>>
>>> Sure, will try with "*ceph osd crush reweight 0.0" *and update the
>>> status.
>>>
>>> Thanks
>>> Swami
>>>
>>> On Fri, Dec 2, 2016 at 8:15 PM, David Turner <
>>> david.tur...@storagecraft.com> wrote:
>>>
>>>> If you want to reweight only once when you have a failed disk that is
>>>> being balanced off of, set the crush weight for that osd to 0.0.  Then when
>>>> you fully remove the disk from the cluster it will not do any additional
>>>> backfilling.  Any change to the crush map will likely move data around,
>>>> even if you're removing an already "removed" osd.
>>>>
>>>> ------------------------------
>>>>
>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | 
>>>> StorageCraft
>>>> Technology Corporation <https://storagecraft.com>
>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943
>>>> <%28385%29%20224-2943>
>>>>
>>>> ------------------------------
>>>>
>>>> If you are not the intended recipient of this message or received it
>>>> erroneously, please notify the sender and delete it, together with any
>>>> attachments, and be advised that any dissemination or copying of this
>>>> message is prohibited.
>>>>
>>>> ------------------------------
>>>>
>>>> ------------------------------
>>>> *From:* M Ranga Swami Reddy [swamire...@gmail.com]
>>>> *Sent:* Thursday, December 01, 2016 11:45 PM
>>>> *To:* David Turner
>>>> *Cc:* ceph-users
>>>> *Subject:* Re: [ceph-users] node and its OSDs down...
>>>>
>>>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which
>>>> started the recovery.
>>>> My worries is - why Ceph is doing the recovery, if an OSD is already
>>>> down and no more in the cluster. That means, ceph already
>>>> maintained down OSDs objects copied to another OSDs.. here is the ceph osd
>>>> tree o/p:
>>>> ===
>>>>
>>>> 227     0.91                            osd.227 down    0
>>>>
>>>> ....
>>>>
>>>> 250     0.91                            osd.250 down    0
>>>>
>>>> ===
>>>>
>>>>
>>>> So to avoid the recovery/rebalance , can I set the weight of OSD (which
>>>> was in down state). But is this weight setting also lead to rebalance
>>>> activity.
>>>>
>>>>
>>>> Thanks
>>>>
>>>> Swami
>>>>
>>>>
>>>>
>>>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner <
>>>> david.tur...@storagecraft.com> wrote:
>>>>
>>>>> I assume you also did ceph osd crush remove osd.<id>.  When you
>>>>> removed the osd that was down/out and balanced off of, you changed the
>>>>> weight of the host that it was on which triggers additional backfilling to
>>>>> balance the crush map.
>>>>>
>>>>> ------------------------------
>>>>>
>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | 
>>>>> StorageCraft
>>>>> Technology Corporation <https://storagecraft.com>
>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943
>>>>> <%28385%29%20224-2943>
>>>>>
>>>>> ------------------------------
>>>>>
>>>>> If you are not the intended recipient of this message or received it
>>>>> erroneously, please notify the sender and delete it, together with any
>>>>> attachments, and be advised that any dissemination or copying of this
>>>>> message is prohibited.
>>>>>
>>>>> ------------------------------
>>>>>
>>>>> ------------------------------
>>>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of M
>>>>> Ranga Swami Reddy [swamire...@gmail.com]
>>>>> *Sent:* Thursday, December 01, 2016 3:03 AM
>>>>> *To:* ceph-users
>>>>> *Subject:* [ceph-users] node and its OSDs down...
>>>>>
>>>>> Hello,
>>>>> One of my ceph node with 20 OSDs down...After a couple of hours,
>>>>> ceph health is in OK state.
>>>>>
>>>>> Now, I tried to remove those OSDs, which were down state from
>>>>> ceph cluster...
>>>>> using the "ceh osd remove osd.<id>"
>>>>> then ceph clsuter started rebalancing...which is strange ..because
>>>>> thsoe OSDs are down for a long time and health also OK..
>>>>> my question - why recovery or reblance started when I remove the OSD
>>>>> (which was down).
>>>>>
>>>>> Thanks
>>>>> Swami
>>>>>
>>>>>
>>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>


-- 
Cheers,
Brad
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to