On Fri, Dec 9, 2016 at 3:28 PM, M Ranga Swami Reddy <swamire...@gmail.com>
wrote:

> Confused ...
> a few OSDs down and cluster done the recovery and reblanced to HEALTH OK
> state.
> Now I can could that down OSDs are down state from crushmap and are not
> part of OSD up or in state.
> After 5 days or says, still the same state.
> How or when Ceph will make the down state OSDs to out state? I guess
> ceph don't do it.
>

 300 seconds by default

Any flags set? noout perhaps?

Now I ran the OSD out - (after 5days of down state), still recovery and
> reblanced stating...worried about it...
>
> Thanks
> Swami
>
>
> On Thu, Dec 8, 2016 at 6:40 AM, Brad Hubbard <bhubb...@redhat.com> wrote:
>
>>
>>
>> On Wed, Dec 7, 2016 at 9:11 PM, M Ranga Swami Reddy <swamire...@gmail.com
>> > wrote:
>>
>>> That's right..
>>> But, my question was: when an OSD down, all data will be moved to other
>>> OSDs from downed OSD. - Is this correct?
>>>
>>
>> No, only after it is marked out.
>>
>> "If an OSD is down and the degraded condition persists, Ceph may mark
>> the down OSD as out of the cluster and remap the data from the down OSD
>> to another OSD. The time between being marked down and being marked out
>> is controlled by mon osd down out interval, which is set to 300 seconds
>> by default."
>>
>>
>>> Now, I change the crushmap as out an OSD, then again data will be moved
>>> across the cluster?
>>>
>>
>>
>>>
>>> Thanks
>>> Swami
>>>
>>> On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote:
>>>
>>>> Hi
>>>>
>>>> Because "down" and "out" are different to ceph cluster
>>>>
>>>> Crush map of ceph is depends on how many osds are in ths cluster.
>>>>
>>>> Crush map doesn't change when osds are down. However crush map would
>>>> chage when the osds are absolutelly out.
>>>> Data location also will change, there fore rebalancing starts.
>>>>
>>>> Thanks
>>>> John Haan
>>>>
>>>>
>>>>
>>>> 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이
>>>> 작성:
>>>>
>>>>> Sure, will try with "*ceph osd crush reweight 0.0" *and update the
>>>>> status.
>>>>>
>>>>> Thanks
>>>>> Swami
>>>>>
>>>>> On Fri, Dec 2, 2016 at 8:15 PM, David Turner <
>>>>> david.tur...@storagecraft.com> wrote:
>>>>>
>>>>>> If you want to reweight only once when you have a failed disk that is
>>>>>> being balanced off of, set the crush weight for that osd to 0.0.  Then 
>>>>>> when
>>>>>> you fully remove the disk from the cluster it will not do any additional
>>>>>> backfilling.  Any change to the crush map will likely move data around,
>>>>>> even if you're removing an already "removed" osd.
>>>>>>
>>>>>> ------------------------------
>>>>>>
>>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer |
>>>>>> StorageCraft Technology Corporation <https://storagecraft.com>
>>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943
>>>>>> <%28385%29%20224-2943>
>>>>>>
>>>>>> ------------------------------
>>>>>>
>>>>>> If you are not the intended recipient of this message or received it
>>>>>> erroneously, please notify the sender and delete it, together with any
>>>>>> attachments, and be advised that any dissemination or copying of this
>>>>>> message is prohibited.
>>>>>>
>>>>>> ------------------------------
>>>>>>
>>>>>> ------------------------------
>>>>>> *From:* M Ranga Swami Reddy [swamire...@gmail.com]
>>>>>> *Sent:* Thursday, December 01, 2016 11:45 PM
>>>>>> *To:* David Turner
>>>>>> *Cc:* ceph-users
>>>>>> *Subject:* Re: [ceph-users] node and its OSDs down...
>>>>>>
>>>>>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which
>>>>>> started the recovery.
>>>>>> My worries is - why Ceph is doing the recovery, if an OSD is already
>>>>>> down and no more in the cluster. That means, ceph already
>>>>>> maintained down OSDs objects copied to another OSDs.. here is the ceph 
>>>>>> osd
>>>>>> tree o/p:
>>>>>> ===
>>>>>>
>>>>>> 227     0.91                            osd.227 down    0
>>>>>>
>>>>>> ....
>>>>>>
>>>>>> 250     0.91                            osd.250 down    0
>>>>>>
>>>>>> ===
>>>>>>
>>>>>>
>>>>>> So to avoid the recovery/rebalance , can I set the weight of OSD
>>>>>> (which was in down state). But is this weight setting also lead to
>>>>>> rebalance activity.
>>>>>>
>>>>>>
>>>>>> Thanks
>>>>>>
>>>>>> Swami
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner <
>>>>>> david.tur...@storagecraft.com> wrote:
>>>>>>
>>>>>>> I assume you also did ceph osd crush remove osd.<id>.  When you
>>>>>>> removed the osd that was down/out and balanced off of, you changed the
>>>>>>> weight of the host that it was on which triggers additional backfilling 
>>>>>>> to
>>>>>>> balance the crush map.
>>>>>>>
>>>>>>> ------------------------------
>>>>>>>
>>>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer |
>>>>>>> StorageCraft Technology Corporation <https://storagecraft.com>
>>>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020
>>>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943
>>>>>>> <%28385%29%20224-2943>
>>>>>>>
>>>>>>> ------------------------------
>>>>>>>
>>>>>>> If you are not the intended recipient of this message or received it
>>>>>>> erroneously, please notify the sender and delete it, together with any
>>>>>>> attachments, and be advised that any dissemination or copying of this
>>>>>>> message is prohibited.
>>>>>>>
>>>>>>> ------------------------------
>>>>>>>
>>>>>>> ------------------------------
>>>>>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of
>>>>>>> M Ranga Swami Reddy [swamire...@gmail.com]
>>>>>>> *Sent:* Thursday, December 01, 2016 3:03 AM
>>>>>>> *To:* ceph-users
>>>>>>> *Subject:* [ceph-users] node and its OSDs down...
>>>>>>>
>>>>>>> Hello,
>>>>>>> One of my ceph node with 20 OSDs down...After a couple of hours,
>>>>>>> ceph health is in OK state.
>>>>>>>
>>>>>>> Now, I tried to remove those OSDs, which were down state from
>>>>>>> ceph cluster...
>>>>>>> using the "ceh osd remove osd.<id>"
>>>>>>> then ceph clsuter started rebalancing...which is strange ..because
>>>>>>> thsoe OSDs are down for a long time and health also OK..
>>>>>>> my question - why recovery or reblance started when I remove the OSD
>>>>>>> (which was down).
>>>>>>>
>>>>>>> Thanks
>>>>>>> Swami
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@lists.ceph.com
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>
>>>
>>
>>
>> --
>> Cheers,
>> Brad
>>
>
>


-- 
Cheers,
Brad
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to