On Fri, Dec 9, 2016 at 3:28 PM, M Ranga Swami Reddy <swamire...@gmail.com> wrote:
> Confused ... > a few OSDs down and cluster done the recovery and reblanced to HEALTH OK > state. > Now I can could that down OSDs are down state from crushmap and are not > part of OSD up or in state. > After 5 days or says, still the same state. > How or when Ceph will make the down state OSDs to out state? I guess > ceph don't do it. > 300 seconds by default Any flags set? noout perhaps? Now I ran the OSD out - (after 5days of down state), still recovery and > reblanced stating...worried about it... > > Thanks > Swami > > > On Thu, Dec 8, 2016 at 6:40 AM, Brad Hubbard <bhubb...@redhat.com> wrote: > >> >> >> On Wed, Dec 7, 2016 at 9:11 PM, M Ranga Swami Reddy <swamire...@gmail.com >> > wrote: >> >>> That's right.. >>> But, my question was: when an OSD down, all data will be moved to other >>> OSDs from downed OSD. - Is this correct? >>> >> >> No, only after it is marked out. >> >> "If an OSD is down and the degraded condition persists, Ceph may mark >> the down OSD as out of the cluster and remap the data from the down OSD >> to another OSD. The time between being marked down and being marked out >> is controlled by mon osd down out interval, which is set to 300 seconds >> by default." >> >> >>> Now, I change the crushmap as out an OSD, then again data will be moved >>> across the cluster? >>> >> >> >>> >>> Thanks >>> Swami >>> >>> On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote: >>> >>>> Hi >>>> >>>> Because "down" and "out" are different to ceph cluster >>>> >>>> Crush map of ceph is depends on how many osds are in ths cluster. >>>> >>>> Crush map doesn't change when osds are down. However crush map would >>>> chage when the osds are absolutelly out. >>>> Data location also will change, there fore rebalancing starts. >>>> >>>> Thanks >>>> John Haan >>>> >>>> >>>> >>>> 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이 >>>> 작성: >>>> >>>>> Sure, will try with "*ceph osd crush reweight 0.0" *and update the >>>>> status. >>>>> >>>>> Thanks >>>>> Swami >>>>> >>>>> On Fri, Dec 2, 2016 at 8:15 PM, David Turner < >>>>> david.tur...@storagecraft.com> wrote: >>>>> >>>>>> If you want to reweight only once when you have a failed disk that is >>>>>> being balanced off of, set the crush weight for that osd to 0.0. Then >>>>>> when >>>>>> you fully remove the disk from the cluster it will not do any additional >>>>>> backfilling. Any change to the crush map will likely move data around, >>>>>> even if you're removing an already "removed" osd. >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>>>>> StorageCraft Technology Corporation <https://storagecraft.com> >>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943 >>>>>> <%28385%29%20224-2943> >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> If you are not the intended recipient of this message or received it >>>>>> erroneously, please notify the sender and delete it, together with any >>>>>> attachments, and be advised that any dissemination or copying of this >>>>>> message is prohibited. >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> ------------------------------ >>>>>> *From:* M Ranga Swami Reddy [swamire...@gmail.com] >>>>>> *Sent:* Thursday, December 01, 2016 11:45 PM >>>>>> *To:* David Turner >>>>>> *Cc:* ceph-users >>>>>> *Subject:* Re: [ceph-users] node and its OSDs down... >>>>>> >>>>>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which >>>>>> started the recovery. >>>>>> My worries is - why Ceph is doing the recovery, if an OSD is already >>>>>> down and no more in the cluster. That means, ceph already >>>>>> maintained down OSDs objects copied to another OSDs.. here is the ceph >>>>>> osd >>>>>> tree o/p: >>>>>> === >>>>>> >>>>>> 227 0.91 osd.227 down 0 >>>>>> >>>>>> .... >>>>>> >>>>>> 250 0.91 osd.250 down 0 >>>>>> >>>>>> === >>>>>> >>>>>> >>>>>> So to avoid the recovery/rebalance , can I set the weight of OSD >>>>>> (which was in down state). But is this weight setting also lead to >>>>>> rebalance activity. >>>>>> >>>>>> >>>>>> Thanks >>>>>> >>>>>> Swami >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner < >>>>>> david.tur...@storagecraft.com> wrote: >>>>>> >>>>>>> I assume you also did ceph osd crush remove osd.<id>. When you >>>>>>> removed the osd that was down/out and balanced off of, you changed the >>>>>>> weight of the host that it was on which triggers additional backfilling >>>>>>> to >>>>>>> balance the crush map. >>>>>>> >>>>>>> ------------------------------ >>>>>>> >>>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>>>>>> StorageCraft Technology Corporation <https://storagecraft.com> >>>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943 >>>>>>> <%28385%29%20224-2943> >>>>>>> >>>>>>> ------------------------------ >>>>>>> >>>>>>> If you are not the intended recipient of this message or received it >>>>>>> erroneously, please notify the sender and delete it, together with any >>>>>>> attachments, and be advised that any dissemination or copying of this >>>>>>> message is prohibited. >>>>>>> >>>>>>> ------------------------------ >>>>>>> >>>>>>> ------------------------------ >>>>>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of >>>>>>> M Ranga Swami Reddy [swamire...@gmail.com] >>>>>>> *Sent:* Thursday, December 01, 2016 3:03 AM >>>>>>> *To:* ceph-users >>>>>>> *Subject:* [ceph-users] node and its OSDs down... >>>>>>> >>>>>>> Hello, >>>>>>> One of my ceph node with 20 OSDs down...After a couple of hours, >>>>>>> ceph health is in OK state. >>>>>>> >>>>>>> Now, I tried to remove those OSDs, which were down state from >>>>>>> ceph cluster... >>>>>>> using the "ceh osd remove osd.<id>" >>>>>>> then ceph clsuter started rebalancing...which is strange ..because >>>>>>> thsoe OSDs are down for a long time and health also OK.. >>>>>>> my question - why recovery or reblance started when I remove the OSD >>>>>>> (which was down). >>>>>>> >>>>>>> Thanks >>>>>>> Swami >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@lists.ceph.com >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>>>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>> >> >> >> -- >> Cheers, >> Brad >> > > -- Cheers, Brad
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com