Confused ... a few OSDs down and cluster done the recovery and reblanced to HEALTH OK state. Now I can could that down OSDs are down state from crushmap and are not part of OSD up or in state. After 5 days or says, still the same state. How or when Ceph will make the down state OSDs to out state? I guess ceph don't do it. Now I ran the OSD out - (after 5days of down state), still recovery and reblanced stating...worried about it...
Thanks Swami On Thu, Dec 8, 2016 at 6:40 AM, Brad Hubbard <bhubb...@redhat.com> wrote: > > > On Wed, Dec 7, 2016 at 9:11 PM, M Ranga Swami Reddy <swamire...@gmail.com> > wrote: > >> That's right.. >> But, my question was: when an OSD down, all data will be moved to other >> OSDs from downed OSD. - Is this correct? >> > > No, only after it is marked out. > > "If an OSD is down and the degraded condition persists, Ceph may mark the > down OSD as out of the cluster and remap the data from the down OSD to > another OSD. The time between being marked down and being marked out is > controlled by mon osd down out interval, which is set to 300 seconds by > default." > > >> Now, I change the crushmap as out an OSD, then again data will be moved >> across the cluster? >> > > >> >> Thanks >> Swami >> >> On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote: >> >>> Hi >>> >>> Because "down" and "out" are different to ceph cluster >>> >>> Crush map of ceph is depends on how many osds are in ths cluster. >>> >>> Crush map doesn't change when osds are down. However crush map would >>> chage when the osds are absolutelly out. >>> Data location also will change, there fore rebalancing starts. >>> >>> Thanks >>> John Haan >>> >>> >>> >>> 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이 작성: >>> >>>> Sure, will try with "*ceph osd crush reweight 0.0" *and update the >>>> status. >>>> >>>> Thanks >>>> Swami >>>> >>>> On Fri, Dec 2, 2016 at 8:15 PM, David Turner < >>>> david.tur...@storagecraft.com> wrote: >>>> >>>>> If you want to reweight only once when you have a failed disk that is >>>>> being balanced off of, set the crush weight for that osd to 0.0. Then >>>>> when >>>>> you fully remove the disk from the cluster it will not do any additional >>>>> backfilling. Any change to the crush map will likely move data around, >>>>> even if you're removing an already "removed" osd. >>>>> >>>>> ------------------------------ >>>>> >>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>>>> StorageCraft >>>>> Technology Corporation <https://storagecraft.com> >>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943 >>>>> <%28385%29%20224-2943> >>>>> >>>>> ------------------------------ >>>>> >>>>> If you are not the intended recipient of this message or received it >>>>> erroneously, please notify the sender and delete it, together with any >>>>> attachments, and be advised that any dissemination or copying of this >>>>> message is prohibited. >>>>> >>>>> ------------------------------ >>>>> >>>>> ------------------------------ >>>>> *From:* M Ranga Swami Reddy [swamire...@gmail.com] >>>>> *Sent:* Thursday, December 01, 2016 11:45 PM >>>>> *To:* David Turner >>>>> *Cc:* ceph-users >>>>> *Subject:* Re: [ceph-users] node and its OSDs down... >>>>> >>>>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which >>>>> started the recovery. >>>>> My worries is - why Ceph is doing the recovery, if an OSD is already >>>>> down and no more in the cluster. That means, ceph already >>>>> maintained down OSDs objects copied to another OSDs.. here is the ceph osd >>>>> tree o/p: >>>>> === >>>>> >>>>> 227 0.91 osd.227 down 0 >>>>> >>>>> .... >>>>> >>>>> 250 0.91 osd.250 down 0 >>>>> >>>>> === >>>>> >>>>> >>>>> So to avoid the recovery/rebalance , can I set the weight of OSD >>>>> (which was in down state). But is this weight setting also lead to >>>>> rebalance activity. >>>>> >>>>> >>>>> Thanks >>>>> >>>>> Swami >>>>> >>>>> >>>>> >>>>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner < >>>>> david.tur...@storagecraft.com> wrote: >>>>> >>>>>> I assume you also did ceph osd crush remove osd.<id>. When you >>>>>> removed the osd that was down/out and balanced off of, you changed the >>>>>> weight of the host that it was on which triggers additional backfilling >>>>>> to >>>>>> balance the crush map. >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>>>>> StorageCraft Technology Corporation <https://storagecraft.com> >>>>>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>>>>> Office: 801.871.2760 <%28801%29%20871-2760> | Mobile: 385.224.2943 >>>>>> <%28385%29%20224-2943> >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> If you are not the intended recipient of this message or received it >>>>>> erroneously, please notify the sender and delete it, together with any >>>>>> attachments, and be advised that any dissemination or copying of this >>>>>> message is prohibited. >>>>>> >>>>>> ------------------------------ >>>>>> >>>>>> ------------------------------ >>>>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of >>>>>> M Ranga Swami Reddy [swamire...@gmail.com] >>>>>> *Sent:* Thursday, December 01, 2016 3:03 AM >>>>>> *To:* ceph-users >>>>>> *Subject:* [ceph-users] node and its OSDs down... >>>>>> >>>>>> Hello, >>>>>> One of my ceph node with 20 OSDs down...After a couple of hours, >>>>>> ceph health is in OK state. >>>>>> >>>>>> Now, I tried to remove those OSDs, which were down state from >>>>>> ceph cluster... >>>>>> using the "ceh osd remove osd.<id>" >>>>>> then ceph clsuter started rebalancing...which is strange ..because >>>>>> thsoe OSDs are down for a long time and health also OK.. >>>>>> my question - why recovery or reblance started when I remove the OSD >>>>>> (which was down). >>>>>> >>>>>> Thanks >>>>>> Swami >>>>>> >>>>>> >>>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@lists.ceph.com >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > > -- > Cheers, > Brad >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com