That's right.. But, my question was: when an OSD down, all data will be moved to other OSDs from downed OSD. - Is this correct? Now, I change the crushmap as out an OSD, then again data will be moved across the cluster?
Thanks Swami On Wed, Dec 7, 2016 at 2:14 PM, 한승진 <yongi...@gmail.com> wrote: > Hi > > Because "down" and "out" are different to ceph cluster > > Crush map of ceph is depends on how many osds are in ths cluster. > > Crush map doesn't change when osds are down. However crush map would chage > when the osds are absolutelly out. > Data location also will change, there fore rebalancing starts. > > Thanks > John Haan > > > > 2016. 12. 3. 오후 5:27에 "M Ranga Swami Reddy" <swamire...@gmail.com>님이 작성: > >> Sure, will try with "*ceph osd crush reweight 0.0" *and update the >> status. >> >> Thanks >> Swami >> >> On Fri, Dec 2, 2016 at 8:15 PM, David Turner < >> david.tur...@storagecraft.com> wrote: >> >>> If you want to reweight only once when you have a failed disk that is >>> being balanced off of, set the crush weight for that osd to 0.0. Then when >>> you fully remove the disk from the cluster it will not do any additional >>> backfilling. Any change to the crush map will likely move data around, >>> even if you're removing an already "removed" osd. >>> >>> ------------------------------ >>> >>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>> StorageCraft >>> Technology Corporation <https://storagecraft.com> >>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>> Office: 801.871.2760 <(801)%20871-2760> | Mobile: 385.224.2943 >>> <(385)%20224-2943> >>> >>> ------------------------------ >>> >>> If you are not the intended recipient of this message or received it >>> erroneously, please notify the sender and delete it, together with any >>> attachments, and be advised that any dissemination or copying of this >>> message is prohibited. >>> >>> ------------------------------ >>> >>> ------------------------------ >>> *From:* M Ranga Swami Reddy [swamire...@gmail.com] >>> *Sent:* Thursday, December 01, 2016 11:45 PM >>> *To:* David Turner >>> *Cc:* ceph-users >>> *Subject:* Re: [ceph-users] node and its OSDs down... >>> >>> Hi David - Yep, I did the "ceph osd crush remove osd.<id>", which >>> started the recovery. >>> My worries is - why Ceph is doing the recovery, if an OSD is already >>> down and no more in the cluster. That means, ceph already >>> maintained down OSDs objects copied to another OSDs.. here is the ceph osd >>> tree o/p: >>> === >>> >>> 227 0.91 osd.227 down 0 >>> >>> .... >>> >>> 250 0.91 osd.250 down 0 >>> >>> === >>> >>> >>> So to avoid the recovery/rebalance , can I set the weight of OSD (which >>> was in down state). But is this weight setting also lead to rebalance >>> activity. >>> >>> >>> Thanks >>> >>> Swami >>> >>> >>> >>> On Thu, Dec 1, 2016 at 8:07 PM, David Turner < >>> david.tur...@storagecraft.com> wrote: >>> >>>> I assume you also did ceph osd crush remove osd.<id>. When you removed >>>> the osd that was down/out and balanced off of, you changed the weight of >>>> the host that it was on which triggers additional backfilling to balance >>>> the crush map. >>>> >>>> ------------------------------ >>>> >>>> <https://storagecraft.com> David Turner | Cloud Operations Engineer | >>>> StorageCraft >>>> Technology Corporation <https://storagecraft.com> >>>> 380 Data Drive Suite 300 | Draper | Utah | 84020 >>>> Office: 801.871.2760 <(801)%20871-2760> | Mobile: 385.224.2943 >>>> <(385)%20224-2943> >>>> >>>> ------------------------------ >>>> >>>> If you are not the intended recipient of this message or received it >>>> erroneously, please notify the sender and delete it, together with any >>>> attachments, and be advised that any dissemination or copying of this >>>> message is prohibited. >>>> >>>> ------------------------------ >>>> >>>> ------------------------------ >>>> *From:* ceph-users [ceph-users-boun...@lists.ceph.com] on behalf of M >>>> Ranga Swami Reddy [swamire...@gmail.com] >>>> *Sent:* Thursday, December 01, 2016 3:03 AM >>>> *To:* ceph-users >>>> *Subject:* [ceph-users] node and its OSDs down... >>>> >>>> Hello, >>>> One of my ceph node with 20 OSDs down...After a couple of hours, >>>> ceph health is in OK state. >>>> >>>> Now, I tried to remove those OSDs, which were down state from >>>> ceph cluster... >>>> using the "ceh osd remove osd.<id>" >>>> then ceph clsuter started rebalancing...which is strange ..because >>>> thsoe OSDs are down for a long time and health also OK.. >>>> my question - why recovery or reblance started when I remove the OSD >>>> (which was down). >>>> >>>> Thanks >>>> Swami >>>> >>>> >>> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com