Hi,
On 04/26/2016 12:32 PM, SCHAER Frederic wrote:
Hi,
One simple/quick question.
In my ceph cluster, I had a disk wich was in predicted failure. It was
so much in predicted failure that the ceph OSD daemon crashed.
After the OSD crashed, ceph moved data correctly (or at least that’s
what I thought), and a ceph –s was giving a “HEALTH_OK”.
Perfect.
I tride to tell ceph to mark the OSD down : it told me the OSD was
already down… fine.
Then I ran this :
ID=43 ; ceph osd down $ID ; ceph auth del osd.$ID ; ceph osd rm $ID ;
ceph osd crush remove osd.$ID
*snipsnap*
Removing the dead OSD entry changed the CRUSH weight for the host,
resulting in a redistribution of the data.
The easiest way to prevent this is setting the CRUSH weight to 0.0
before removing the OSD. This will trigger backfilling only once.
Regards,
Burkhard.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com