Hi All,

I just removed an OSD from our cluster following the steps on 
http://ceph.com/docs/master/rados/operations/add-or-rm-osds/

First I set the OSD as out,

ceph osd out osd.0

This emptied the OSD and eventually health of the cluster came back to 
normal/ok. and OSD was up and out. (took about 2-3 hours) (OSD.0 used space 
before setting as OUT was 900~ GB after rebalance took place OSD Usage was 
~150MB)

Once this was all ok I then proceeded to STOP the OSD.

service ceph stop osd.0

checked cluster health and all looked ok, then I decided to remove the osd 
using the following commands.

ceph osd crush remove osd.0
ceph auth del osd.0
ceph osd rm 0


Now our cluster says
health HEALTH_WARN 414 pgs backfill; 12 pgs backfilling; 19 pgs recovering; 344 
pgs recovery_wait; 789 pgs stuck unclean; recovery 390967/10986568 objects 
degraded (3.559%)

before using the removal procedure everything was "ok" and the osd.0 had been 
emptied and seemingly rebalanced.

Any ideas why its rebalancing again?

we're using Ubuntu 12.04 w/ Ceph 80.8 & Kernel 3.13.0-43-generic 
#72~precise1-Ubuntu SMP Tue Dec 9 12:14:18 UTC 2014 x86_64 x86_64 x86_64 
GNU/Linux



Regards,
Quenten Grasso
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to