http://ceph.com/docs/master/rados/operations/placement-groups/
2013/10/22 HURTEVENT VINCENT <vincent.hurtev...@univ-lyon1.fr> > Hello, > > we're using a small Ceph cluster with 8 nodes, each 4 osds. People are > using it through instances and volumes in a Openstack platform. > > We're facing a HEALTH_ERR with full or near full osds : > > cluster 5942e110-ea2f-4bac-80f7-243fe3e35732 > health HEALTH_ERR 1 full osd(s); 13 near full osd(s) > monmap e1: 3 mons at {0= > 192.168.73.131:6789/0,1=192.168.73.135:6789/0,2=192.168.73.140:6789/0}, > election epoch 2974, quorum 0,1,2 0,1,2 > osdmap e4127: 32 osds: 32 up, 32 in full > pgmap v6055899: 10304 pgs: 10304 active+clean; 12444 GB data, 24953 GB > used, 4840 GB / 29793 GB avail > mdsmap e792: 1/1/1 up {0=2=up:active}, 2 up:standby > > Here is the dd output on these osds : > > /dev/sdc 932G 785G 147G 85% /data/ceph/osd/data/1 > /dev/sdd 932G 879G 53G 95% /data/ceph/osd/data/2 > /dev/sde 932G 765G 167G 83% /data/ceph/osd/data/3 > /dev/sdf 932G 754G 178G 81% /data/ceph/osd/data/4 > /dev/sdc 932G 799G 133G 86% /data/ceph/osd/data/6 > /dev/sdd 932G 818G 114G 88% /data/ceph/osd/data/7 > /dev/sde 932G 814G 118G 88% /data/ceph/osd/data/8 > /dev/sdf 932G 801G 131G 86% /data/ceph/osd/data/9 > /dev/sdc 932G 764G 168G 83% /data/ceph/osd/data/11 > /dev/sdd 932G 840G 92G 91% /data/ceph/osd/data/12 > /dev/sde 932G 699G 233G 76% /data/ceph/osd/data/13 > /dev/sdf 932G 721G 211G 78% /data/ceph/osd/data/14 > /dev/sdc 932G 778G 154G 84% /data/ceph/osd/data/16 > /dev/sdd 932G 820G 112G 88% /data/ceph/osd/data/17 > /dev/sde 932G 684G 248G 74% /data/ceph/osd/data/18 > /dev/sdf 932G 763G 169G 82% /data/ceph/osd/data/19 > /dev/sdc 932G 757G 175G 82% /data/ceph/osd/data/21 > /dev/sdd 932G 715G 217G 77% /data/ceph/osd/data/22 > /dev/sde 932G 762G 170G 82% /data/ceph/osd/data/23 > /dev/sdf 932G 728G 204G 79% /data/ceph/osd/data/24 > /dev/sdc 932G 841G 91G 91% /data/ceph/osd/data/26 > /dev/sdd 932G 795G 137G 86% /data/ceph/osd/data/27 > /dev/sde 932G 691G 241G 75% /data/ceph/osd/data/28 > /dev/sdf 932G 772G 160G 83% /data/ceph/osd/data/29 > /dev/sdc 932G 738G 195G 80% /data/ceph/osd/data/36 > /dev/sdd 932G 803G 129G 87% /data/ceph/osd/data/37 > /dev/sde 932G 783G 149G 85% /data/ceph/osd/data/38 > /dev/sdf 932G 844G 88G 91% /data/ceph/osd/data/39 > /dev/sdc 932G 885G 47G 96% /data/ceph/osd/data/31 > /dev/sdd 932G 708G 224G 76% /data/ceph/osd/data/32 > /dev/sde 932G 802G 130G 87% /data/ceph/osd/data/33 > /dev/sdf 932G 862G 70G 93% /data/ceph/osd/data/34 > > Some osds are really nearly full whereas some are ok I think (below 80%). > Wer're losing GBs we thought usable. There is no custom CRUSHmap, > > Is there a way to rebalance data between osds ? As I understand the > documentation, we should do this by adding new osd(s), switch off highly > used osd to force the cluster to rebuild data. > > Is there a way to do this without addind osds ? We 're using ceph version > 0.67.1. > > Thank you, > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- С уважением, Фасихов Ирек Нургаязович Моб.: +79229045757
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com