If you are not sure about what weight to put , ‘ceph osd reweight-by-utilization’ should also do the job for you automatically..
Thanks & Regards Somnath From: ceph-users [mailto:ceph-users-boun...@lists.ceph.com] On Behalf Of Jan Schermer Sent: Wednesday, February 17, 2016 12:48 PM To: Lukáš Kubín Cc: ceph-users@lists.ceph.com Subject: Re: [ceph-users] How to recover from OSDs full in small cluster Ahoj ;-) You can reweight them temporarily, that shifts the data from the full drives. ceph osd reweight osd.XX YY (XX = the number of full OSD, YY is "weight" which default to 1) This is different from "crush reweight" which defaults to drive size in TB. Beware that reweighting will (afaik) only shuffle the data to other local drives, so you should reweight both the full drives at the same time and only by little bit at a time (0.95 is a good starting point). Jan On 17 Feb 2016, at 21:43, Lukáš Kubín <lukas.ku...@gmail.com<mailto:lukas.ku...@gmail.com>> wrote: Hi, I'm running a very small setup of 2 nodes with 6 OSDs each. There are 2 pools, each of size=2. Today, one of our OSDs got full, another 2 near full. Cluster turned into ERR state. I have noticed uneven space distribution among OSD drives between 70 and 100 perce. I have realized there's a low amount of pgs in those 2 pools (128 each) and increased one of them to 512, expecting a magic to happen and redistribute the space evenly. Well, something happened - another OSD became full during the redistribution and cluster stopped both OSDs and marked them down. After some hours the remaining drives partially rebalanced and cluster get to WARN state. I've deleted 3 placement group directories from one of the full OSD's filesystem which allowed me to start it up again. Soon, however this drive became full again. So now, there are 2 of 12 OSDs down, cluster is in WARN and I have no drives to add. Is there a way how to get out of this situation without adding OSDs? I will attempt to release some space, just waiting for colleague to identify RBD volumes (openstack images and volumes) which can be deleted. Thank you. Lukas This is my cluster state now: [root@compute1 ~]# ceph -w cluster d35174e9-4d17-4b5e-80f2-02440e0980d5 health HEALTH_WARN 10 pgs backfill_toofull 114 pgs degraded 114 pgs stuck degraded 147 pgs stuck unclean 114 pgs stuck undersized 114 pgs undersized 1 requests are blocked > 32 sec recovery 56923/640724 objects degraded (8.884%) recovery 29122/640724 objects misplaced (4.545%) 3 near full osd(s) monmap e3: 3 mons at {compute1=10.255.242.14:6789/0,compute2=10.255.242.15:6789/0,compute3=10.255.242.16:6789/0<http://10.255.242.14:6789/0,compute2=10.255.242.15:6789/0,compute3=10.255.242.16:6789/0>} election epoch 128, quorum 0,1,2 compute1,compute2,compute3 osdmap e1073: 12 osds: 10 up, 10 in; 39 remapped pgs pgmap v21609066: 640 pgs, 2 pools, 2390 GB data, 309 kobjects 4365 GB used, 890 GB / 5256 GB avail 56923/640724 objects degraded (8.884%) 29122/640724 objects misplaced (4.545%) 493 active+clean 108 active+undersized+degraded 29 active+remapped 6 active+undersized+degraded+remapped+backfill_toofull 4 active+remapped+backfill_toofull [root@ceph1 ~]# df|grep osd /dev/sdg1 580496384 500066812 80429572 87% /var/lib/ceph/osd/ceph-3 /dev/sdf1 580496384 502131428 78364956 87% /var/lib/ceph/osd/ceph-2 /dev/sde1 580496384 506927100 73569284 88% /var/lib/ceph/osd/ceph-0 /dev/sdb1 287550208 287550188 20 100% /var/lib/ceph/osd/ceph-5 /dev/sdd1 580496384 580496364 20 100% /var/lib/ceph/osd/ceph-4 /dev/sdc1 580496384 478675672 101820712 83% /var/lib/ceph/osd/ceph-1 [root@ceph2 ~]# df|grep osd /dev/sdf1 580496384 448689872 131806512 78% /var/lib/ceph/osd/ceph-7 /dev/sdb1 287550208 227054336 60495872 79% /var/lib/ceph/osd/ceph-11 /dev/sdd1 580496384 464175196 116321188 80% /var/lib/ceph/osd/ceph-10 /dev/sdc1 580496384 489451300 91045084 85% /var/lib/ceph/osd/ceph-6 /dev/sdg1 580496384 470559020 109937364 82% /var/lib/ceph/osd/ceph-9 /dev/sde1 580496384 490289388 90206996 85% /var/lib/ceph/osd/ceph-8 [root@ceph2 ~]# ceph df GLOBAL: SIZE AVAIL RAW USED %RAW USED 5256G 890G 4365G 83.06 POOLS: NAME ID USED %USED MAX AVAIL OBJECTS glance 6 1714G 32.61 385G 219579 cinder 7 676G 12.86 385G 97488 [root@ceph2 ~]# ceph osd pool get glance pg_num pg_num: 512 [root@ceph2 ~]# ceph osd pool get cinder pg_num pg_num: 128 _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com<mailto:ceph-users@lists.ceph.com> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com PLEASE NOTE: The information contained in this electronic mail message is intended only for the use of the designated recipient(s) named above. If the reader of this message is not the intended recipient, you are hereby notified that you have received this message in error and that any review, dissemination, distribution, or copying of this message is strictly prohibited. If you have received this communication in error, please notify the sender by telephone or e-mail (as shown above) immediately and destroy any and all copies of this message in your possession (whether hard copies or electronically stored copies).
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com