Hi Jan, Bryan, the cluster is healthy now again. Thanks for your support and tips!
I have identified all active+clean pgs which had their directories also located on those 2 full disks and deleted those directories from /var/lib/ceph/osd/ceph-N/current. That released lot of space from those disks and I could restart OSDs one by one. I had to manipulate weights using "ceph osd reweight" when the utilization of those OSDs exceeded other drives' average again. That helped. One strange thing - few pgs still remained in degraded or misplaced state until I set weights of all OSDs back to 1. When all degraded pgs turned to active+clean state, I still had "1 requests are blocked > 32 sec" error on one OSD (other than those which get full). Fixed by restarting that OSD's daemon. After that, I got the HEALTH_OK state. Thanks and greetings, Lukas On Thu, Feb 18, 2016 at 11:51 PM Stillwell, Bryan < bryan.stillw...@twcable.com> wrote: > When I've run into this situation I look for PGs that are on the full > drives, but are in an active+clean state in the cluster. That way I can > safely remove the PGs from the full drives and not have to risk data loss. > > It usually doesn't take much before you can restart the OSDs and let ceph > take care of the rest. > > Bryan > > From: ceph-users <ceph-users-boun...@lists.ceph.com> on behalf of Lukáš > Kubín <lukas.ku...@gmail.com> > Date: Thursday, February 18, 2016 at 2:39 PM > To: "ceph-users@lists.ceph.com" <ceph-users@lists.ceph.com> > Subject: Re: [ceph-users] How to recover from OSDs full in small cluster > > > >Hi, > >we've managed to release some space from our cluster. Now I would like to > >restart those 2 full OSDs. As they're completely full I probably need to > >delete some data from them. > > > >I would like to ask: Is it OK to delete all pg directories (eg. all > >subdirectories in /var/lib/ceph/osd/ceph-5/current/) and start the > >stopped OSD daemon then? This process seems most simple I'm just not sure > >if it is correct - if ceph can handle such > > situation. (I've noticed similar advice here: > > > http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd > >/ ) > > > >Another option as suggested by Jan is to remove OSD from cluster, and > >recreate them back. That presents more steps though and perhaps some more > >safety prerequirements (nobackfill?) to prevent more block > >movements/disks full while removing/readding. > > > >Thanks! > > > >Lukas > > > > > >Current status: > > > >[root@ceph1 ~]# ceph osd stat > > osdmap e1107: 12 osds: 10 up, 10 in; 29 remapped pgs > > > >[root@ceph1 ~]# ceph pg stat > > > >v21691144: 640 pgs: 503 active+clean, 29 active+remapped, 108 > >active+undersized+degraded; 1892 GB data, 3476 GB used, 1780 GB / 5256 GB > >avail; 0 B/s rd, 323 kB/s wr, 49 op/s; 42998/504482 objects degraded > >(8.523%); 10304/504482 objects misplaced (2.042%) > >[root@ceph1 ~]# df -h|grep osd > >/dev/sdg1 554G 383G 172G 70% /var/lib/ceph/osd/ceph-3 > >/dev/sdf1 554G 401G 154G 73% /var/lib/ceph/osd/ceph-2 > >/dev/sde1 554G 381G 174G 69% /var/lib/ceph/osd/ceph-0 > >/dev/sdb1 275G 275G 20K 100% /var/lib/ceph/osd/ceph-5 > >/dev/sdd1 554G 554G 20K 100% /var/lib/ceph/osd/ceph-4 > >/dev/sdc1 554G 359G 196G 65% /var/lib/ceph/osd/ceph-1 > > > >[root@ceph1 ~]# ceph osd tree > >ID WEIGHT TYPE NAME UP/DOWN REWEIGHT PRIMARY-AFFINITY > >-1 5.93991 root default > >-2 2.96996 host ceph1 > > 0 0.53999 osd.0 up 1.00000 1.00000 > > 1 0.53999 osd.1 up 1.00000 1.00000 > > 2 0.53999 osd.2 up 1.00000 1.00000 > > 3 0.53999 osd.3 up 1.00000 1.00000 > > 4 0.53999 osd.4 down 0 1.00000 > > 5 0.26999 osd.5 down 0 1.00000 > >-3 2.96996 host ceph2 > > 6 0.53999 osd.6 up 1.00000 1.00000 > > 7 0.53999 osd.7 up 1.00000 1.00000 > > 8 0.53999 osd.8 up 1.00000 1.00000 > > 9 0.53999 osd.9 up 1.00000 1.00000 > >10 0.53999 osd.10 up 1.00000 1.00000 > >11 0.26999 osd.11 up 1.00000 1.00000 > > > > > > > >On Wed, Feb 17, 2016 at 9:43 PM Lukáš Kubín <lukas.ku...@gmail.com> > wrote: > > > > > >Hi, > >I'm running a very small setup of 2 nodes with 6 OSDs each. There are 2 > >pools, each of size=2. Today, one of our OSDs got full, another 2 near > >full. Cluster turned into ERR state. I have noticed uneven space > >distribution among OSD drives between 70 and > > 100 perce. I have realized there's a low amount of pgs in those 2 pools > >(128 each) and increased one of them to 512, expecting a magic to happen > >and redistribute the space evenly. > > > >Well, something happened - another OSD became full during the > >redistribution and cluster stopped both OSDs and marked them down. After > >some hours the remaining drives partially rebalanced and cluster get to > >WARN state. > > > >I've deleted 3 placement group directories from one of the full OSD's > >filesystem which allowed me to start it up again. Soon, however this > >drive became full again. > > > >So now, there are 2 of 12 OSDs down, cluster is in WARN and I have no > >drives to add. > > > >Is there a way how to get out of this situation without adding OSDs? I > >will attempt to release some space, just waiting for colleague to > >identify RBD volumes (openstack images and volumes) which can be deleted. > > > >Thank you. > > > >Lukas > > > > > >This is my cluster state now: > > > >[root@compute1 ~]# ceph -w > > cluster d35174e9-4d17-4b5e-80f2-02440e0980d5 > > health HEALTH_WARN > > 10 pgs backfill_toofull > > 114 pgs degraded > > 114 pgs stuck degraded > > 147 pgs stuck unclean > > 114 pgs stuck undersized > > 114 pgs undersized > > 1 requests are blocked > 32 sec > > recovery 56923/640724 objects degraded (8.884%) > > recovery 29122/640724 objects misplaced (4.545%) > > 3 near full osd(s) > > monmap e3: 3 mons at > >{compute1= > 10.255.242.14:6789/0,compute2=10.255.242.15:6789/0,compute3=10.2 > >55.242.16:6789/0 > >< > http://10.255.242.14:6789/0,compute2=10.255.242.15:6789/0,compute3=10.255 > >.242.16:6789/0>} > > election epoch 128, quorum 0,1,2 compute1,compute2,compute3 > > osdmap e1073: 12 osds: 10 up, 10 in; 39 remapped pgs > > pgmap v21609066: 640 pgs, 2 pools, 2390 GB data, 309 kobjects > > 4365 GB used, 890 GB / 5256 GB avail > > 56923/640724 objects degraded (8.884%) > > 29122/640724 objects misplaced (4.545%) > > 493 active+clean > > 108 active+undersized+degraded > > 29 active+remapped > > 6 active+undersized+degraded+remapped+backfill_toofull > > 4 active+remapped+backfill_toofull > > > > > >[root@ceph1 ~]# df|grep osd > >/dev/sdg1 580496384 500066812 80429572 87% > >/var/lib/ceph/osd/ceph-3 > >/dev/sdf1 580496384 502131428 78364956 87% > >/var/lib/ceph/osd/ceph-2 > >/dev/sde1 580496384 506927100 73569284 88% > >/var/lib/ceph/osd/ceph-0 > >/dev/sdb1 287550208 287550188 20 100% > >/var/lib/ceph/osd/ceph-5 > >/dev/sdd1 580496384 580496364 20 100% > >/var/lib/ceph/osd/ceph-4 > >/dev/sdc1 580496384 478675672 101820712 83% > >/var/lib/ceph/osd/ceph-1 > > > > > >[root@ceph2 ~]# df|grep osd > >/dev/sdf1 580496384 448689872 131806512 78% > >/var/lib/ceph/osd/ceph-7 > >/dev/sdb1 287550208 227054336 60495872 79% > >/var/lib/ceph/osd/ceph-11 > >/dev/sdd1 580496384 464175196 116321188 80% > >/var/lib/ceph/osd/ceph-10 > >/dev/sdc1 580496384 489451300 91045084 85% > >/var/lib/ceph/osd/ceph-6 > >/dev/sdg1 580496384 470559020 109937364 82% > >/var/lib/ceph/osd/ceph-9 > >/dev/sde1 580496384 490289388 90206996 85% > >/var/lib/ceph/osd/ceph-8 > > > > > >[root@ceph2 ~]# ceph df > >GLOBAL: > > SIZE AVAIL RAW USED %RAW USED > > 5256G 890G 4365G 83.06 > >POOLS: > > NAME ID USED %USED MAX AVAIL OBJECTS > > glance 6 1714G 32.61 385G 219579 > > cinder 7 676G 12.86 385G 97488 > > > > > >[root@ceph2 ~]# ceph osd pool get glance pg_num > >pg_num: 512 > > > >[root@ceph2 ~]# ceph osd pool get cinder pg_num > >pg_num: 128 > > > ________________________________ > > This E-mail and any of its attachments may contain Time Warner Cable > proprietary information, which is privileged, confidential, or subject to > copyright belonging to Time Warner Cable. This E-mail is intended solely > for the use of the individual or entity to which it is addressed. If you > are not the intended recipient of this E-mail, you are hereby notified that > any dissemination, distribution, copying, or action taken in relation to > the contents of and attachments to this E-mail is strictly prohibited and > may be unlawful. If you have received this E-mail in error, please notify > the sender immediately and permanently delete the original and any copy of > this E-mail and any printout. > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com