You're right, the "full" osd was still up and in until I increased the pg values of one of the pools. The redistribution has not completed yet and perhaps that's what is still filling the drive. With this info - do you think I'm still safe to follow the steps suggested in previous post?
Thanks! Lukas On Wed, Feb 17, 2016 at 10:29 PM Jan Schermer <j...@schermer.cz> wrote: > Something must be on those 2 OSDs that ate all that space - ceph by > default doesn't allow OSD to get completely full (filesystem-wise) and from > what you've shown those filesystems are really really full. > OSDs don't usually go down when "full" (95%) .. or do they? I don't think > so... so the reason they stopped is likely a completely full filfeystem. > You have to move something out of the way, restart those OSDs with lower > reweight and hopefully everything will be good. > > Jan > > > On 17 Feb 2016, at 22:22, Lukáš Kubín <lukas.ku...@gmail.com> wrote: > > Ahoj Jan, thanks for the quick hint! > > Those 2 OSDs are currently full and down. How should I handle that? Is it > ok that I delete some pg directories again and start the OSD daemons, on > both drives in parallel. Then set the weights as recommended ? > > What effect should I expect then - will the cluster attempt to move some > pgs out of these drives to different local OSDs? I'm asking because when > I've attempted to delete pg dirs and restart OSD for the first time, the > OSD get full again very fast. > > Thank you. > > Lukas > > > > On Wed, Feb 17, 2016 at 9:48 PM Jan Schermer <j...@schermer.cz> wrote: > >> Ahoj ;-) >> >> You can reweight them temporarily, that shifts the data from the full >> drives. >> >> ceph osd reweight osd.XX YY >> (XX = the number of full OSD, YY is "weight" which default to 1) >> >> This is different from "crush reweight" which defaults to drive size in >> TB. >> >> Beware that reweighting will (afaik) only shuffle the data to other local >> drives, so you should reweight both the full drives at the same time and >> only by little bit at a time (0.95 is a good starting point). >> >> Jan >> >> >> >> On 17 Feb 2016, at 21:43, Lukáš Kubín <lukas.ku...@gmail.com> wrote: >> >> Hi, >> I'm running a very small setup of 2 nodes with 6 OSDs each. There are 2 >> pools, each of size=2. Today, one of our OSDs got full, another 2 near >> full. Cluster turned into ERR state. I have noticed uneven space >> distribution among OSD drives between 70 and 100 perce. I have realized >> there's a low amount of pgs in those 2 pools (128 each) and increased one >> of them to 512, expecting a magic to happen and redistribute the space >> evenly. >> >> Well, something happened - another OSD became full during the >> redistribution and cluster stopped both OSDs and marked them down. After >> some hours the remaining drives partially rebalanced and cluster get to >> WARN state. >> >> I've deleted 3 placement group directories from one of the full OSD's >> filesystem which allowed me to start it up again. Soon, however this drive >> became full again. >> >> So now, there are 2 of 12 OSDs down, cluster is in WARN and I have no >> drives to add. >> >> Is there a way how to get out of this situation without adding OSDs? I >> will attempt to release some space, just waiting for colleague to identify >> RBD volumes (openstack images and volumes) which can be deleted. >> >> Thank you. >> >> Lukas >> >> >> This is my cluster state now: >> >> [root@compute1 ~]# ceph -w >> cluster d35174e9-4d17-4b5e-80f2-02440e0980d5 >> health HEALTH_WARN >> 10 pgs backfill_toofull >> 114 pgs degraded >> 114 pgs stuck degraded >> 147 pgs stuck unclean >> 114 pgs stuck undersized >> 114 pgs undersized >> 1 requests are blocked > 32 sec >> recovery 56923/640724 objects degraded (8.884%) >> recovery 29122/640724 objects misplaced (4.545%) >> 3 near full osd(s) >> monmap e3: 3 mons at {compute1= >> 10.255.242.14:6789/0,compute2=10.255.242.15:6789/0,compute3=10.255.242.16:6789/0 >> } >> election epoch 128, quorum 0,1,2 compute1,compute2,compute3 >> osdmap e1073: 12 osds: 10 up, 10 in; 39 remapped pgs >> pgmap v21609066: 640 pgs, 2 pools, 2390 GB data, 309 kobjects >> 4365 GB used, 890 GB / 5256 GB avail >> 56923/640724 objects degraded (8.884%) >> 29122/640724 objects misplaced (4.545%) >> 493 active+clean >> 108 active+undersized+degraded >> 29 active+remapped >> 6 active+undersized+degraded+remapped+backfill_toofull >> 4 active+remapped+backfill_toofull >> >> [root@ceph1 ~]# df|grep osd >> /dev/sdg1 580496384 500066812 80429572 87% >> /var/lib/ceph/osd/ceph-3 >> /dev/sdf1 580496384 502131428 78364956 87% >> /var/lib/ceph/osd/ceph-2 >> /dev/sde1 580496384 506927100 73569284 88% >> /var/lib/ceph/osd/ceph-0 >> /dev/sdb1 287550208 287550188 20 100% >> /var/lib/ceph/osd/ceph-5 >> /dev/sdd1 580496384 580496364 20 100% >> /var/lib/ceph/osd/ceph-4 >> /dev/sdc1 580496384 478675672 101820712 83% >> /var/lib/ceph/osd/ceph-1 >> >> [root@ceph2 ~]# df|grep osd >> /dev/sdf1 580496384 448689872 131806512 78% >> /var/lib/ceph/osd/ceph-7 >> /dev/sdb1 287550208 227054336 60495872 79% >> /var/lib/ceph/osd/ceph-11 >> /dev/sdd1 580496384 464175196 116321188 80% >> /var/lib/ceph/osd/ceph-10 >> /dev/sdc1 580496384 489451300 91045084 85% >> /var/lib/ceph/osd/ceph-6 >> /dev/sdg1 580496384 470559020 109937364 82% >> /var/lib/ceph/osd/ceph-9 >> /dev/sde1 580496384 490289388 90206996 85% >> /var/lib/ceph/osd/ceph-8 >> >> [root@ceph2 ~]# ceph df >> GLOBAL: >> SIZE AVAIL RAW USED %RAW USED >> 5256G 890G 4365G 83.06 >> POOLS: >> NAME ID USED %USED MAX AVAIL OBJECTS >> glance 6 1714G 32.61 385G 219579 >> cinder 7 676G 12.86 385G 97488 >> >> [root@ceph2 ~]# ceph osd pool get glance pg_num >> pg_num: 512 >> [root@ceph2 ~]# ceph osd pool get cinder pg_num >> pg_num: 128 >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com