I don't see anything related to lost objects in your output. I just see waiting on backfill, backfill_toofull, remapped, and so forth. You can read a bit about what is going on here: http://ceph.com/docs/next/rados/operations/monitoring-osd-pg/
Keep us posted as to the recovery, and let me know what I can do to improve the docs for scenarios like this. On Sat, Apr 20, 2013 at 10:52 AM, Marco Aroldi <marco.aro...@gmail.com>wrote: > John, > thanks for the quick reply. > Below you can see my ceph osd tree > The problem is caused not by the failure itself, but by the "renamed" > bunch of devices. > It was like a deadly 15-puzzle > I think that the solution was to mount the devices in fstab using UUID > (/dev/disk/by-uuid) instead of /dev/sdX > > However, yes I have an entry in my ceph.conf (devs = /dev/sdX1 -- > osd_journal = /dev/sdX2) *and* an entry in my fstab for each OSD > > The node with failed disk is s103 (osd.59) > > Now i have 5 osd from s203 up and in to try to let ceph rebalance > data... but is still a bloody mess. > Look at ceph -w output: is reported a total of 110TB: is wrong... al > drives are 2TB and i have 49 drives up and in -- total 98Tb > I think that 110TB (55 osd) was the size before cluster became inaccessible > > # id weight type name up/down reweight > -1 130 root default > -9 65 room p1 > -3 44 rack r14 > -4 22 host s101 > 11 2 osd.11 up 1 > 12 2 osd.12 up 1 > 13 2 osd.13 up 1 > 14 2 osd.14 up 1 > 15 2 osd.15 up 1 > 16 2 osd.16 up 1 > 17 2 osd.17 up 1 > 18 2 osd.18 up 1 > 19 2 osd.19 up 1 > 20 2 osd.20 up 1 > 21 2 osd.21 up 1 > -6 22 host s102 > 33 2 osd.33 up 1 > 34 2 osd.34 up 1 > 35 2 osd.35 up 1 > 36 2 osd.36 up 1 > 37 2 osd.37 up 1 > 38 2 osd.38 up 1 > 39 2 osd.39 up 1 > 40 2 osd.40 up 1 > 41 2 osd.41 up 1 > 42 2 osd.42 up 1 > 43 2 osd.43 up 1 > -13 21 rack r10 > -12 21 host s103 > 55 2 osd.55 up 0 > 56 2 osd.56 up 0 > 57 2 osd.57 up 0 > 58 2 osd.58 up 0 > 59 2 osd.59 down 0 > 60 2 osd.60 down 0 > 61 2 osd.61 down 0 > 62 2 osd.62 up 0 > 63 2 osd.63 up 0 > 64 1.5 osd.64 up 0 > 65 1.5 osd.65 down 0 > -10 65 room p2 > -7 22 rack r20 > -5 22 host s202 > 22 2 osd.22 up 1 > 23 2 osd.23 up 1 > 24 2 osd.24 up 1 > 25 2 osd.25 up 1 > 26 2 osd.26 up 1 > 27 2 osd.27 up 1 > 28 2 osd.28 up 1 > 29 2 osd.29 up 1 > 30 2 osd.30 up 1 > 31 2 osd.31 up 1 > 32 2 osd.32 up 1 > -8 22 rack r22 > -2 22 host s201 > 0 2 osd.0 up 1 > 1 2 osd.1 up 1 > 2 2 osd.2 up 1 > 3 2 osd.3 up 1 > 4 2 osd.4 up 1 > 5 2 osd.5 up 1 > 6 2 osd.6 up 1 > 7 2 osd.7 up 1 > 8 2 osd.8 up 1 > 9 2 osd.9 up 1 > 10 2 osd.10 up 1 > -14 21 rack r21 > -11 21 host s203 > 44 2 osd.44 up 1 > 45 2 osd.45 up 1 > 46 2 osd.46 up 1 > 47 2 osd.47 up 1 > 48 2 osd.48 up 1 > 49 2 osd.49 up 0 > 50 2 osd.50 up 0 > 51 2 osd.51 up 0 > 52 1.5 osd.52 up 0 > 53 1.5 osd.53 up 0 > 54 2 osd.54 up 0 > > > ceph -w > > 2013-04-20 19:46:48.608988 mon.0 [INF] pgmap v1352767: 17280 pgs: 58 > active, 12581 active+clean, 1686 active+remapped+wait_backfill, 24 > active+degraded+wait_backfill, 224 > active+remapped+wait_backfill+backfill_toofull, 1061 > active+recovery_wait, 4 > active+degraded+wait_backfill+backfill_toofull, 629 peering, 626 > active+remapped, 72 active+remapped+backfilling, 89 active+degraded, > 14 active+remapped+backfill_toofull, 1 active+clean+scrubbing, 8 > active+degraded+remapped+wait_backfill, 20 > active+recovery_wait+remapped, 5 > active+degraded+remapped+wait_backfill+backfill_toofull, 162 > remapped+peering, 1 active+degraded+remapped+backfilling, 2 > active+degraded+remapped+backfill_toofull, 13 active+recovering; 49777 > GB data, 72863 GB used, 40568 GB / 110 TB avail; 2965687/21848501 > degraded (13.574%); recovering 5 o/s, 16363B/s > > 2013/4/20 John Wilkins <john.wilk...@inktank.com>: > > Marco, > > > > If you do a "ceph tree" can you see if your OSDs are all up? You seem to > > have at least one problem related to the backfill OSDs being too full, > and > > some which are near full or full for the purposes of storage. See the > > following in the documentation to see if this helps: > > > > > http://ceph.com/docs/master/rados/configuration/mon-config-ref/#storage-capacity > > > http://ceph.com/docs/master/rados/configuration/osd-config-ref/#backfilling > > > http://ceph.com/docs/master/rados/operations/troubleshooting-osd/#no-free-drive-space > > > > Before you start deleting data as a remedy, you'd want to at least try to > > get the OSDs back up and running first. > > > > If rebooting changed the drive names, you might look here: > > > http://ceph.com/docs/master/rados/configuration/osd-config-ref/#general-settings > > > > We have default settings for OSD and journal paths, which you could > override > > if you can locate the data and journal sources on the renamed drives. If > you > > mounted them, but didn't add them to the fstab, that might be the source > of > > the problem. I'd rather see you use the default paths, as it would be > easier > > to troubleshoot later. So did you mount the drives, but not add the mount > > points to fstab? > > > > John > > > > > > > > > > On Sat, Apr 20, 2013 at 8:46 AM, Marco Aroldi <marco.aro...@gmail.com> > > wrote: > >> > >> Hi, > >> due a harware failure during expanding ceph, I'm in big trouble > >> because the cephfs doesn't mount anymore. > >> I was adding a couple storage nodes, but a disk has failed and after a > >> reboot the OS (ubuntu 12.04) renamed the remaining devices, so the > >> entire node has been screwed out. > >> > >> Now, from the "sane new node", I'm taking some new osd up and in > >> because the cluster is near full and I can't revert completely the > >> situation as before > >> > >> *I can* afford data loss, but i need to regain access to the filesystem > >> > >> My setup: > >> 3 mon + 3 mds > >> 4 storage nodes (i was adding no. 5 and 6) > >> > >> Ceph 0.56.4 > >> > >> > >> ceph health: > >> HEALTH_ERR 2008 pgs backfill; 246 pgs backfill_toofull; 74 pgs > >> backfilling; 134 pgs degraded; 790 pgs peering; 10 pgs recovering; > >> 1116 pgs recovery_wait; 790 pgs stuck inactive; 4782 pgs stuck > >> unclean; recovery 3049459/21926624 degraded (13.908%); recovering 6 > >> o/s, 16316KB/s; 4 full osd(s); 30 near full osd(s); full,noup,nodown > >> flag(s) set > >> > >> > >> > >> ceph mds dump: > >> dumped mdsmap epoch 44 > >> epoch 44 > >> flags 0 > >> created 2013-03-18 14:42:29.330548 > >> modified 2013-04-20 17:14:32.969332 > >> tableserver 0 > >> root 0 > >> session_timeout 60 > >> session_autoclose 300 > >> last_failure 43 > >> last_failure_osd_epoch 18160 > >> compat compat={},rocompat={},incompat={1=base v0.20,2=client > >> writeable ranges,3=default file layouts on dirs,4=dir inode in > >> separate object} > >> max_mds 1 > >> in 0 > >> up {0=6376} > >> failed > >> stopped > >> data_pools [0] > >> metadata_pool 1 > >> 6376: 192.168.21.11:6800/13457 'm1' mds.0.9 up:replay seq 1 > >> 5945: 192.168.21.13:6800/12999 'm3' mds.-1.0 up:standby seq 1 > >> 5963: 192.168.21.12:6800/22454 'm2' mds.-1.0 up:standby seq 1 > >> > >> > >> > >> ceph mon dump: > >> epoch 1 > >> fsid d634f7b3-8a8a-4893-bdfb-a95ccca7fddd > >> last_changed 2013-03-18 14:39:42.253923 > >> created 2013-03-18 14:39:42.253923 > >> 0: 192.168.21.11:6789/0 mon.m1 > >> 1: 192.168.21.12:6789/0 mon.m2 > >> 2: 192.168.21.13:6789/0 mon.m3 > >> _______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > > > > > -- > > John Wilkins > > Senior Technical Writer > > Intank > > john.wilk...@inktank.com > > (415) 425-9599 > > http://inktank.com > -- John Wilkins Senior Technical Writer Intank john.wilk...@inktank.com (415) 425-9599 http://inktank.com
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com