Sure. Here's a complete query dump of one of the 30 pgs: http://pastebin.com/NFSYTbUP
Brian On Wed, Jun 29, 2016 at 6:25 PM, Brad Hubbard <bhubb...@redhat.com> wrote: > On Thu, Jun 30, 2016 at 3:22 AM, Brian Felton <bjfel...@gmail.com> wrote: > > Greetings, > > > > I have a lab cluster running Hammer 0.94.6 and being used exclusively for > > object storage. The cluster consists of four servers running 60 6TB OSDs > > each. The main .rgw.buckets pool is using k=3 m=1 erasure coding and > > contains 8192 placement groups. > > > > Last week, one of our guys out-ed and removed one OSD from each of three > of > > the four servers in the cluster, which resulted in some general badness > (the > > disks were wiped post-removal, so the data are gone). After a proper > > education in why this is a Bad Thing, we got the OSDs added back. When > all > > was said and done, we had 30 pgs that were stuck incomplete, and no > amount > > of magic has been able to get them to recover. From reviewing the data, > we > > knew that all of these pgs contained at least 2 of the removed OSDs; I > > understand and accept that the data are gone, and that's not a concern > (yay > > lab). > > > > Here are the things I've tried: > > > > - Restarted all OSDs > > - Stopped all OSDs, removed all OSDs from the crush map, and started > > everything back up > > - Executed a 'ceph pg force_create_pg <id>' for each of the 30 stuck pgs > > - Executed a 'ceph pg send_pg_creates' to get the ball rolling on creates > > - Executed several 'ceph pg <id> query' commands to ensure we were > > referencing valid OSDs after the 'force_create_pg' > > - Ensured those OSDs were really removed (e.g. 'ceph auth del', 'ceph osd > > crush remove', and 'ceph osd rm') > > Can you share some of the pg query output? > > > > > At this point, I've got the same 30 pgs that are stuck creating. I've > run > > out of ideas for getting this back to a healthy state. In reviewing the > > other posts on the mailing list, the overwhelming solution was a bad OSD > in > > the crush map, but I'm all but certain that isn't what's hitting us here. > > Normally, being the lab, I'd consider nuking the .rgw.buckets pool and > > starting from scratch, but we've recently spent a lot of time pulling > 140TB > > of data into this cluster for some performance and recovery tests, and > I'd > > prefer not to have to start that process again. I am willing to > entertain > > most any other idea irrespective to how destructive it is to these PGs, > so > > long as I don't have to lose the rest of the data in the pool. > > > > Many thanks in advance for any assistance here. > > > > Brian Felton > > > > > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > > > -- > Cheers, > Brad >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com