Ceph cluster with 60 OSDs, Giant 0.87.2. One of the OSDs failed due to a hardware error, however after normal recovery it seems stuck with one active+undersized+degraded+inconsistent pg.
I haven't been able to get repair to happen using "ceph pg repair 12.28a"; I can see the activity logged in the mon logs, however the repair doesn't actually seem to happen in any of the actual osd logs. I tried folowing Sebiastien's instructions for manually locating the inconsistent object ( http://www.sebastien-han.fr/blog/2015/04/27/ceph-manually-repair-object/), however the md5sum from the objects both match, so I'm not quite sure how to proceed. Any ideas on how to return to a healthy cluster? [root@soi-ceph2 ceph]# ceph status cluster 6cc00165-4956-4947-8605-53ba51acd42b health HEALTH_ERR 1023 pgs degraded; 1 pgs inconsistent; 1023 pgs stuck degraded; 1099 pgs stuck unclean; 1023 pgs stuck undersized; 1023 pgs undersized; recovery 132091/23742762 objects degraded (0.556%); 7745/23742762 objects misplaced (0.033%); 1 scrub errors monmap e5: 3 mons at {soi-ceph1= 10.2.2.11:6789/0,soi-ceph2=10.2.2.12:6789/0,soi-ceph3=10.2.2.13:6789/0}, election epoch 4132, quorum 0,1,2 soi-ceph1,soi-ceph2,soi-ceph3 osdmap e41120: 60 osds: 59 up, 59 in pgmap v37432002: 61440 pgs, 15 pools, 30513 GB data, 7728 kobjects 91295 GB used, 73500 GB / 160 TB avail 132091/23742762 objects degraded (0.556%); 7745/23742762 objects misplaced (0.033%) 60341 active+clean 76 active+remapped 1022 active+undersized+degraded 1 active+undersized+degraded+inconsistent client io 44548 B/s rd, 19591 kB/s wr, 1095 op/s [root@soi-ceph2 ceph]# ceph health detail | grep inconsistent pg 12.28a is stuck unclean for 126274.215835, current state active+undersized+degraded+inconsistent, last acting [36,52] pg 12.28a is stuck undersized for 3499.099747, current state active+undersized+degraded+inconsistent, last acting [36,52] pg 12.28a is stuck degraded for 3499.107051, current state active+undersized+degraded+inconsistent, last acting [36,52] pg 12.28a is active+undersized+degraded+inconsistent, acting [36,52] [root@soi-ceph2 ceph]# zgrep 'ERR' *.gz ceph-osd.36.log-20160325.gz:2016-03-24 12:00:43.568221 7fe7b2897700 -1 log_channel(default) log [ERR] : 12.28a shard 20: soid c5cf428a/default.64340.11__shadow_.VR0pEp1Nea8buLSqa9TGhLFZQ6co3KO_106/head//12 candidate had a read error, digest 2029411064 != known digest 2692480864 ceph-osd.36.log-20160325.gz:2016-03-24 12:01:25.970413 7fe7b2897700 -1 log_channel(default) log [ERR] : 12.28a deep-scrub 0 missing, 1 inconsistent objects ceph-osd.36.log-20160325.gz:2016-03-24 12:01:25.970423 7fe7b2897700 -1 log_channel(default) log [ERR] : 12.28a deep-scrub 1 errors [root@soi-ceph2 ceph]# md5sum /var/lib/ceph/osd/ceph-36/current/12.28a_head/DIR_A/DIR_8/DIR_2/DIR_4/default.64340.11\\u\\ushadow\\u.VR0pEp1Nea8buLSqa9TGhLFZQ6co3KO\\u106__head_C5CF428A__c \fb57b1f17421377bf2c35809f395e9b9 /var/lib/ceph/osd/ceph-36/current/12.28a_head/DIR_A/DIR_8/DIR_2/DIR_4/default.64340.11\\u\\ushadow\\u.VR0pEp1Nea8buLSqa9TGhLFZQ6co3KO\\u106__head_C5CF428A__c [root@soi-ceph3 ceph]# md5sum /var/lib/ceph/osd/ceph-52/current/12.28a_head/DIR_A/DIR_8/DIR_2/DIR_4/default.64340.11\\u\\ushadow\\u.VR0pEp1Nea8buLSqa9TGhLFZQ6co3KO\\u106__head_C5CF428A__c \fb57b1f17421377bf2c35809f395e9b9 /var/lib/ceph/osd/ceph-52/current/12.28a_head/DIR_A/DIR_8/DIR_2/DIR_4/default.64340.11\\u\\ushadow\\u.VR0pEp1Nea8buLSqa9TGhLFZQ6co3KO\\u106__head_C5CF428A__c
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com