It won't kick off right away if other deep scrubs are going to those OSDs. You can set nodeep-scrub on the cluster, wait till your other deep scrubs have finished, then turn deep scrubs back on and immediately run the repair. You should see that pg do a deep scrubs then repair.
On Sat, May 18, 2019, 6:41 PM Jorge Garcia <jgar...@soe.ucsc.edu> wrote: > I have tried ceph pg repair several times. It claims "instructing pg > 2.798s0 on osd.41 to repair" but then nothing happens as far as I can tell. > Any way of knowing if it's doing more? > > On Sat, May 18, 2019 at 3:33 PM Brett Chancellor < > bchancel...@salesforce.com> wrote: > >> I would try the ceph pg repair. If you see the pg go into deep scrubbing, >> then back to inconsistent you probably have a bad drive. Find which of the >> drives in the pg are bad (pg query or go to the host and look through >> dmesg). Take that osd offline and mark it out. Once backfill is complete, >> it should clear up. >> >> On Sat, May 18, 2019, 6:05 PM Jorge Garcia <jgar...@soe.ucsc.edu> wrote: >> >>> We are testing a ceph cluster mostly using cephfs. We are using an >>> erasure-code pool, and have been loading it up with data. Recently, we got >>> a HEALTH_ERR response when we were querying the ceph status. We stopped all >>> activity to the filesystem, and waited to see if the error would go away. >>> It didn't. Then we tried a couple of suggestions from the internet (ceph pg >>> repair, ceph pg scrub, ceph pg deep-scrub) to no avail. I'm not sure how to >>> find out more information about what the problem is, and how to repair the >>> filesystem to bring it back to normal health. Any suggestions? >>> >>> Current status: >>> >>> # ceph -s >>> >>> cluster: >>> >>> id: 28ef32f1-4350-491b-9003-b19b9c3a2076 >>> >>> health: HEALTH_ERR >>> >>> 5 scrub errors >>> >>> Possible data damage: 1 pg inconsistent >>> >>> >>> >>> services: >>> >>> mon: 3 daemons, quorum gi-cba-01,gi-cba-02,gi-cba-03 >>> >>> mgr: gi-cba-01(active), standbys: gi-cba-02, gi-cba-03 >>> >>> mds: backups-1/1/1 up {0=gi-cbmd=up:active} >>> >>> osd: 87 osds: 87 up, 87 in >>> >>> >>> >>> data: >>> >>> pools: 2 pools, 4096 pgs >>> >>> objects: 90.98 M objects, 134 TiB >>> >>> usage: 210 TiB used, 845 TiB / 1.0 PiB avail >>> >>> pgs: 4088 active+clean >>> >>> 5 active+clean+scrubbing+deep >>> >>> 2 active+clean+scrubbing >>> >>> 1 active+clean+inconsistent >>> >>> # ceph health detail >>> >>> HEALTH_ERR 5 scrub errors; Possible data damage: 1 pg inconsistent >>> >>> OSD_SCRUB_ERRORS 5 scrub errors >>> >>> PG_DAMAGED Possible data damage: 1 pg inconsistent >>> >>> pg 2.798 is active+clean+inconsistent, acting [41,50,17,2,86,70,61] >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com