Could you, please paste the output of pg 37.9c query pt., 21 wrz 2018 o 14:39 Olivier Bonvalet <ceph.l...@daevel.fr> napisał(a):
> In fact, one object (only one) seem to be blocked on the cache tier > (writeback). > > I tried to flush the cache with "rados -p cache-bkp-foo cache-flush- > evict-all", but it blocks on the object > "rbd_data.f66c92ae8944a.00000000000f2596". > > So I reduced (a lot) the cache tier to 200MB, "rados -p cache-bkp-foo > ls" now show only 3 objects : > > rbd_directory > rbd_data.f66c92ae8944a.00000000000f2596 > rbd_header.f66c92ae8944a > > And "cache-flush-evict-all" still hangs. > > I also switched the cache tier to "readproxy", to avoid using this > cache. But, it's still blocked. > > > > > Le vendredi 21 septembre 2018 à 02:14 +0200, Olivier Bonvalet a écrit : > > Hello, > > > > on a Luminous cluster, I have a PG incomplete and I can't find how to > > fix that. > > > > It's an EC pool (4+2) : > > > > pg 37.9c is incomplete, acting [32,50,59,1,0,75] (reducing pool > > bkp-sb-raid6 min_size from 4 may help; search ceph.com/docs for > > 'incomplete') > > > > Of course, we can't reduce min_size from 4. > > > > And the full state : https://pastebin.com/zrwu5X0w > > > > So, IO are blocked, we can't access thoses damaged data. > > OSD blocks too : > > osds 32,68,69 have stuck requests > 4194.3 sec > > > > OSD 32 is the primary of this PG. > > And OSD 68 and 69 are for cache tiering. > > > > Any idea how can I fix that ? > > > > Thanks, > > > > Olivier > > > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com