> 在 2021年10月4日,04:18,Michael Thomas <w...@caltech.edu> 写道:
> 
> On 10/3/21 12:08, 胡 玮文 wrote:
>>>> 在 2021年10月4日,00:53,Michael Thomas <w...@caltech.edu> 写道:
>>> 
>>> I recently started getting inconsistent PGs in my Octopus (15.2.14) ceph 
>>> cluster.  I was able to determine that they are all coming from the same 
>>> OSD: osd.143.  This host recently suffered from an unplanned power loss, so 
>>> I'm not surprised that there may be some corruption.  This PG is part of a 
>>> EC 8+2 pool.
>>> 
>>> The OSD logs from the PG's primary OSD show this and similar errors from 
>>> the PG's most recent deep scrub:
>>> 
>>> 2021-10-03T03:25:25.969-0500 7f6e6801f700 -1 log_channel(cluster) log [ERR] 
>>> : 23.1fa shard 143(1) soid 23:5f8c3d4e:::10000179969.00000168:head : 
>>> candidate had a read error
>>> 
>>> In attempting to fix it, I first ran 'ceph pg repair 23.1fa' on the PG. 
>>> This accomplished nothing.  Next I ran a shallow fsck on the OSD:
>> I expect this ‘ceph pg repair’ command could handle this kind of errors. 
>> After issuing this command, the pg should enter a state like 
>> “active+clean+scrubbing+deep+inconsistent+repair”, then you wait for the 
>> repair to finish (this can take hours), and you should be able to recover 
>> from the inconsistent state. What do you mean by “This accomplished nothing”?
> 
> The PG never entered the 'repair' state, nor did anything appear in the 
> primary OSD logs about a request for repair.  After more than 24 hours, the 
> PG remained listed as 'inconsistent'.
> 
> --Mike

I have encountered a similar situation. My case is the pg being repaired cannot 
get all the scrub reservations to enter the scrubbing state. Could you try 
“ceph tell osd.<primary OSD ID> dump_scrubs”, and see whether 23.1fa is listed 
and has forced == true? If so, this may also be your case. I think you may wait 
even longer, raise “osd_max_scrubs” config, or try set then unset noscrub to 
interrupt the running scrubs.

> _______________________________________________
> ceph-users mailing list -- ceph-users@ceph.io
> To unsubscribe send an email to ceph-users-le...@ceph.io
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to