Hi Josh,
> On Sep 16, 2017, at 3:13 AM, Josh Durgin wrote:
>
> (Sorry for top posting, this email client isn't great at editing)
Thanks for taking the time to respond. :)
> The mitigation strategy I mentioned before of forcing backfill could be
> backported to jewel, but I don't think it's a
> All the crashed OSDs had the same rocksdb corruption error? What kind >
> hardware (or vm?) are you using?
yes, All the crashed OSDs Has the some rocksdb corruption.
Our Cluster has 3 Nodes, and Per Node has
2 * Intel(R) Xeon(R) E5-2620 v3 @ 2.40GHz CPU
4 * 16G DDR4-2133 memory
10 * O
There is a ceph command "reweight-by-utilization" you can run to
adjust the OSD weights automatically based on their utilization:
http://docs.ceph.com/docs/master/rados/operations/control/#osd-subsystem
Some people run this on a periodic basis (cron script)
Check the mailing list archives, for exa