On Thu, May 4, 2017 at 8:40 AM Osama Hasebou <osama.hase...@csc.fi> wrote:
> Hi Everyone, > > We keep running into stalled IOs / they also drop almost to zero, whenever > a node suddenly would go down or if there was a large amount of rebalancing > going on and once rebalancing is completed, we would also get stalled io > for 2-10 mins. > > Has anyone seen this behaviour before and found a way to fix this? We are > seeing this on Ceph Hammer and also on Jewel. > Please check the setting "mon osd down out subtree limit". Setting to host would prevent automatic marking OSDs down when a host fails. Also osd-recovery-max-active (my setting is 5) osd-recovery-threads (my setting is 3) osd-max-backfills (my setting is 5) Hth, Alex > > Thanks. > > Regards, > Ossi > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > -- -- Alex Gorbachev Storcium
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com