When you say 250+, you mean 250+ host spread in lots of cluster, right? If I am not mistaken, ACS limits the number of KVM hosts in a cluster, something like 50? I do not remember now if that value can be configured, may it can be.
I recall to have read something in a Red Hat doc about the KVM that it does not have limit of hosts in a cluster. Actually, it does not seem to have the figure of cluster at all. That is created solely in ACS, to facilitate the management. To debug the problem, I would start with the following questions: Is every single cluster of your environment is presenting that problem? What is the size of physical hosts that you have in your environment? Do all of them have the same configuration? Do you know the load (resource allocated and used) that is being imposed in those hosts that had shown those problems? What is your over commitment/provisioning factor that you are using? On Wed, Nov 18, 2015 at 8:19 PM, Daan Hoogland <daan.hoogl...@gmail.com> wrote: > sounds like a bad limit Ilya, i'll keep an eye out. > > On Wed, Nov 18, 2015 at 10:10 PM, ilya <ilya.mailing.li...@gmail.com> > wrote: > > > I'm curious if anyone runs ACS with atleast 250+ KVM hosts. > > > > We've been noticing weird issues with KVM where occasionally lots of KVM > > agents get Nio connection closed issue followed by barrage of alerts. > > > > In some instances the agent reconnects right away and in other - it > > attempts to reconnect but never receives an ACK from MS. > > > > Please let me know if you notice anything like it and if you found a > > solution. > > > > Also, it would help to know what global settings have been tuned to make > > things work better (aside from direct.agent.*) and how MS are running. > > > > Thanks > > ilya > > > > > > -- > Daan > -- Rafael Weingärtner