btzq opened a new issue, #9336: URL: https://github.com/apache/cloudstack/issues/9336
<!-- Verify first that your issue/request is not already reported on GitHub. Also test if the latest release and main branch are affected too. Always add information AFTER of these HTML comments, but no need to delete the comments. --> ##### ISSUE TYPE <!-- Pick one below and delete the rest --> * Bug Report ##### COMPONENT NAME <!-- Categorize the issue, e.g. API, VR, VPN, UI, etc. --> ~~~ Autoscale ~~~ ##### CLOUDSTACK VERSION <!-- New line separated list of affected versions, commit ID for issues on main branch. --> ~~~ 4.19.0 ~~~ ##### CONFIGURATION <!-- Information about the configuration if relevant, e.g. basic network, advanced networking, etc. N/A otherwise --> ##### OS / ENVIRONMENT <!-- Information about the environment if relevant, N/A otherwise --> ##### SUMMARY <!-- Explain the problem/feature briefly --> ##### STEPS TO REPRODUCE <!-- For bugs, show exactly how to reproduce the problem, using a minimal test-case. Use Screenshots if accurate. For new features, show how the feature would be used. --> We are actively using Autoscale Groups which have the following scale down rules. Autoscale Rule - Name: ScaleDownPolicy-0 - Duration (in sec) = 60 Seconds - Quiet Time (in sec) = 30 seconds Conditions: - Counter: VM CPU - Average Percentage - Operator: Less Than - Threshold: 35 These rules work well on normal days. But today, we had a node failure, which the node was hosting some autoscaleVMs As a result, we experienced this issue, which have already been reported: - https://github.com/apache/cloudstack/issues/9145 In the case where the VMs were not 'Orphaned', and managed to find its way back to the Autoscale Group (or just werent affected), i noticed that the scale down rule did not work anymore. I had VMs that: - Min : 2 Members - Max : 6 Members - Available Instances: 6 _<- This should be 2 Instead._ And after 5 minutes, it still did not scale down. All VMs will running. Ive verified the CPU Utilisation of all VMs were only 1% consistently. In order to have the service resume, i had to: - Disable the Autoscale Group - Delete the VMs - Re-enable the Autoscale VM After that, the scaledown rule works, and i had the setup below again: I had VMs that: - Min : 2 Members - Max : 6 Members - Available Instances: 2 <!-- Paste example playbooks or commands between quotes below --> ~~~ NA ~~~ <!-- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!-- What did you expect to happen when running the steps above? --> ~~~ NA ~~~ ##### ACTUAL RESULTS <!-- What actually happened? --> <!-- Paste verbatim command output between quotes below --> ~~~ In the case where the VMs were not 'Orphaned', and managed to find its way back to the Autoscale Group (or just happened to be in a host that was not affected), i noticed that the scale down rule did not work anymore. ~~~ -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@cloudstack.apache.org.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org