We had this issue in Havana and applying the patch from that gist
fixed the issue for us. We just noticed last week that it has returned
in our upgraded Trusty/Icehouse cloud since we neglected to repatch
the new eventlet package, and patching *that* fixed it again for us.

I reported an issue against eventlet for this a while ago, but looks
like it has since been closed:

https://github.com/eventlet/eventlet/issues/63


Cheers,
Kieran

On 14 October 2014 02:17, Joe Topjian <j...@topjian.net> wrote:
> That's really interesting - thanks for the link.
>
> We've been able to narrow down why didn't see this with Cinder or Swift: in
> short, Swift isn't using Oslo (AFAIK) and we had some previous logging
> issues with Cinder in Havana so we altered the logging setup a bit.
>
>
> On Mon, Oct 13, 2014 at 3:05 AM, Francois Deppierraz
> <franc...@ctrlaltdel.ch> wrote:
>>
>> Hi Joe,
>>
>> Yes, same problem here running Ubuntu 14.04.
>>
>> The symptom is nova-api, nova-conductor and glance-api eating all CPU
>> without responding to API requests anymore.
>>
>> It is possible to reproduce it thanks to the following script.
>>
>> https://gist.github.com/dbishop/7a2e224f3aafea1a1fc3
>>
>> François
>>
>> On 11. 10. 14 00:40, Joe Topjian wrote:
>> > Hello,
>> >
>> > This morning we noticed various nova, glance, and keystone services (not
>> > cinder or swift) not working in two different clouds and required a
>> > restart.
>> >
>> > We thought it was a network issue since one of the only commonalities
>> > between the two clouds was that they are on the same network.
>> >
>> > Then later in the day I logged into a test cloud on a totally separate
>> > network and had the same problem.
>> >
>> > Looking at all three environments, the commonality is now that they have
>> > Ubuntu security updates automatically applied in the morning and this
>> > morning rsyslog was patched and restarted.
>> >
>> > I found this oslo bug that kind of sounds like the issue we saw:
>> >
>> > https://bugs.launchpad.net/oslo.log/+bug/1076466
>> >
>> > Doing further investigation, log files do indeed show a lack of entries
>> > for various services/daemons until they were restarted.
>> >
>> > Has anyone else run into this? Maybe even this morning, too? :)
>> >
>> > Thanks,
>> > Joe
>> >
>> >
>> >
>> >
>> > _______________________________________________
>> > OpenStack-operators mailing list
>> > OpenStack-operators@lists.openstack.org
>> > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>> >
>>
>>
>> _______________________________________________
>> OpenStack-operators mailing list
>> OpenStack-operators@lists.openstack.org
>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>
>
>
> _______________________________________________
> OpenStack-operators mailing list
> OpenStack-operators@lists.openstack.org
> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
>

_______________________________________________
OpenStack-operators mailing list
OpenStack-operators@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators

Reply via email to