On Wed, Aug 29, 2012 at 10:52 AM, Ryan Lane <rl...@wikimedia.org> wrote:
>> There was talk of trying to set up test infrastructure that would roll out 
>> Essex and then upgrade it to Folsom in some automated fashion so we could 
>> start learning where it breaks. Was there any forward momentum on that?
>>
>
> This would be awesome. Wrapping automated tests around upgrades would
> greatly improve the situation. Most of the issues that ops runs into
> during upgrades are unexpected changes, which are the same things that
> will likely be hit when testing upgrades in an automated way.

It would be fascinating (for me at least :)) to know the upgrade
process you use - how many stages you use, do you have multiple
regions and use one/some as canaries? Does the downtime required to do
an upgrade affect you? Do you run skewed versions (e.g. folsom nova,
essex glance) or do you do lock-step upgrades of all the components?

For Launchpad we've been moving more and more to a model of permitting
temporary skew so that we can do rolling upgrades of the component
services. That seems in-principle doable here - and could make it
easier to smoothly transition between versions, at the cost of a
(small) amount of attention to detail while writing changes to the
various apis.

-Rob

_______________________________________________
Mailing list: https://launchpad.net/~openstack
Post to     : openstack@lists.launchpad.net
Unsubscribe : https://launchpad.net/~openstack
More help   : https://help.launchpad.net/ListHelp

Reply via email to