Hi Martyn,

Le 22/08/2013 13:23, Martyn Taylor a écrit :
Hi Sylvain,

We are currently working on design docs. We'll be adding some architecture diagrams and description to our documentation soon.

Nice to know, thanks.

To answer your question re: provisioning and images.

In our currently implementation (which is very early days), we took images that were built from by Triple O, namely the overcloud non-compute and compute images and we use these directly. in the demo environment you seen in the video, we used Triple O CI to set up the machine, register the relevant overcloud images with glance and so on.

In Tuskar, we lifted from TripleO a copy of the triple-o overcloud heat template and made some modifications. We split out the non-compute and compute sections, this allows us to add multiple entries of each of the non-compute and compute sections (based on what is registered in Tuskar) we then add a section to enforce deployment of the particular images onto particular bare metal machines. (This allows us to match hardware to OpenStack services). We do this by using the force_hosts capability in the nova bare metal driver: https://blueprints.launchpad.net/nova/+spec/baremetal-force-node.

We also add some extra commands to the Heat template to registers flavors and associates the flavors, host aggregates and baremetal nodes in the overcloud nova control instance. This allows us to tell the nova scheduler to match any instance requests with flavors that were registered with a resource class in Tuskar with particular hardware that has also been added to that resource class.


Thanks for the explanation, I'm understanding more. So, basically, *and I understand this is a POC*, your API allows to dynamically build yaml templates for injecting Openstack components thanks to os-apply-config and os-refresh-config, packed in some images built by disk-image-builder ?

As Tomas mentioned, our initial release is really just a Proof of Concept. We'll be working to add more complex features and probably rework much of our "short cuts". Our aim though is to contribute as much as possible (or as much that makes sense) of Tuskar upstream into TripleO or any other component that we utilize and extend and have Tuskar really concentrate on how to utilize existing components to manage and deploy an OpenStack at large scale.


Totally understand. So, my next question is : do you have kind of a roadmap, which could tell us when you feel confident for releasing something ready for lab usage ? =)

Thanks,
-Sylvain


Regards
Martyn


On 21/08/13 16:15, Sylvain Bauza wrote:
Hi Tomas,

Are there any design docs which could explain how you provision the baremetal hosts ? As far as I can see, it seeems you're relying on TripleO heat templates, right ?

Are you then using disk-image-builder ?

Thanks,
-Sylvain

PS : I just looked at the Youtube demo https://www.youtube.com/watch?v=VEY035-Lyzo


Le 21/08/2013 14:32, Tomas Sedovic a écrit :
Hi everyone,

We would like to announce Tuskar, an OpenStack management service.

Our goal is to provide an API and UI to install and manage OpenStack at larger scale: where you deal with racks, different hardware classes for different purposes (storage, memory vs. cpu-intensive compute), the burn-in process, monitoring the HW utilisation, etc.

Some of this will overlap with TripleO, Ceilometer and possibly other projects. In that case, we will work with the projects to figure out the best place to fix rather than duplicating effort and playing in our own sandbox.


Current status:

There's a saying that if you're not embarrassed by your first release, you've shipped too late.

I'm happy to say, we are quite embarrassed :-)

We've got a prototype that allows us to define different hardware classes and provision the racks with the appropriate images, then add new racks and have them provisioned.

We've got a Horizon dashboard plugin that shows the general direction we want to follow and we're looking into integrating Ceilometer metrics and alarms.

However, we're still tossing around different ideas and things are very likely to change.

Our repositories are on Stackforge:

https://github.com/stackforge/tuskar
https://github.com/stackforge/python-tuskarclient
https://github.com/stackforge/tuskar-ui

And we're using Launchpad to manage our bugs and blueprints:

https://launchpad.net/tuskar
https://launchpad.net/tuskar-ui

If you want to talk to us, pop in the #tuskar IRC channel on Freenode or send an email to openstack-...@lists.launchpad.net with "[Tuskar]" in the subject.


PTL:

Talking to OpenStack developers, we were advised to elect the PTL early.

Since we're nearing the end of the Havana cycle, we'll elect the PTL for a slightly longer term -- the rest of Havana and throughout Icehouse. The next election will coincide with those of the official OpenStack projects.

If you are a Tuskar developer and want to nominate yourself, please send an email to openstack-...@lists.launchpad.net with subject "Tuskar PTL candidacy".

The self-nomination period will end on Monday, 26th August 2013, 23:59 UTC.


--
Tomas Sedovic

_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev



_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev


_______________________________________________
OpenStack-dev mailing list
OpenStack-dev@lists.openstack.org
http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-dev

Reply via email to