Once you dropped the database, follow the installation instructions on
docs.openstack.org and create a new database for neutron, grant the
privileges and sync it. Then restart your neutron-server service and
recreate the provider and tenant networks.

You don't have to recreate the user, service and endpoint - they are not
saved in the neutron database but in keystone DB - which you hopefully didn't 
drop.

Am 29.06.2015 um 18:32 schrieb Yngvi Páll Þorfinnsson:
> Hi Uwe
> Is there something more to do than
> 
> drop database neutron;
> 
> will all information be deleted ?
> 
> I get this error when doing this task
> 
> root@controller2:/# keystone user-create --name neutron --pass netpass
> Conflict occurred attempting to store user - Duplicate Entry (HTTP 409)
> 
> best regards
> Yngvi
> 
> -----Original Message-----
> From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] 
> Sent: 29. júní 2015 16:11
> To: Yngvi Páll Þorfinnsson; Andreas Scheuring
> Cc: openstack@lists.openstack.org
> Subject: Re: [Openstack] error creating instance
> 
> If this is just a test setup it might be worth to drop and recreate the 
> neutron database, then recreate external and tenant networks. This way you 
> could get rid of any left-overs from before your switch.
> 
> 
> 
> Am 29.06.2015 um 18:06 schrieb Yngvi Páll Þorfinnsson:
>> Yes it's strange, this error is now reported
>>
>> 2015-06-29 16:03:38.779 5328 DEBUG 
>> neutron.plugins.ml2.drivers.mech_openvswitch 
>> [req-60dac17f-5280-4e0a-988f-ee763df18632 None] Checking segment: 
>> {'segmentation_id': 1102L, 'physical_network': u'external', 'id': 
>> u'cf6489c4-7ed6-43dc-85aa-f4b8c6b501ca', 'network_type': u'vlan'} for 
>> mappings: {} with tunnel_types: [u'gre'] check_segment_for_agent 
>> /usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_open
>> vswitch.py:52
>> root@controller2:/#
>>
>> but I can't see this network with id
>>
>> " cf6489c4-7ed6-43dc-85aa-f4b8c6b501ca"
>>
>>
>> Best regards
>> Yngvi
>>
>> -----Original Message-----
>> From: Andreas Scheuring [mailto:scheu...@linux.vnet.ibm.com]
>> Sent: 29. júní 2015 15:25
>> To: Yngvi Páll Þorfinnsson
>> Cc: uwe.sauter...@gmail.com; openstack@lists.openstack.org
>> Subject: Re: [Openstack] error creating instance
>>
>>
>>
>> Attempting to bind port 2bf4a49b-2ad6-4ead-a656-65814ad0724e on 
>> network 7a344656-815c-4116-b697-b52f9fdc6e4c
>> bind_port 
>> /usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_agen
>> t.py:57
>> 2015-06-29 14:28:55.924 5328 DEBUG
>> neutron.plugins.ml2.drivers.mech_agent
>> [req-9fe66e60-1a70-4ad6-b21e-ef91aca8a931 None] Checking agent:
>> {'binary': u'neutron-openvswitch-agent', 'description': None,
>> 'admin_state_up': True, 'heartbeat_timestamp': datetime.datetime(2015, 6, 
>> 29, 14, 28, 45), 'alive': True, 'id':
>> u'1c06fb08-105c-4659-ae0e-4a905931311e', 'topic': u'N/A', 'host':
>> u'compute5', 'agent_type': u'Open vSwitch agent', 'started_at':
>> datetime.datetime(2015, 6, 29, 14, 27, 45), 'created_at':
>> datetime.datetime(2015, 6, 26, 14, 51, 14), 'configurations':
>> {u'arp_responder_enabled': False, u'tunneling_ip': u'172.22.15.17',
>> u'devices': 0, u'l2_population': False, u'tunnel_types': [u'gre'],
>> u'enable_distributed_routing': False, u'bridge_mappings': {}}} 
>> bind_port 
>> /usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_agen
>> t.py:65
>> 2015-06-29 14:28:55.925 5328 DEBUG
>> neutron.plugins.ml2.drivers.mech_openvswitch
>> [req-9fe66e60-1a70-4ad6-b21e-ef91aca8a931 None] Checking segment:
>> {'segmentation_id': 1102L, 'physical_network': u'external', 'id':
>> u'cf6489c4-7ed6-43dc-85aa-f4b8c6b501ca', 'network_type': u'vlan'} for
>> mappings: {} with tunnel_types: [u'gre'] check_segment_for_agent 
>> /usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_open
>> vswitch.py:52
>>
>>
>>
>> ===
>> Checking segment: {'segmentation_id': 1102L, 'physical_network':u'external', 
>> 'id':
>> u'cf6489c4-7ed6-43dc-85aa-f4b8c6b501ca', 'network_type': u'vlan'} for 
>> mappings: {} with tunnel_types: [u'gre']
>>
>> This looks strange: Seems like your tenant network has a physical_network of 
>> type vlan assigned. That shouldn't be the case.
>>
>> Could you please provide the following information:
>>
>> Information of all Openstack networks available:
>>
>>> neutron net-list
>>
>>> neutron net-show <uuid>
>>
>> Especially of this one:
>>> neturon net-show cf6489c4-7ed6-43dc-85aa-f4b8c6b501ca
>>
>>
>> Usually your network should look like this (in this case vxlan):
>>
>> +---------------------------+--------------------------------------+
>> | Field                     | Value                                |
>> +---------------------------+--------------------------------------+
>> | admin_state_up            | True                                 |
>> | id                        | ef6552a5-be39-4bcc-9dde-2a200eaca64d |
>> | mtu                       | 0                                    |
>> | name                      | private                              |
>> | provider:network_type     | vxlan                                |
>> | provider:physical_network |                                      |
>> | provider:segmentation_id  | 1001                                 |
>> | router:external           | False                                |
>> | shared                    | False                                |
>> | status                    | ACTIVE                               |
>> | subnets                   | 4b539feb-b104-4f69-83ba-76f746a2c592 |
>> |                           | ac255618-afe9-4aea-b86d-b662b68e9d9d |
>> | tenant_id                 | 3c4ddcff52a74f2b97b71392300aa74d     |
>> +---------------------------+--------------------------------------+
>>
>> How did you create yours? via the UI? Or are you attaching your instance to 
>> the external network instead? In any case you need to attach it to your 
>> tenant network!! If it's not visible via the UI, maybe you have to switch to 
>> another tenant to get it.
>>
>> Hope we're close to finding the issue ;)
>>
>>
>> Andreas
>>
>>
>> On Mo, 2015-06-29 at 14:33 +0000, Yngvi Páll Þorfinnsson wrote:
>>> OK, I've enabled
>>> Debug=True
>>>
>>> I did try to create an instance, I've attached the server.log file 
>>> for the neutron server
>>>
>>> Best regards
>>> Yngvi
>>>
>>>
>>> -----Original Message-----
>>> From: Andreas Scheuring [mailto:scheu...@linux.vnet.ibm.com]
>>> Sent: 29. júní 2015 14:02
>>> To: Yngvi Páll Þorfinnsson
>>> Cc: uwe.sauter...@gmail.com; openstack@lists.openstack.org
>>> Subject: Re: [Openstack] error creating instance
>>>
>>> Correct, as the error messages indicated, get rid of the bridgemapping - 
>>> fine. Your agent is running now without error messages.
>>>
>>> Could you please enable debug=true in your neutron.conf, in order to also 
>>> get the debug logs (restart neutron server and ovs-agent).
>>>
>>> Can you trigger the start of another instance, so that an error message 
>>> appears?
>>>
>>> Andreas
>>>
>>>
>>>
>>>
>>> On Mo, 2015-06-29 at 13:45 +0000, Yngvi Páll Þorfinnsson wrote:
>>>> Hi Andreas,
>>>>
>>>> OK, so I'm providing the correct log file now i.e.
>>>> So, I'm not putting the line
>>>> #bridge_mappings = external:br-ex
>>>> In the conf file
>>>> /etc/neutron/plugins/ml2/ml2_conf.ini
>>>> On computer host for now.
>>>>
>>>> Best regards
>>>> Yngvi
>>>>
>>>> -----Original Message-----
>>>> From: Andreas Scheuring [mailto:scheu...@linux.vnet.ibm.com]
>>>> Sent: 29. júní 2015 13:07
>>>> To: uwe.sauter...@gmail.com
>>>> Cc: openstack@lists.openstack.org
>>>> Subject: Re: [Openstack] error creating instance
>>>>
>>>> Uwe,
>>>> along the configuration files Yngvi is using gre networking. 
>>>> So there should no bridge mapping be required at all for spawning an 
>>>> instance, right?
>>>>
>>>> The vlan tagging is done via a vlan device on the bond. So in fact it's a 
>>>> static configured vlan. Openstack does gre tunneling on this vlan.
>>>>
>>>>
>>>> Yngvi: on the compute node, there should be a 
>>>> "neutron-openvswtich-agent" log file. You sent over the log file of 
>>>> openvswitch itself. But what is required is the logfile of the 
>>>> openstack neutron agent managing openvswitch
>>>>
>>>> Andreas
>>>>
>>>>
>>>>
>>>> On Mo, 2015-06-29 at 13:40 +0200, Uwe Sauter wrote:
>>>>> Hi,
>>>>>
>>>>> I ran into a similar problem. Make sure that you include the
>>>>>
>>>>> [ml2_type_vlan]
>>>>> tenant_network_type = vlan
>>>>> network_vlan_ranges = physnet1:1501:1509
>>>>>
>>>>>
>>>>> on your controller network (as the controller needs this info to 
>>>>> make a proper decission on which VLAN IDs are available for tenant 
>>>>> networks).
>>>>>
>>>>> On other trick that isn't mentioned in any documentation is:
>>>>>
>>>>> Instead of symlinking  /etc/neutron/plugins.ini -> 
>>>>> /etc/neutron/plugins/ml2/ml2_conf.ini do it the other way around:
>>>>> Keep your neutron plugin configuration centralized in 
>>>>> /etc/neutron/plugin.ini and create one or two symlinks:
>>>>>
>>>>> /etc/neutron/plugins/ml2/ml2_conf.ini -> /etc/neutron/plugins.ini 
>>>>> (on each host running a neutron service) 
>>>>> /etc/neutron/plugins/openvswitch/ovs_neutron_plugin.ini -> 
>>>>> /etc/neutron/plugins.ini (on each host running
>>>>> neutron-openvswitch-agent)
>>>>>
>>>>> Just make sure that /etc/neutron/plugins.ini keeps both ML2 and 
>>>>> OpenVSwitch config options.
>>>>>
>>>>>
>>>>> The whole trick works because  Neutron services are started using 
>>>>> the option "--config <file>". And it will only read the files mentioned 
>>>>> on the commandline / init script / systemd unit file.
>>>>>
>>>>>
>>>>> Regards,
>>>>>
>>>>>   Uwe
>>>>>
>>>>>
>>>>>
>>>>> Am 29.06.2015 um 13:01 schrieb Yngvi Páll Þorfinnsson:
>>>>>> Hi
>>>>>>
>>>>>> I should add to this information, I'm using VLANs on compute, 
>>>>>> network and swift nodes.
>>>>>>
>>>>>> But not on the controller node.
>>>>>> I'm not sure if that causes problems ?
>>>>>>
>>>>>> Best regards
>>>>>> Yngvi
>>>>>>
>>>>>>
>>>>>> Hi Andreas
>>>>>>
>>>>>> I'v attached those files from a network node:
>>>>>>
>>>>>> Ml2_conf.ini
>>>>>> Neutron.conf
>>>>>> L3_agent.ini
>>>>>> Nova-compute.log
>>>>>>
>>>>>> I've setup 3 vlans on one interface and configured bonding It works and 
>>>>>> I have good connection between the servers:
>>>>>>
>>>>>> root@network2:/# cat /proc/net/vlan/config
>>>>>> VLAN Dev name    | VLAN ID
>>>>>> Name-Type: VLAN_NAME_TYPE_RAW_PLUS_VID_NO_PAD
>>>>>> bond0.48       | 48  | bond0
>>>>>> bond0.47       | 47  | bond0
>>>>>> bond0.45       | 45  | bond0
>>>>>>
>>>>>> Tunnel network   -> bond0.47
>>>>>> Mgmt network             -> bond0.48
>>>>>> Extrenal network         -> bond0.45
>>>>>>
>>>>>> And when I check for bridges and ports on the network node:
>>>>>>
>>>>>> root@network2:/# ovs-vsctl list-br br-ex br-int br-tun
>>>>>>
>>>>>> root@network2:/# ovs-vsctl list-ports br-ex
>>>>>> bond0.45
>>>>>> phy-br-ex
>>>>>> qg-eb22e091-62
>>>>>>
>>>>>> best regards
>>>>>> Yngvi
>>>>>>
>>>>>>
>>>>>> -----Original Message-----
>>>>>> From: Andreas Scheuring [mailto:scheu...@linux.vnet.ibm.com]
>>>>>> Sent: 29. júní 2015 06:59
>>>>>> To: Yngvi Páll Þorfinnsson
>>>>>> Cc: openstack@lists.openstack.org
>>>>>> Subject: Re: [Openstack] error creating instance
>>>>>>
>>>>>> The issue seems to be related to a neutron missconfiguration. 
>>>>>> --> Unexpected vif_type=binding_failed
>>>>>>
>>>>>> Please have a look at your neutron server config file in the 
>>>>>> network
>>>>>> node(s) and the l2 agent config files (ovs?). You should find additional 
>>>>>> information there.
>>>>>>
>>>>>>
>>>>>> If this doesn't help, please provide these log files, the neutron config 
>>>>>> files and a brief description of how your nodes network are set up.
>>>>>>
>>>>>>
>>>>>> Andreas
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fr, 2015-06-26 at 15:05 +0000, Yngvi Páll Þorfinnsson wrote:
>>>>>>> Hi
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> Can someone please help on this matter?
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> I have installed and configured OpenStack(Juno) environment on 
>>>>>>> Ubuntu(14.04).
>>>>>>>
>>>>>>> We have currently 1 controller node, 3 network nodes and 4 
>>>>>>> compute nodes as well as 4 swift nodes.
>>>>>>>
>>>>>>> I'm using OpenStack Networking (neutron).  
>>>>>>>
>>>>>>> I‘v recently introduced VLANs for tunnel-, mgmt- and external 
>>>>>>> networks.
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> When I try to create a instance, with this cmd:
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> nova boot --flavor m1.tiny --image cirros-0.3.3-x86_64 --nic 
>>>>>>> net-id=7a344656-815c-4116-b697-b52f9fdc6e4c --security-group 
>>>>>>> default --key-name demo-key demo-instance3
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> it fails.  The status from nova list is :
>>>>>>>
>>>>>>> root@controller2:/# nova list
>>>>>>>
>>>>>>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>>>>>>>
>>>>>>> | ID                                   | Name            | Status |
>>>>>>> Task State | Power State | Networks              |
>>>>>>>
>>>>>>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>>>>>>>
>>>>>>> | ca662fc0-2417-4da1-be2c-d6ccf90ed732 | demo-instance22 | ERROR  | -
>>>>>>> | NOSTATE     |                       |
>>>>>>>
>>>>>>> | 17d26ca3-f56c-4a87-ae0a-acfafea4838c | demo-instance30 | ERROR  | -
>>>>>>> | NOSTATE     | demo-net=x.x.x.x |
>>>>>>>
>>>>>>> +--------------------------------------+-----------------+--------+------------+-------------+-----------------------+
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> This error apperas in the /var/log/syslog file on the computer node:
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.597951]  nbd8: p1
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.668430] EXT4-fs (nbd8): VFS:
>>>>>>> Can't find ext4 filesystem
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.668521] EXT4-fs (nbd8): VFS:
>>>>>>> Can't find ext4 filesystem
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.668583] EXT4-fs (nbd8): VFS:
>>>>>>> Can't find ext4 filesystem
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.668899] FAT-fs (nbd8): 
>>>>>>> bogus number of reserved sectors
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.668936] FAT-fs (nbd8): 
>>>>>>> Can't find a valid FAT filesystem
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.753989] block nbd8:
>>>>>>> NBD_DISCONNECT
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.754056] block nbd8: 
>>>>>>> Receive control failed (result -32)
>>>>>>>
>>>>>>> Jun 26 14:55:04 compute5 kernel: [ 2187.754161] block nbd8: 
>>>>>>> queue cleared
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> Also, this is logged on the computer node, in 
>>>>>>> /var/log/nova/nova-compute.log
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> 2015-06-26 14:55:02.591 7961 AUDIT nova.compute.claims [-] [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c] disk limit not specified, 
>>>>>>> defaulting to unlimited
>>>>>>>
>>>>>>> 2015-06-26 14:55:02.606 7961 AUDIT nova.compute.claims [-] [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c] Claim successful
>>>>>>>
>>>>>>> 2015-06-26 14:55:02.721 7961 INFO nova.scheduler.client.report 
>>>>>>> [-] Compute_service record updated for ('compute5')
>>>>>>>
>>>>>>> 2015-06-26 14:55:02.836 7961 INFO nova.scheduler.client.report 
>>>>>>> [-] Compute_service record updated for ('compute5')
>>>>>>>
>>>>>>> 2015-06-26 14:55:03.115 7961 INFO nova.virt.libvirt.driver [-]
>>>>>>> [instance: 17d26ca3-f56c-4a87-ae0a-acfafea4838c] Creating image
>>>>>>>
>>>>>>> 2015-06-26 14:55:03.118 7961 INFO nova.openstack.common.lockutils 
>>>>>>> [-] Created lock path:
>>>>>>> /var/lib/nova/instances/locks
>>>>>>>
>>>>>>> 2015-06-26 14:55:03.458 7961 INFO nova.scheduler.client.report 
>>>>>>> [-] Compute_service record updated for ('compute5',
>>>>>>> 'compute5.siminn.is')
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.088 7961 INFO nova.virt.disk.vfs.api [-] 
>>>>>>> Unable to import guestfsfalling back to VFSLocalFS
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 ERROR nova.compute.manager [-] [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c] Instance failed to spawn
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c] Traceback (most recent call
>>>>>>> last):
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py",
>>>>>>> line 2267, in _build_resources
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]     yield resources
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py",
>>>>>>> line 2137, in _build_and_run_instance
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]
>>>>>>> block_device_info=block_device_info)
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py",
>>>>>>> line 2620, in spawn
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]     write_to_disk=True)
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py",
>>>>>>> line 4159, in _get_guest_xml
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]     context)
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py",
>>>>>>> line 3937, in _get_guest_config
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]     flavor,
>>>>>>> CONF.libvirt.virt_type)
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]   File
>>>>>>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/vif.py",
>>>>>>> line 352, in get_config
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]     _("Unexpected vif_type=%s")
>>>>>>> % vif_type)
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c] NovaException: Unexpected 
>>>>>>> vif_type=binding_failed
>>>>>>>
>>>>>>> 2015-06-26 14:55:04.363 7961 TRACE nova.compute.manager [instance:
>>>>>>> 17d26ca3-f56c-4a87-ae0a-acfafea4838c]
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>>  
>>>>>>>
>>>>>>> Best regards
>>>>>>>
>>>>>>> Yngvi
>>>>>>>
>>>>>>>
>>>>>>> _______________________________________________
>>>>>>> Mailing list: 
>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>> Post to     : openstack@lists.openstack.org
>>>>>>> Unsubscribe : 
>>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>
>>>>>> --
>>>>>> Andreas
>>>>>> (IRC: scheuran)
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> Mailing list: 
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>> Post to     : openstack@lists.openstack.org
>>>>>> Unsubscribe : 
>>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Mailing list: 
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>> Post to     : openstack@lists.openstack.org
>>>>> Unsubscribe : 
>>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>>
>>>> --
>>>> Andreas
>>>> (IRC: scheuran)
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>> Post to     : openstack@lists.openstack.org
>>>> Unsubscribe : 
>>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
>>>
>>> --
>>> Andreas
>>> (IRC: scheuran)
>>>
>>>
>>
>> --
>> Andreas
>> (IRC: scheuran)
>>
>>
> 

_______________________________________________
Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
Post to     : openstack@lists.openstack.org
Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack

Reply via email to