Hi, It does not work ... ON network node: root@network2:/# nmap -sO 172.22.14.17
Starting Nmap 6.40 ( http://nmap.org ) at 2015-06-29 23:19 GMT Warning: 172.22.14.17 giving up on port because retransmission cap hit (10). root@network2:/# nmap -sO 172.22.15.17 Starting Nmap 6.40 ( http://nmap.org ) at 2015-06-29 23:21 GMT Warning: 172.22.15.17 giving up on port because retransmission cap hit (10). ON compute node: root@compute5:/# nmap -sO 172.22.14.14 Starting Nmap 6.40 ( http://nmap.org ) at 2015-06-29 23:21 GMT Warning: 172.22.14.14 giving up on port because retransmission cap hit (10). root@compute5:/# nmap -sO 172.22.15.14 Starting Nmap 6.40 ( http://nmap.org ) at 2015-06-29 23:21 GMT Warning: 172.22.15.14 giving up on port because retransmission cap hit (10). Best regards Yngvi -----Original Message----- From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] Sent: 29. júní 2015 23:18 To: Yngvi Páll Þorfinnsson; YANG LI Cc: openstack@lists.openstack.org Subject: RE: [Openstack] network question on openstack installation Hm, I'm running out of ideas. Can you run those two commands to verify that GRE traffic can pass the firewalls: Network node: nmap -sO <IP of compute node> Compute node: nmap -sO <IP of network node> In both cases, that's a big o, not a zero. Am 30. Juni 2015 01:07:04 MESZ, schrieb "Yngvi Páll Þorfinnsson" <yngv...@siminn.is>: >OK, so I ran the command you sent me , and now it looks it‘s allowed, > >Controller: > >root@controller2:/# iptables -L -nv --line-numbers Chain INPUT (policy >ACCEPT 6437 packets, 2609K bytes) >num pkts bytes target prot opt in out source > destination >1 54411 17M nova-api-INPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 0 0 ACCEPT 47 -- * * 172.22.15.0/24 > 172.22.15.0/24 > >Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 0 0 nova-api-FORWARD all -- * * 0.0.0.0/0 > 0.0.0.0/0 > >Chain OUTPUT (policy ACCEPT 6228 packets, 2573K bytes) >num pkts bytes target prot opt in out source > destination >1 51439 16M nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 51439 16M nova-api-OUTPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > >Chain nova-api-FORWARD (1 references) >num pkts bytes target prot opt in out source > destination > >Chain nova-api-INPUT (1 references) >num pkts bytes target prot opt in out source > destination >1 0 0 ACCEPT tcp -- * * 0.0.0.0/0 > 172.22.14.22 tcp dpt:8775 > >Chain nova-api-OUTPUT (1 references) >num pkts bytes target prot opt in out source > destination > >Chain nova-api-local (1 references) >num pkts bytes target prot opt in out source > destination > >Chain nova-filter-top (2 references) >num pkts bytes target prot opt in out source > destination >1 51439 16M nova-api-local all -- * * 0.0.0.0/0 > 0.0.0.0/0 >root@controller2:/# > > >Network: > >root@network2:/# iptables -L -nv --line-numbers Chain INPUT (policy >ACCEPT 8 packets, 512 bytes) >num pkts bytes target prot opt in out source > destination >1 2215 700K neutron-openvswi-INPUT all -- * * >0.0.0.0/0 0.0.0.0/0 >2 0 0 ACCEPT 47 -- * * 172.22.15.0/24 > 172.22.15.0/24 > >Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 neutron-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 0 0 neutron-openvswi-FORWARD all -- * * >0.0.0.0/0 0.0.0.0/0 > >Chain OUTPUT (policy ACCEPT 5 packets, 664 bytes) >num pkts bytes target prot opt in out source > destination >1 1827 332K neutron-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 1827 332K neutron-openvswi-OUTPUT all -- * * >0.0.0.0/0 0.0.0.0/0 > >Chain neutron-filter-top (2 references) >num pkts bytes target prot opt in out source > destination >1 1827 332K neutron-openvswi-local all -- * * >0.0.0.0/0 0.0.0.0/0 > >Chain neutron-openvswi-FORWARD (1 references) >num pkts bytes target prot opt in out source > destination > >Chain neutron-openvswi-INPUT (1 references) >num pkts bytes target prot opt in out source > destination > >Chain neutron-openvswi-OUTPUT (1 references) >num pkts bytes target prot opt in out source > destination > >Chain neutron-openvswi-local (1 references) >num pkts bytes target prot opt in out source > destination > >Chain neutron-openvswi-sg-chain (0 references) >num pkts bytes target prot opt in out source > destination > >Chain neutron-openvswi-sg-fallback (0 references) >num pkts bytes target prot opt in out source > destination >1 0 0 DROP all -- * * 0.0.0.0/0 > 0.0.0.0/0 >root@network2:/# > >compute: > >root@compute5:/# iptables -L -nv --line-numbers Chain INPUT (policy >ACCEPT 24 packets, 7092 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 ACCEPT udp -- virbr0 * 0.0.0.0/0 > 0.0.0.0/0 udp dpt:53 >2 0 0 ACCEPT 47 -- * * 172.22.15.0/24 > 172.22.15.0/24 >3 0 0 ACCEPT tcp -- virbr0 * 0.0.0.0/0 > 0.0.0.0/0 tcp dpt:53 >4 0 0 ACCEPT udp -- virbr0 * 0.0.0.0/0 > 0.0.0.0/0 udp dpt:67 >5 0 0 ACCEPT tcp -- virbr0 * 0.0.0.0/0 > 0.0.0.0/0 tcp dpt:67 > >Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 ACCEPT all -- * virbr0 0.0.0.0/0 > 192.168.122.0/24 ctstate RELATED,ESTABLISHED >2 0 0 ACCEPT all -- virbr0 * 192.168.122.0/24 > 0.0.0.0/0 >3 0 0 ACCEPT all -- virbr0 virbr0 0.0.0.0/0 > 0.0.0.0/0 >4 0 0 REJECT all -- * virbr0 0.0.0.0/0 > 0.0.0.0/0 reject-with icmp-port-unreachable >5 0 0 REJECT all -- virbr0 * 0.0.0.0/0 > 0.0.0.0/0 reject-with icmp-port-unreachable > >Chain OUTPUT (policy ACCEPT 14 packets, 2340 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 ACCEPT udp -- * virbr0 0.0.0.0/0 > 0.0.0.0/0 udp dpt:68 > > >I tried once more to create an instance, but if failed as well > >Best regards >Yngvi > >From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] >Sent: 29. júní 2015 23:01 >To: Yngvi Páll Þorfinnsson; YANG LI >Cc: openstack@lists.openstack.org >Subject: RE: [Openstack] network question on openstack installation > >I'm not sure if there is something wrong but on both hosts I don't see >any rule that accepts GRE traffic. You need to allow GRE traffic on >your internal network so that tunneling can actually work. Without that >it's like having your network configured but not plugged in... >Am 30. Juni 2015 00:50:28 MESZ, schrieb "Yngvi Páll Þorfinnsson" ><yngv...@siminn.is<mailto:yngv...@siminn.is>>: >Network node: > > >root@network2:/# iptables -L -nv --line-numbers Chain INPUT (policy >ACCEPT 1286 packets, 351K bytes) >num pkts bytes target prot opt in out source > destination >1 1171 338K neutron-openvswi-INPUT all -- * * >0.0.0.0/0 0.0.0.0/0 > > >Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 neutron-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 0 0 neutron-openvswi-FORWARD all -- * * >0.0.0.0/0 0.0.0.0/0 > > >Chain OUTPUT (policy ACCEPT 1124 packets, 180K bytes) >num pkts bytes target prot opt in out source > destination >1 988 164K neutron-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 988 164K neutron-openvswi-OUTPUT all -- * * >0.0.0.0/0 0.0.0.0/0 > > >Chain neutron-filter-top (2 references) >num pkts bytes target prot opt in out source > destination >1 988 164K neutron-openvswi-local all -- * * >0.0.0.0/0 0.0.0.0/0 > > >Chain neutron-openvswi-FORWARD (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain neutron-openvswi-INPUT (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain neutron-openvswi-OUTPUT (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain neutron-openvswi-local (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain neutron-openvswi-sg-chain (0 references) >num pkts bytes target prot opt in out source > destination > > >Chain neutron-openvswi-sg-fallback (0 references) >num pkts bytes target prot opt in out source > destination >1 0 0 DROP all -- * * 0.0.0.0/0 > 0.0.0.0/0 >root@network2:/# > > >Controller node > > >root@controller2:/# iptables -L -nv --line-numbers Chain INPUT (policy >ACCEPT 25498 packets, 7540K bytes) >num pkts bytes target prot opt in out source > destination >1 25498 7540K nova-api-INPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > >Chain FORWARD (policy ACCEPT 0 packets, 0 bytes) >num pkts bytes target prot opt in out source > destination >1 0 0 nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 0 0 nova-api-FORWARD all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > >Chain OUTPUT (policy ACCEPT 24216 packets, 7244K bytes) >num pkts bytes target prot opt in out source > destination >1 24216 7244K nova-filter-top all -- * * 0.0.0.0/0 > 0.0.0.0/0 >2 24216 7244K nova-api-OUTPUT all -- * * 0.0.0.0/0 > 0.0.0.0/0 > > >Chain nova-api-FORWARD (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain nova-api-INPUT (1 references) >num pkts bytes target prot opt in out source > destination >1 0 0 ACCEPT tcp -- * * 0.0.0.0/0 > 172.22.14.22 tcp dpt:8775 > > >Chain nova-api-OUTPUT (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain nova-api-local (1 references) >num pkts bytes target prot opt in out source > destination > > >Chain nova-filter-top (2 references) >num pkts bytes target prot opt in out source > destination >1 24216 7244K nova-api-local all -- * * 0.0.0.0/0 > 0.0.0.0/0 >root@controller2:/# > > > > > > >From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] >Sent: 29. júní 2015 22:46 >To: Yngvi Páll Þorfinnsson; YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> >Subject: RE: [Openstack] network question on openstack installation > > >One more thing. Please provide > >iptables -L -nv --line-numbers > >for network and compute nodes. >Am 30. Juni 2015 00:25:45 MESZ, schrieb "Yngvi Páll Þorfinnsson" ><yngv...@siminn.is<mailto:yngv...@siminn.is>>: >In fact, I don‘t think I‘ll need more than one „external network“ so, >am I on the wrong path here, i.e. when I‘m configuing the external >network as a VLAN ? > > >Best regards >Yngvi > > >From: Yngvi Páll Þorfinnsson >Sent: 29. júní 2015 21:57 >To: Uwe Sauter; YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> >Subject: Re: [Openstack] network question on openstack installation > > >OK, I only found one fresh error >Compute node; nova-compute.log ( as usually when I create instance) > > >grep ERR nova-compute.log >2015-06-29 21:11:11.801 4166 ERROR nova.compute.manager [-] [instance: >af901a2b-2462-4c19-b1f1-237371fd8177] Instance failed to spawn > > >I‘ve attached the neutron agent-show and neutron (sub)net-list in the >attached file. > > >Best regards >Yngvi > > > > >From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] >Sent: 29. júní 2015 21:34 >To: Yngvi Páll Þorfinnsson; YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> >Subject: RE: [Openstack] network question on openstack installation > > >Can you check for ERRORs in: >Network node: neutron server log, neutron openvswitch agent log, >openvswitch log Nova controller node: nova api log, nova scheduler log >Compute node: nova compute log, neutron openvswitch agent log, >openvswitch log > >Also please list again neutron agent-show for the different agents and >neutron net-show and neutron subnet-show for your (sub)networks. >Am 29. Juni 2015 23:24:48 MESZ, schrieb "Yngvi Páll Þorfinnsson" ><yngv...@siminn.is<mailto:yngv...@siminn.is>>: >Thanks a lot for your effort Uwe ;-) >It‘s relly helpful ! > > >Now , I keep creating instances, and have the same error. >I still get a strange comment in the >Neutron server.log file when I try to create an instance: > > >2015-06-29 21:11:11.576 1960 DEBUG >neutron.plugins.ml2.drivers.mech_openvswitch >[req-1e603e4b-61e6-4896-8f81-208daba8569b None] Checking segment: >{'segmentation_id': 1102L, 'physical_network': u'external', 'id': >u'11fab5ad-c457-4175-9e5a-f505fc0e072d', 'network_type': u'vlan'} for >mappings: {u'external': u'br-ex'} with tunnel_types: [u'gre'] >check_segment_for_agent >/usr/lib/python2.7/dist-packages/neutron/plugins/ml2/drivers/mech_openv >switch.py:52 > > >But still this segment is not listed for ‚neutron net-list‘ nor >‚neutron-subnet-list‘ ! > > > > >root@controller2:/# source admin-openrc.sh root@controller2:/# >root@controller2:/# neutron net-list >+--------------------------------------+-------------+-----------------------------------------------------+ >| id | name | subnets > | >+--------------------------------------+-------------+-----------------------------------------------------+ >| b43da44a-42d5-4b1f-91c2-d06a923deb29 | ext_net1101 | >c40fa8e3-cd8e-4566-ade6-5f3eabed121c 157.157.8.0/24 | >| 3446e54b-346f-45e5-89a2-1ec4eef251ab | demo-net | >2c79bb00-0ace-4319-8151-81210ee3dfb2 172.22.18.0/24 | >+--------------------------------------+-------------+-----------------------------------------------------+ >root@controller2:/# >root@controller2:/# neutron subnet-list >+--------------------------------------+--------------------+----------------+---------------------------------------------------+ >| id | name | cidr > | allocation_pools | >+--------------------------------------+--------------------+----------------+---------------------------------------------------+ >| c40fa8e3-cd8e-4566-ade6-5f3eabed121c | ext_net1101-subnet | >157.157.8.0/24 | {"start": "157.157.8.51", "end": "157.157.8.250"} | >| 2c79bb00-0ace-4319-8151-81210ee3dfb2 | demo-subnet | >172.22.18.0/24 | {"start": "172.22.18.2", "end": "172.22.18.254"} | >+--------------------------------------+--------------------+----------------+---------------------------------------------------+ > > > > >And the output of listing it is empty: > > >root@controller2:/# neutron net-show >11fab5ad-c457-4175-9e5a-f505fc0e072d >Unable to find network with name '11fab5ad-c457-4175-9e5a-f505fc0e072d' >root@controller2:/# >root@controller2:/# source demo-openrc.sh root@controller2:/# neutron >net-show 11fab5ad-c457-4175-9e5a-f505fc0e072d >Unable to find network with name '11fab5ad-c457-4175-9e5a-f505fc0e072d' > > >This is after I have dropped the neutron, and resynced.... > > >Best regards >Yngvi > > >From: Uwe Sauter [mailto:uwe.sauter...@gmail.com] >Sent: 29. júní 2015 21:16 >To: Yngvi Páll Þorfinnsson; YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> >Subject: RE: [Openstack] network question on openstack installation > > >Yes. Just keep in mind that if you extend your configuration with a new >config file, then you must change your init script / unit file to >reference that file. And it would probably be a good idea to re-sync >the DB with that additional file as an option. Or you keep your plugin >configuration in a single file and be happy with the current layout. >Am 29. Juni 2015 22:47:22 MESZ, schrieb "Yngvi Páll Þorfinnsson" ><yngv...@siminn.is<mailto:yngv...@siminn.is>>: > >Hi Uwe > >I just ran this some minutes ago, i.e. did the "population of db" >again, >According to the manual. >Should'nt this be enough ? > > >root@controller2:/# su -s /bin/sh -c "neutron-db-manage --config-file >/etc/neutron/neutron.conf \ > >--config-file /etc/neutron/plugins/ml2/ml2_conf.ini upgrade juno" >neutron > >INFO [alembic.migration] Context impl MySQLImpl. >INFO [alembic.migration] Will assume non-transactional DDL. >INFO [alembic.migration] Running upgrade None -> havana, >havana_initial INFO [alembic.migration] Running upgrade havana -> >e197124d4b9, add unique constraint to members INFO [alembic.migration] >Running upgrade e197124d4b9 -> 1fcfc149aca4, Add a unique constraint on >(agent_type, host) columns to prevent a race condition when an agent > >entry is 'upserted'. >INFO [alembic.migration] Running upgrade 1fcfc149aca4 -> 50e86cb2637a, >nsx_mappings INFO [alembic.migration] Running upgrade 50e86cb2637a -> >1421183d533f, NSX DHCP/metadata support INFO [alembic.migration] >Running upgrade 1421183d533f -> 3d3cb89d84ee, nsx_switch_mappings INFO >[alembic.migration] Running upgrade 3d3cb89d84ee -> 4ca36cfc898c, >nsx_router_mappings INFO [alembic.migration] Running upgrade >4ca36cfc898c -> 27cc183af192, ml2_vnic_type INFO [alembic.migration] >Running upgrade 27cc183af192 -> 50d5ba354c23, >ml2 binding:vif_details >INFO [alembic.migration] Running upgrade 50d5ba354c23 -> 157a5d299379, >ml2 binding:profile >INFO [alembic.migration] Running upgrade 157a5d299379 -> 3d2585038b95, >VMware NSX rebranding INFO [alembic.migration] Running upgrade >3d2585038b95 -> abc88c33f74f, lb stats INFO > >[alembic.migration] > >Running upgrade abc88c33f74f -> 1b2580001654, > >nsx_sec_group_mapping >INFO [alembic.migration] Running upgrade 1b2580001654 -> e766b19a3bb, >nuage_initial INFO [alembic.migration] Running upgrade e766b19a3bb -> >2eeaf963a447, floatingip_status INFO [alembic.migration] Running >upgrade 2eeaf963a447 -> 492a106273f8, Brocade ML2 Mech. Driver INFO >[alembic.migration] Running upgrade 492a106273f8 -> 24c7ea5160d7, Cisco >CSR VPNaaS INFO [alembic.migration] Running upgrade 24c7ea5160d7 -> >81c553f3776c, bsn_consistencyhashes INFO [alembic.migration] Running >upgrade 81c553f3776c -> 117643811bca, >nec: delete old ofc mapping tables >INFO [alembic.migration] Running upgrade 117643811bca -> 19180cf98af6, >nsx_gw_devices INFO [alembic.migration] Running upgrade 19180cf98af6 >-> 33dd0a9fa487, embrane_lbaas_driver INFO [alembic.migration] Running >upgrade 33dd0a9fa487 -> 2447ad0e9585, Add IPv6 Subnet properties INFO >[alembic.migration] Running upgrade 2447ad0e9585 > >-> 538732fa21e1, NEC Rename quantum_id to neutron_id >INFO [alembic.migration] Running upgrade 538732fa21e1 -> 5ac1c354a051, >n1kv segment allocs for cisco n1kv plugin INFO [alembic.migration] >Running upgrade 5ac1c354a051 -> icehouse, icehouse INFO >[alembic.migration] Running upgrade icehouse -> 54f7549a0e5f, >set_not_null_peer_address INFO [alembic.migration] Running upgrade >54f7549a0e5f -> 1e5dd1d09b22, set_not_null_fields_lb_stats INFO >[alembic.migration] Running upgrade 1e5dd1d09b22 -> b65aa907aec, >set_length_of_protocol_field INFO [alembic.migration] Running upgrade >b65aa907aec -> 33c3db036fe4, set_length_of_description_field_metering >INFO [alembic.migration] Running upgrade 33c3db036fe4 -> 4eca4a84f08a, >Remove ML2 Cisco Credentials DB INFO [alembic.migration] Running >upgrade 4eca4a84f08a -> d06e871c0d5, >set_admin_state_up_not_null_ml2 >INFO > >[alembic.migration] Running upgrade d06e871c0d5 -> > >6be312499f9, set_not_null_vlan_id_cisco INFO [alembic.migration] >Running upgrade 6be312499f9 -> 1b837a7125a9, Cisco APIC Mechanism >Driver INFO [alembic.migration] Running upgrade 1b837a7125a9 -> >10cd28e692e9, nuage_extraroute INFO [alembic.migration] Running >upgrade 10cd28e692e9 -> 2db5203cb7a9, nuage_floatingip INFO >[alembic.migration] Running upgrade 2db5203cb7a9 -> 5446f2a45467, >set_server_default INFO [alembic.migration] Running upgrade >5446f2a45467 -> db_healing, Include all tables and make migrations >unconditional. >INFO [alembic.migration] Context impl MySQLImpl. >INFO [alembic.migration] Will assume non-transactional DDL. >INFO [alembic.autogenerate.compare] Detected server default on column >'cisco_ml2_apic_epgs.provider' >INFO [alembic.autogenerate.compare] Detected removed index >'cisco_n1kv_vlan_allocations_ibfk_1' on 'cisco_n1kv_vlan_allocations' >INFO [alembic.autogenerate.compare] Detected server > >default on column 'cisco_n1kv_vxlan_allocations.allocated' >INFO [alembic.autogenerate.compare] Detected removed index >'cisco_n1kv_vxlan_allocations_ibfk_1' on 'cisco_n1kv_vxlan_allocations' >INFO [alembic.autogenerate.compare] Detected removed index >'embrane_pool_port_ibfk_2' on 'embrane_pool_port' >INFO [alembic.autogenerate.compare] Detected removed index >'firewall_rules_ibfk_1' on 'firewall_rules' >INFO [alembic.autogenerate.compare] Detected removed index >'firewalls_ibfk_1' on 'firewalls' >INFO [alembic.autogenerate.compare] Detected server default on column >'meteringlabelrules.excluded' >INFO [alembic.autogenerate.compare] Detected server default on column >'ml2_port_bindings.host' >INFO [alembic.autogenerate.compare] Detected added column >'nuage_routerroutes_mapping.destination' >INFO [alembic.autogenerate.compare] Detected added column >'nuage_routerroutes_mapping.nexthop' >INFO > >[alembic.autogenerate.compare] Detected server default > >on column 'poolmonitorassociations.status' >INFO [alembic.autogenerate.compare] Detected added index >'ix_quotas_tenant_id' on '['tenant_id']' >INFO [alembic.autogenerate.compare] Detected NULL on column >'tz_network_bindings.phy_uuid' >INFO [alembic.autogenerate.compare] Detected NULL on column >'tz_network_bindings.vlan_id' >INFO [neutron.db.migration.alembic_migrations.heal_script] Detected >removed foreign key u'nuage_floatingip_pool_mapping_ibfk_2' on table >u'nuage_floatingip_pool_mapping' >INFO [alembic.migration] Running upgrade db_healing -> 3927f7f7c456, >L3 extension distributed mode >INFO [alembic.migration] Running upgrade 3927f7f7c456 -> 2026156eab2f, >L2 models to support DVR >INFO [alembic.migration] Running upgrade 2026156eab2f -> 37f322991f59, >removing_mapping_tables INFO [alembic.migration] Running upgrade >37f322991f59 -> 31d7f831a591, add constraint for routerid INFO >[alembic.migration] Running upgrade > >31d7f831a591 -> 5589aa32bf80, L3 scheduler additions to support DVR >INFO [alembic.migration] Running upgrade 5589aa32bf80 -> 884573acbf1c, >Drop NSX table in favor of the extra_attributes one INFO >[alembic.migration] Running upgrade 884573acbf1c -> 4eba2f05c2f4, >correct Vxlan Endpoint primary key INFO [alembic.migration] Running >upgrade 4eba2f05c2f4 -> 327ee5fde2c7, set_innodb_engine INFO >[alembic.migration] Running upgrade 327ee5fde2c7 -> 3b85b693a95f, Drop >unused servicedefinitions and servicetypes tables. >INFO [alembic.migration] Running upgrade 3b85b693a95f -> aae5706a396, >nuage_provider_networks INFO [alembic.migration] Running upgrade >aae5706a396 -> 32f3915891fd, cisco_apic_driver_update INFO >[alembic.migration] Running upgrade 32f3915891fd -> 58fe87a01143, >cisco_csr_routing INFO [alembic.migration] Running upgrade >58fe87a01143 -> 236b90af57ab, > >ml2_type_driver_refactor_dynamic_segments >INFO > >[alembic.migration] Running upgrade 236b90af57ab -> 86d6d9776e2b, Cisco >APIC Mechanism Driver INFO [alembic.migration] Running upgrade >86d6d9776e2b -> 16a27a58e093, ext_l3_ha_mode INFO [alembic.migration] >Running upgrade 16a27a58e093 -> 3c346828361e, metering_label_shared >INFO [alembic.migration] Running upgrade 3c346828361e -> 1680e1f0c4dc, >Remove Cisco Nexus Monolithic Plugin INFO [alembic.migration] Running >upgrade 1680e1f0c4dc -> 544673ac99ab, add router port relationship INFO >[alembic.migration] Running upgrade 544673ac99ab -> juno, juno >root@controller2:/# > >-----Original Message----- >From: Uwe Sauter [mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com] >Sent: 29. júní 2015 20:45 >To: Yngvi Páll Þorfinnsson; YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> >Subject: Re: [Openstack] network > >question > >on > >openstack installation > >Go to http://docs.openstack.org/ , select the OpenStack version, then >the installation guide for your distribution and navigate to > >6. Add a networking component > - OpenStack Networking (neutron) > - Install and configure controller node > >and follow the database related stuff: >- create DB >- grant privileges to neutron user >- populate the DB (search for "neutron-db-manage") > >Depending on your distribution, you probably need to use sudo. > >The number of config files depends on the file layout you are working >with, so this is not an exact answer. > >I did the following: > >[on controller node] > >cd /etc/neutron >cat plugins/ml2/ml2_conf.ini > plugin.ini cd plugins/ml2 rm -f >ml2_config.ini ln -s ../../plugin.ini ml2_conf.ini > > >[on network and compute nodes] > >cd > >/etc/neutron >cat plugins/ml2/ml2_conf.ini > plugin.ini cat > >plugins/openvswitch/ovs_neutron_plugin.ini >> plugin.ini cd plugins/ml2 >rm -f ml2_config.ini ln -s ../../plugin.ini ml2_conf.ini cd >../openvswitch rm -f ovs_neutron_plugin.ini ln -s ../../plugin.ini >ovs_neutron_plugin.ini > > > >Then I did that sed trick that is needed on RDO because of a packaging >bug with Juno (changes systemd unit file to load >/etc/neutron/plugin.ini instead of >/etc/neutron/plugins/ml2/ml2_plugin.ini). >And then 'su -s /bin/sh -c "neutron-db-manage --config-file >/etc/neutron/neutron.conf --config-file /etc/neutron/plugin.ini upgrade >juno" neutron' on the controller node. > >As far as I understood the configuration system used in Neutron, each >config file will be merged into the same configuration namespace; only >the order will determine, which value is the current one if they exist >in more then one file [] section. >So you could have just one ini file in addition to >/etc/neutron/neutron.conf (my > >way) > >or > >several. But then you > >need to load them all when sync'ing the DB *and* starting the services. > > >Hopefully this brightens the dark side of Neutron configuration, > > Uwe > > > > > >Am 29.06.2015 um 22:19 schrieb Yngvi Páll Þorfinnsson: > > HI Uwe > No, I did'nt drop the keystone ;-) > > But is this the correct way to resync neutron ? > > # neutron-db-manage --config-file /etc/neutron/neutron.conf \ # > --config-file /etc/neutron/plugins/ml2/ml2_plugin.ini > > I mean, how many config files is necessary to have in the cmd ? > > best regards > Yngvi > > -----Original Message----- >From: Uwe Sauter [mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com] > Sent: 29. júní 2015 18:08 > To: YANG LI >Cc: openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> > Subject: Re: > >[Openstack] network question on openstack installation > >It depends on your switch… some drop tagged packets on an access port, >others allow tagged packets, if the packet VLAN ID equals the >configured VLAN ID. > > I'd reconfigure the provider network type to "flat" but that's > personal taste. You could also reconfigure the switch port to be a > trunking port (with only one VLAN ID). Currently your in between those > configurations that might or might not work lateron… > > > Regards, > > Uwe > > > Am 29.06.2015 um 19:42 schrieb YANG LI: > > thank you, Uwe. our provider network actually is untagged, but I did >specified VLAN ID when I create our external network and everything >still works. will this cause issue later on? > > eutron net-create --provider:network_type=vlan > —provider:segmentation_id=<vlan id> > --provider:physical_network=physnet1 > --router:external=true public > > > Thanks, > Yang > >On Jun 27, 2015, at 11:17 AM, Uwe Sauter ><uwe.sauter.de<http://uwe.sauter.de>@gmail.com ><mailto:uwe.sauter.de<http://uwe.sauter.de>@gmail.com>> wrote: > > Hi Yang, > >it depends on whether your provider network is tagged or untagged. If >it is untagged (the switch port is an "access" > port) then you don't specify the VLAN ID for the external network (as >it will get tagged by the switch). If the provider network is tagged >(the switch port is a "trunk" port) then you have to configure the VLAN >ID because the switch might refuse the traffic (depending if there is a >default VLAN ID defined on the switch's port). > >Regards, > > Uwe > > > > Am 27.06.2015 um 13:47 schrieb YANG LI: > > Thank you so much, James! This is so helpful. Another confusion I >have is about network_vlan_ranges. Is this network VLAN id range? >If so, does it has to match external network? For example, we only have >one external VLAN we can use as Our provider network and that VLAN id >is 775 (xxx.xxx.xxx.0/26). Should I define network_vlan_ranges as >following? > > [ml2] > type_drivers=vlan > tenant_network_types = vlan > mechanism_drivers=openvswitch > # > [ml2_type_vlan] > #thistellsOpenstackthattheinternalname"physnet1"providesthevlanrang > e > 100-199 > network_vlan_ranges=physnet1:775 > # > > Thanks, > Yang > Sent from my iPhone > > On Jun 26, 2015, at 8:54 AM, "James Denton" > ><james.den...@rackspace.com ><mailto:james.den...@rackspace.com<mailto:james.den...@rackspace.com%20 >%3cmailto:james.den...@rackspace.com>> > <mailto:james.den...@rackspace.com>> wrote: > > You can absolutely have instances in the same network span different >compute nodes. As an admin, you can run ‘nova show <instanceid>’ and >see the host in the output: > > root@controller01:~# nova show > 7bb18175-87da-4d1f-8dca-2ef07fee9d21 > | grep host >| OS-EXT-SRV-ATTR:host | compute02 > | > > That info is not available to non-admin users by default. > > James > > On Jun 26, 2015, at 7:38 AM, YANG LI <yan...@clemson.edu ><mailto:yan...@clemson.edu<mailto:yan...@clemson.edu%20%0b%20%3cmailto: >yan...@clemson.edu>> ><mailto:yan...@clemson.edu>> > >wrote: > > Thanks, James for the explanation. it make more sense now. ><http://now.it/> it is possible that a instances on same tenant network >reside on different compute nodes right? how do I tell which compute >node a instance is on? > > Thanks, > Yang > > On Jun 24, 2015, at 10:27 AM, James Denton <james.den...@rackspace.com ><mailto:james.den...@rackspace.com<mailto:james.den...@rackspace.com%20 >%3cmailto:james.den...@rackspace.com>> > <mailto:james.den...@rackspace.com>> wrote: > > Hello. > > all three nodes will have eth0 on management/api network. since I am > using ml2 plugin with vlan for tenant network, I think all compute > node should have eth1 as the second nic on provider network. Is this > >correct? I understand provider network is for instance to get external >access to internet, but how is instance live on compute1 communicate >with instance live on compute2? are they also go through provider >network? > > In short, yes. If you’re connecting instances to vlan “provider” > networks, traffic between instances on different compute nodes will >traverse the “provider bridge”, get tagged out eth1, and hit the >physical switching fabric. Your external gateway device could also sit >in that vlan, and the default route on the instance would direct >external traffic to that device. > > In reality, every network has ‘provider’ attributes that describe the > network type, segmentation id, and bridge interface (for vlan/flat > only). So tenant networks that leverage vlans would have provider > attributes set by Neutron automatically based on the configuration set > in the ML2 config file. If you > >use Neutron routers that connect to both ‘tenant’ vlan-based networks >and external ‘provider’ networks, all of that traffic could traverse >the same provider bridge on the controller/network node, but would be >tagged accordingly based on the network (ie. vlan 100 for external >network, vlan 200 for tenant network). > > Hope that’s not too confusing! > > James > > On Jun 24, 2015, at 8:54 AM, YANG LI <yan...@clemson.edu ><mailto:yan...@clemson.edu<mailto:yan...@clemson.edu%20%0b%20%3cmailto: >yan...@clemson.edu>> ><mailto:yan...@clemson.edu>> > wrote: > > I am working on install openstack from scratch, but get confused with >network part. I want to have one controller node, two compute nodes. > > the controller node will only handle following services: > glance-api > glance-registry > keystone > nova-api > nova-cert > >nova-conductor > nova-consoleauth > nova-novncproxy > nova-scheduler > qpid > mysql > neutron-server > > compute nodes will have following services: > neutron-dhcp-agent > neutron-l3-agent > neutron-metadata-agent > neutron-openvswitch-agent > neutron-ovs-cleanup > openvswtich > nova-compute > > all three nodes will have eth0 on management/api network. since I am >using ml2 plugin with vlan for tenant network, I think all compute >node should have eth1 as the second nic on provider network. Is this >correct? I understand provider network is for instance to get external >access to internet, but how is instance live on compute1 communicate >with instance live on compute2? are they also go through provider >network? > >________________________________ > >Mailing list: >http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > Post to : > >openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> ><mailto:openstack@lists.openstack.org> > <mailto:openstack@lists.openstack.org> > Unsubscribe : > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > > > > > > > >________________________________ > >Mailing list: >http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >Post to : >openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> ><mailto:openstack@lists.openstack.org> > Unsubscribe : > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > >________________________________ > >Mailing list: >http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > >Post to : >openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> ><mailto:openstack@lists.openstack.org> > Unsubscribe : > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > > > >________________________________ > >Mailing list: >http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >Post to : >openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> > Unsubscribe : > http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack > > > > > > > >-- >Diese Nachricht wurde von meinem Android-Mobiltelefon mit K-9 Mail >gesendet. -- Diese Nachricht wurde von meinem Android-Mobiltelefon mit K-9 Mail gesendet. _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack