Hello, I'm configuring it manually. DHCP is not working because vxlan tunnels are not working at all :/ Compute nodes and network nodes con ping each other:
admin@network:~$ ping 10.1.0.4 PING 10.1.0.4 (10.1.0.4) 56(84) bytes of data. 64 bytes from 10.1.0.4: icmp_seq=1 ttl=64 time=8.83 ms 64 bytes from 10.1.0.4: icmp_seq=2 ttl=64 time=0.282 ms ^C --- 10.1.0.4 ping statistics --- 2 packets transmitted, 2 received, 0% packet loss, time 1001ms rtt min/avg/max/mdev = 0.282/4.560/8.838/4.278 ms -- Best regards / Pozdrawiam Sławek Kapłoński sla...@kaplonski.pl On Sat, 03 Oct 2015, James Denton wrote: > Are your instances getting their ip from DHCP server or are you manually > configuring them? Can the network node ping the compute node at 10.1.0.4 and > vice-versa? > > Sent from my iPhone > > > On Oct 3, 2015, at 3:55 AM, Sławek Kapłoński <sla...@kaplonski.pl> wrote: > > > > This vlan bridge_mapping I set just to be sure if it will not help for > > some reason :) Before I tested it without this mapping configured. And > > in fact I'm not using vlan networks at all (at least now) - I only want > > to have local vxlan network between instances :) > > When I booted one instance on host in brqXXX bridge I got vxlan-10052 > > port and tapXXX port (10052 is vni used assigned to network in neutron). > > After boot second vm I got in same bridge second tap interface so it > > looks like: > > > > root@compute-2:~# brctl show > > bridge name bridge id STP enabled interfaces > > brq8fe8a32f-e6 8000.ce544d0c0e5d no tap691a138a-6c > > tapbc1e5179-53 > > vxlan-10052 > > virbr0 8000.5254007611ab yes virbr0-nic > > > > > > So it looks fine for me. I have no idea what is this vibr0 bridge - maybe it > > should be used somehow? > > > > One more think. Those two vms on one host are pinging each other. So bridge > > looks that is working fine. Problem is with vxlan tunnels. > > > > About security groups: by default there is rule to allow traffic from > > different > > vms using same SG. All my instances are using same security group so it > > should > > be no problem IMHO. > > > > -- > > Best regards / Pozdrawiam > > Sławek Kapłoński > > sla...@kaplonski.pl > > > >> On Fri, 02 Oct 2015, James Denton wrote: > >> > >> If eth1 is used for the vxlan tunnel end points, it can't also be used in > >> a bridge ala provider_bridge_mappings. You should have a dedicated > >> interface or a vlan interface off eth1 (i.e. Eth1.20) that is dedicated to > >> the overlay traffic. Move the local_ip address to that interface on > >> respective nodes. Verify that you can ping between nodes at each address. > >> If this doesn't work, the Neutron pieces won't work. You shouldn't have to > >> restart any neutron services, since the IP isn't changing. > >> > >> Once you create a vxlan tenant network and boot some instances, verify > >> that the vxlan interface is being setup and placed in the respective > >> bridge. You can use 'brctl show' to look at the brq bridge that > >> corresponds to the network. You should see a vxlan interface and the tap > >> interfaces of your instances. > >> > >> As always, verify your security groups first when troubleshooting instance > >> to instance communication. > >> > >> James > >> > >> Sent from my iPhone > >> > >>> On Oct 2, 2015, at 3:48 PM, Sławek Kapłoński <sla...@kaplonski.pl> wrote: > >>> > >>> Hello, > >>> > >>> I'm trying to configure small openstack infra (one network node, 2 > >>> compute nodes) with linux bridge and vxlan tenant networks. I don't know > >>> what I'm doing wrong but my instances have no connection between > >>> each other. On compute hosts I run neutron-plugin-linuxbrigde-agent > >>> with config like: > >>> > >>> ------------------ > >>> [ml2_type_vxlan] > >>> # (ListOpt) Comma-separated list of <vni_min>:<vni_max> tuples > >>> # enumerating > >>> # ranges of VXLAN VNI IDs that are available for tenant network > >>> # allocation. > >>> # > >>> vni_ranges = 10000:20000 > >>> > >>> # (StrOpt) Multicast group for the VXLAN interface. When configured, > >>> # will > >>> # enable sending all broadcast traffic to this multicast group. When > >>> # left > >>> # unconfigured, will disable multicast VXLAN mode. > >>> # > >>> # vxlan_group = > >>> # Example: vxlan_group = 239.1.1.1 > >>> > >>> [securitygroup] > >>> # Controls if neutron security group is enabled or not. > >>> # It should be false when you use nova security group. > >>> enable_security_group = True > >>> > >>> # Use ipset to speed-up the iptables security groups. Enabling ipset > >>> # support > >>> # requires that ipset is installed on L2 agent node. > >>> enable_ipset = True > >>> > >>> firewall_driver = > >>> neutron.agent.linux.iptables_firewall.IptablesFirewallDriver > >>> > >>> [ovs] > >>> local_ip = 10.1.0.4 > >>> > >>> [agent] > >>> tunnel_types = vxlan > >>> > >>> [linuxbridge] > >>> physical_interface_mappings = physnet1:eth1 > >>> > >>> [vxlan] > >>> local_ip = 10.1.0.4 > >>> l2_population = True > >>> enable_vxlan = True > >>> ------------------- > >>> > >>> Eth1 is my "tunnel network" which should be used for tunnels. When I > >>> spawn vms on compute 1 and 2 and after configuring network manually on > >>> both vms (dhcp is not working also because of broken tunnels probably) > >>> it not pings. > >>> Even when I started two instances on same host and they are both > >>> connected to one bridge: > >>> > >>> ------------------- > >>> root@compute-2:/usr/lib/python2.7/dist-packages/neutron# brctl show > >>> bridge name bridge id STP enabled interfaces > >>> brq8fe8a32f-e6 8000.ce544d0c0e5d no tap691a138a-6c > >>> tapbc1e5179-53 > >>> vxlan-10052 > >>> virbr0 8000.5254007611ab yes virbr0-nic > >>> ------------------- > >>> > >>> those 2 vms are not pinging each other :/ > >>> I don't have any expeirence with linux bridge in fact (For now I was > >>> always > >>> using ovs). Maybe someone of You will know what I should check or what I > >>> should > >>> configure wrong :/ Generally I was installing this openstack according to > >>> official openstack documentation but in this docs there is info about > >>> ovs+gre > >>> tunnels and that is what I changed. I'm using Ubuntu 14.04 and Openstack > >>> Kilo > >>> installed from cloud archive repo. > >>> > >>> -- > >>> Best regards / Pozdrawiam > >>> Sławek Kapłoński > >>> sla...@kaplonski.pl > >>> > >>> _______________________________________________ > >>> OpenStack-operators mailing list > >>> openstack-operat...@lists.openstack.org > >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack-operators
signature.asc
Description: Digital signature
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack