Rohit, The routes are not the issue. Both SSVM VIFs are being bound to the host-only NIC. VirtualBox explicitly prevents routing between host-only and NAT NICs in its virtual switch. Therefore, the routing configuration below will not allow Internet access if the VIFs are not bound to the proper physical NIC.
Thanks, -John On Dec 15, 2012, at 9:35 PM, Rohit Yadav <rohit.ya...@citrix.com> wrote: > Hi John, > > I think I've the same issue you're facing. > The systemvms and other instances running on top of xen running on > hostonly/nat on vbox (devcloud like vm) won't be able to reach the Internet. > To fix the issue for dom0, I configured my network to route all traffic > through nat and make it the default gateway. > > auto xenbr0 > iface xenbr0 inet static > bridge_ports eth0 > address 192.168.56.10 > netmask 255.255.255.0 > network 192.168.56.0 > broadcast 192.168.56.255 > gateway 192.168.56.1 > dns_nameservers 8.8.8.8 8.8.4.4 > post-up route del default gw 192.168.56.1; route add default gw > 192.168.56.1 metric 100; > > auto xenbr1 > iface xenbr1 inet dhcp > bridge_ports eth1 > dns_nameservers 8.8.8.8 8.8.4.4 > post-up route add default gw 10.0.3.2 > > This fixed the problem on dom0, but I'm not sure what to do about the > systemvms and instances. May be you can try to swap nics such that nic1 is > nat and nic2 is host-only. > The problem I found using tcpdump was that only arp packets were going > through. So, maybe by enabling promiscuous networking mode on xen (just like > that was done on vbox) may fix this, but I've never tried that on devcloud. > > Regards. > ________________________________________ > From: John Burwell [jburw...@basho.com] > Sent: Sunday, December 16, 2012 2:55 AM > To: cloudstack-dev@incubator.apache.org > Subject: Re: SSVM Network Configuration Issue > > Rohit, > > As I stated below, I know which the VIF->PIF->device Xen mapping needs > to be used for each network. My question is how do I configure > CloudStack with that information. > > Thanks for your help, > -John > > > > > On Dec 15, 2012, at 2:51 PM, Rohit Yadav <rohit.ya...@citrix.com> wrote: > >> About xen bridging, these can help; >> >> brctl show all (bridge vif mappings) >> ip show addr xenbr0 (bridge specific info) >> brctl showmacs br0 (bridge mac mappings) >> >> Wiki: >> http://wiki.xen.org/wiki/Xen_FAQ_Networking >> http://wiki.xen.org/wiki/XenNetworking >> >> Regards. >> >> ________________________________________ >> From: John Burwell [jburw...@basho.com] >> Sent: Saturday, December 15, 2012 8:35 PM >> To: cloudstack-dev@incubator.apache.org >> Subject: Re: SSVM Network Configuration Issue >> >> Marcus, >> >> That's what I thought. The Xen physical bridge names are xenbr0 (to >> eth0) and xenbr1 (to eth1). Using basic network configuration, I set >> the Xen network traffic labels for each to the appropriate bridge >> device name. I receive errors regarding invalid network device when >> it attempts to create a VM. Does anyone else know how determine the >> mapping of physical devices to CloudStack Xen network traffic labels? >> >> Thanks, >> -John >> >> On Dec 15, 2012, at 1:20 AM, Marcus Sorensen <shadow...@gmail.com> wrote: >> >>> Vlans in advanced/KVM should only be required for the guest networks. If I >>> create a bridge on physical eth0, name it 'br0', and a bridge on physical >>> eth1, naming it 'br1', and then set my management network to label 'br0', >>> my public network to 'br1', and my guest network to 'br1', it should use >>> the bridges you asked for when connecting the system VMs for the specified >>> traffic. I'd just leave the 'vlan' blank when specifying public and >>> pod(management) IPs. In this scenario, the only place you need to enter >>> vlans is on the guest, and it should create new tagged interfaces/bridges >>> on eth1(per your label of br1) as new guest networks are brought online. >>> This is how my dev VMs are usually set up. >>> >>> >>> On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote: >>> >>>> Marcus, >>>> >>>> My question, more specifically, is are VLANs required to implement traffic >>>> labels? Also, can traffic labels be configured in Basic networking mode or >>>> do I need to switch my configuration to Advanced? >>>> >>>> I am not disagreeing on the how DNS servers should be associated with >>>> interfaces nor do I think a network operator should be required to make any >>>> upstream router configuration changes. I am simply saying that CloudStack >>>> should not make assumptions about the gateways that have been specified. >>>> The behavior I experienced of CloudStack attempting to >>>> "correct" my configuration by injecting another route fails the rule of >>>> least surprise and is based on incomplete knowledge. In my opinion, >>>> CloudStack (or any system of its ilk) should faithfully (or slavishly) >>>> realize the routes on the system VM as specified. If the configuration is >>>> incorrect, networking will fail in an expected manner, and the operator can >>>> adjust their environment as necessary. Otherwise, there is an upstream >>>> router configuration to which CloudStack has no visibility, but with which >>>> it is completely compatible. Essentially, I am asking CloudStack to do >>>> less, assume I know what I am doing, and break in a manner consistent with >>>> other network applications. >>>> >>>> Thanks, >>>> -John >>>> >>>> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote: >>>> >>>>> Traffic labels essentially tell the system which physical network to use. >>>>> So if you've allocated a vlan for a specific traffic type, it will first >>>>> look at the tag associated with that traffic type, figure out which >>>>> physical interface goes with that, and then create a tagged interface and >>>>> bridge also on that physical. >>>>> >>>>> I guess we'll just have to disagree, I think the current behavior makes >>>>> total sense. To me, internal DNS should always use the management >>>>> interface, since it's internally facing. There's no sane way to do that >>>>> other than a static route on the system vm (it seems you're suggesting >>>> that >>>>> the network operator force something like this on the upstream router, >>>>> which seems really strange to require everyone to create static routes on >>>>> their public network to force specific IPs back into their internal >>>>> networks, so correct me if I have the wrong impression). Cloudstack is >>>>> doing exactly what you tell it to. You told it that 10.0.3.2 should be >>>>> accessible via your internal network by setting it as your internal DNS. >>>>> The fact that a broken config doesn't work isn't CloudStack's fault. >>>>> >>>>> Note that internal DNS is just the default for the ssvm, public DNS is >>>>> still offered as a backup, so had you not said that 10.0.3.2 was >>>> available >>>>> on your internal network (perhaps offering a dummy internal DNS >>>>> address or 192.68.56.1), >>>>> lookups would fall back to public and everything would work as expected >>>> as >>>>> well. >>>>> >>>>> There is also a global config called 'use.external.dns', but in setting >>>>> this, restarting the management server, recreating system VMs, I don't >>>> see >>>>> a noticeable difference on any of this, so perhaps that would solve your >>>>> issue as well but it's either broken or doesn't do what I thought it >>>> would. >>>>> >>>>> >>>>> On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote: >>>>> >>>>>> Marcus, >>>>>> >>>>>> Are traffic labels independent of VLANs? I ask because my current XCP >>>>>> network configuration is bridged, and I am not using Open vSwitch. >>>>>> >>>>>> I disagree on the routing issue. CloudStack should do what's told >>>> because >>>>>> it does not have insight into or control of the configuration of the >>>> routes >>>>>> in the layers beneath it. If CloudStack simply did as it was told, it >>>>>> would fail as expected in a typical networking environment while >>>> preserving >>>>>> the flexibility of configuration expected by a network engineer. >>>>>> >>>>>> Thanks, >>>>>> -John >>>>>> >>>>>> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com> >>>> wrote: >>>>>> >>>>>>> I can't really tell you for xen, although it might be similar to KVM. >>>>>>> During setup I would set a traffic label matching the name of the >>>> bridge, >>>>>>> for example if my public interface were eth0 and the bridge I had set >>>> up >>>>>>> was br0, I'd go to the zone network settings, find public traffic, and >>>>>> set >>>>>>> a label on it of "br0". Maybe someone more familiar with the xen setup >>>>>> can >>>>>>> help. >>>>>>> >>>>>>> On the DNS, it makes sense from the perspective that the ssvm has >>>> access >>>>>> to >>>>>>> your internal networks, thus it uses your internal DNS. Its default >>>>>> gateway >>>>>>> is public. So if I have a DNS server on an internal network at >>>>>>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this >>>> route >>>>>>> has to be set in order for the DNS server to be reachable. You would >>>>>> under >>>>>>> normal circumstances not want to use a DNS server on public net as your >>>>>>> internal DNS setting anyway, although I agree that the route insertion >>>>>>> should have a bit more sanity checking and not set a static route to >>>> your >>>>>>> default gateway. >>>>>>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote: >>>>>>> >>>>>>>> Marcus, >>>>>>>> >>>>>>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS >>>>>> for >>>>>>>> the management network to use it, and the route table in the SSVM is >>>> now >>>>>>>> correct. However, this behavior does not seem correct. At a minimum, >>>>>> it >>>>>>>> violates the rule of least surprise. CloudStack shouldn't be adding >>>>>>>> gateways that are not configured. Therefore, I have entered a >>>>>> defect[1] to >>>>>>>> remove the behavior. >>>>>>>> >>>>>>>> With the route table fixed, I am now experiencing a new problem. The >>>>>>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the >>>>>> internal >>>>>>>> NIC (192.168.56.0/24) on the host. The host-only network >>>>>> (192.168.56.15) >>>>>>>> is configured on xenbr0 and the NAT network is configured on xenbr1. >>>>>> As a >>>>>>>> reference, the following is the contents of the >>>> /etc/network/interfaces >>>>>>>> file and ifconfig from devcloud host: >>>>>>>> >>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat >>>>>>>> /etc/network/interfaces >>>>>>>> # The loopback network interface >>>>>>>> auto lo >>>>>>>> iface lo inet loopback >>>>>>>> >>>>>>>> auto eth0 >>>>>>>> iface eth0 inet manual >>>>>>>> >>>>>>>> allow-hotplug eth1 >>>>>>>> iface eth1 inet manual >>>>>>>> >>>>>>>> # The primary network interface >>>>>>>> auto xenbr0 >>>>>>>> iface xenbr0 inet static >>>>>>>> address 192.168.56.15 >>>>>>>> netmask 255.255.255.0 >>>>>>>> network 192.168.56.0 >>>>>>>> broadcast 192.168.56.255 >>>>>>>> dns_nameserver 192.168.56.15 >>>>>>>> bridge_ports eth0 >>>>>>>> >>>>>>>> auto xenbr1 >>>>>>>> iface xenbr1 inet dhcp >>>>>>>> bridge_ports eth1 >>>>>>>> dns_nameserver 8.8.8.8 8.8.4.4 >>>>>>>> post-up route add default gw 10.0.3.2 >>>>>>>> >>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig >>>>>>>> eth0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:777 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:188 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:109977 (109.9 KB) TX bytes:11900 (11.9 KB) >>>>>>>> >>>>>>>> eth1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:478719 (478.7 KB) TX bytes:2542459 (2.5 MB) >>>>>>>> >>>>>>>> lo Link encap:Local Loopback >>>>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>>>> inet6 addr: ::1/128 Scope:Host >>>>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>>>> RX packets:360285 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:0 >>>>>>>> RX bytes:169128181 (169.1 MB) TX bytes:169128181 (169.1 MB) >>>>>>>> >>>>>>>> vif1.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:6 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:152 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:292 (292.0 B) TX bytes:9252 (9.2 KB) >>>>>>>> >>>>>>>> vif1.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:566 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:44227 (44.2 KB) TX bytes:173995 (173.9 KB) >>>>>>>> >>>>>>>> vif1.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:838 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:84 (84.0 B) TX bytes:111361 (111.3 KB) >>>>>>>> >>>>>>>> vif4.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:64 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:197 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:10276 (10.2 KB) TX bytes:18453 (18.4 KB) >>>>>>>> >>>>>>>> vif4.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:2051 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:233914 (233.9 KB) TX bytes:364243 (364.2 KB) >>>>>>>> >>>>>>>> vif4.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:582 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:84 (84.0 B) TX bytes:74700 (74.7 KB) >>>>>>>> >>>>>>>> vif4.3 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>>> RX packets:0 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:585 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:32 >>>>>>>> RX bytes:0 (0.0 B) TX bytes:74826 (74.8 KB) >>>>>>>> >>>>>>>> xapi0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>>> inet addr:169.254.0.1 Bcast:169.254.255.255 Mask:255.255.0.0 >>>>>>>> inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:568 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:0 >>>>>>>> RX bytes:76284 (76.2 KB) TX bytes:109085 (109.0 KB) >>>>>>>> >>>>>>>> xenbr0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>>>> inet addr:192.168.56.15 Bcast:192.168.56.255 >>>>>> Mask:255.255.255.0 >>>>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:4162 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:0 >>>>>>>> RX bytes:469199 (469.1 KB) TX bytes:485688 (485.6 KB) >>>>>>>> >>>>>>>> xenbr1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>>>> inet addr:10.0.3.15 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:0 >>>>>>>> RX bytes:404327 (404.3 KB) TX bytes:2501443 (2.5 MB) >>>>>>>> >>>>>>>> These physical NICs on the host translate to the following Xen PIFs: >>>>>>>> >>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list >>>>>>>> uuid ( RO) : 207413c9-5058-7a40-6c96-2dab21057f30 >>>>>>>> device ( RO): eth1 >>>>>>>> currently-attached ( RO): true >>>>>>>> VLAN ( RO): -1 >>>>>>>> network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72 >>>>>>>> >>>>>>>> >>>>>>>> uuid ( RO) : c0274787-e768-506f-3191-f0ac17b0c72b >>>>>>>> device ( RO): eth0 >>>>>>>> currently-attached ( RO): true >>>>>>>> VLAN ( RO): -1 >>>>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>>>> >>>>>>>> The following is the ifconfig from the SSVM: >>>>>>>> >>>>>>>> root@s-5-TEST:~# ifconfig >>>>>>>> eth0 Link encap:Ethernet HWaddr 0e:00:a9:fe:03:8b >>>>>>>> inet addr:169.254.3.139 Bcast:169.254.255.255 >>>>>> Mask:255.255.0.0 >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:235 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:92 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:21966 (21.4 KiB) TX bytes:16404 (16.0 KiB) >>>>>>>> Interrupt:8 >>>>>>>> >>>>>>>> eth1 Link encap:Ethernet HWaddr 06:bc:62:00:00:05 >>>>>>>> inet addr:192.168.56.104 Bcast:192.168.56.255 >>>>>>>> Mask:255.255.255.0 >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:2532 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:341242 (333.2 KiB) TX bytes:272183 (265.8 KiB) >>>>>>>> Interrupt:10 >>>>>>>> >>>>>>>> eth2 Link encap:Ethernet HWaddr 06:12:72:00:00:37 >>>>>>>> inet addr:10.0.3.204 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:600 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:3 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:68648 (67.0 KiB) TX bytes:126 (126.0 B) >>>>>>>> Interrupt:11 >>>>>>>> >>>>>>>> eth3 Link encap:Ethernet HWaddr 06:25:e2:00:00:15 >>>>>>>> inet addr:192.168.56.120 Bcast:192.168.56.255 >>>>>>>> Mask:255.255.255.0 >>>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>>> RX packets:603 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:1000 >>>>>>>> RX bytes:68732 (67.1 KiB) TX bytes:0 (0.0 B) >>>>>>>> Interrupt:12 >>>>>>>> >>>>>>>> lo Link encap:Local Loopback >>>>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>>>> RX packets:61 errors:0 dropped:0 overruns:0 frame:0 >>>>>>>> TX packets:61 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>>> collisions:0 txqueuelen:0 >>>>>>>> RX bytes:5300 (5.1 KiB) TX bytes:5300 (5.1 KiB) >>>>>>>> >>>>>>>> Finally, the following are the vif params for the eth2 device on the >>>>>> SSVM >>>>>>>> depicting its connection to eth0 instead of eth1: >>>>>>>> >>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243 >>>>>>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210 >>>>>>>> uuid ( RO) : >>>> be44bb30-5700-b461-760e-10fe93079210 >>>>>>>> vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>>>> vm-name-label ( RO): s-5-TEST >>>>>>>> allowed-operations (SRO): attach; unplug_force; unplug >>>>>>>> current-operations (SRO): >>>>>>>> device ( RO): 2 >>>>>>>> MAC ( RO): 06:12:72:00:00:37 >>>>>>>> MAC-autogenerated ( RO): false >>>>>>>> MTU ( RO): 1500 >>>>>>>> currently-attached ( RO): true >>>>>>>> qos_algorithm_type ( RW): ratelimit >>>>>>>> qos_algorithm_params (MRW): kbps: 25600 >>>>>>>> qos_supported_algorithms (SRO): >>>>>>>> other-config (MRW): nicira-iface-id: >>>>>>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id: >>>>>>>> 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>>>> network-name-label ( RO): Pool-wide network associated with >>>>>> eth0 >>>>>>>> io_read_kbs ( RO): 0.007 >>>>>>>> io_write_kbs ( RO): 0.000 >>>>>>>> >>>>>>>> How do I configure CloudStack such that the guest network NIC on the >>>> VM >>>>>>>> will be connected to correct physical NIC? >>>>>>>> >>>>>>>> Thanks for your help, >>>>>>>> -John >>>>>>>> >>>>>>>> >>>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590 >>>>>>>> >>>>>>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com> >>>>>> wrote: >>>>>>>> >>>>>>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the >>>> use >>>>>> of >>>>>>>>> management network to route to 10.0.3.2 for DNS. that's where the >>>> route >>>>>>>> is >>>>>>>>> coming from. you will want to use something on your management net >>>> for >>>>>>>>> internal DNS, or something other than that router. >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com> >>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Anthony, >>>>>>>>>> >>>>>>>>>> I apologize for forgetting to response to the part of your answer >>>> the >>>>>>>>>> first part of the question. I had set the management.network.cidr >>>> and >>>>>>>> host >>>>>>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively. >>>>>>>> Please >>>>>>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my >>>>>> original >>>>>>>>>> email for the actual setting, as well as, the network configurations >>>>>>>> used >>>>>>>>>> when this problem occurs. >>>>>>>>>> >>>>>>>>>> Thanks, >>>>>>>>>> -John >>>>>>>>>> >>>>>>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com> >>>> wrote: >>>>>>>>>> >>>>>>>>>>> Hi join, >>>>>>>>>>> >>>>>>>>>>> Try following, >>>>>>>>>>> >>>>>>>>>>> Set global configuration management.network.cidr to your management >>>>>>>>>> server CIDR, if this configuration is not available in UI, you can >>>>>>>> change >>>>>>>>>> it in DB directly. >>>>>>>>>>> >>>>>>>>>>> Restart management, >>>>>>>>>>> Stop/Start SSVM and CPVM. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> And could you post "cat /proc/cmdline" in SSVM? >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Anthony >>>>>>>>>>> >>>>>>>>>>>> -----Original Message----- >>>>>>>>>>>> From: John Burwell [mailto:jburw...@basho.com] >>>>>>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM >>>>>>>>>>>> To: cloudstack-dev@incubator.apache.org >>>>>>>>>>>> Subject: Re: SSVM Network Configuration Issue >>>>>>>>>>>> >>>>>>>>>>>> All, >>>>>>>>>>>> >>>>>>>>>>>> I was wondering if anyone else is experiencing this problem when >>>>>> using >>>>>>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT >>>>>>>>>>>> adapter. One aspect of this issue that seems interesting is that >>>>>>>>>>>> following route table from the SSVM: >>>>>>>>>>>> >>>>>>>>>>>> root@s-5-TEST:~# route >>>>>>>>>>>> Kernel IP routing table >>>>>>>>>>>> Destination Gateway Genmask Flags Metric Ref >>>>>>>> Use >>>>>>>>>>>> Iface >>>>>>>>>>>> 10.0.3.2 192.168.56.1 255.255.255.255 UGH 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth1 >>>>>>>>>>>> 10.0.3.0 * 255.255.255.0 U 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth2 >>>>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth1 >>>>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth3 >>>>>>>>>>>> link-local * 255.255.0.0 U 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth0 >>>>>>>>>>>> default 10.0.3.2 0.0.0.0 UG 0 0 >>>>>>>> 0 >>>>>>>>>>>> eth2 >>>>>>>>>>>> >>>>>>>>>>>> In particular, the gateways for the management and guest networks >>>> do >>>>>>>>>>>> not match to the configuration provided to the management server >>>>>> (i.e. >>>>>>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and >>>>>> 192.168.56.1 >>>>>>>> is >>>>>>>>>>>> the gateway for the 192.168.56.0/24 network). With this >>>>>>>> configuration, >>>>>>>>>>>> the SSVM has a socket connection to the management server, but is >>>> in >>>>>>>>>>>> alert state. Finally, when I remove the host-only NIC and use >>>> only >>>>>> a >>>>>>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to >>>>>>>>>>>> believe that the segregated network configuration is at the root >>>> of >>>>>>>> the >>>>>>>>>>>> problem. >>>>>>>>>>>> >>>>>>>>>>>> Until I can get the networking on the SSVM configured, I am unable >>>>>> to >>>>>>>>>>>> complete the testing of the S3-backed Secondary Storage >>>> enhancement. >>>>>>>>>>>> >>>>>>>>>>>> Thank you for your help, >>>>>>>>>>>> -John >>>>>>>>>>>> >>>>>>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com> >>>>>> wrote: >>>>>>>>>>>> >>>>>>>>>>>>> All, >>>>>>>>>>>>> >>>>>>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox >>>>>>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1. I have configured the >>>>>>>> base >>>>>>>>>>>> management server VM (zone1) to serve as both the zone1, as well >>>> as, >>>>>>>>>>>> the management server (running MySql) with eth0 as a host-only >>>>>> adapter >>>>>>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see >>>> the >>>>>>>>>>>> attached zone1-interfaces file for the exact network configuration >>>>>> on >>>>>>>>>>>> the VM). The management and guest networks are configured as >>>>>> follows: >>>>>>>>>>>>> >>>>>>>>>>>>> Zone 1 >>>>>>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?) >>>>>>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>>>> Zone 2 >>>>>>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?) >>>>>>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>>>> >>>>>>>>>>>>> The management server deploys and starts without error. I then >>>>>>>>>>>> populate the configuration it using the attached Marvin >>>>>> configuration >>>>>>>>>>>> file (zone1.devcloud.cfg) and restart the management server in >>>> order >>>>>>>> to >>>>>>>>>>>> allow the global configuration option changes to take effect. >>>>>>>>>>>> Following the restart, the CPVM and SSVM start without error. >>>>>>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable >>>>>> to >>>>>>>>>>>> connect outbound through the guest network (very important for my >>>>>>>> tests >>>>>>>>>>>> because I am testing S3-backed secondary storage). >>>>>>>>>>>>> >>>>>>>>>>>>> From the diagnostic checks I have performed on the management >>>>>> server >>>>>>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting >>>>>>>> back >>>>>>>>>>>> to the management server. I have attached a set of diagnostic >>>>>>>>>>>> information from the management server >>>>>> (mgmtsvr-zone1-diagnostics.log) >>>>>>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the >>>>>> results >>>>>>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other >>>>>>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the >>>>>>>> SSVM). >>>>>>>>>>>> Finally, I have attached the vmops log from the management server >>>>>>>>>>>> (vmops-zone1.log). >>>>>>>>>>>>> >>>>>>>>>>>>> What changes need to be made to management server configuration >>>> in >>>>>>>>>>>> order to start up an SSVM that can communicate with the secondary >>>>>>>>>>>> storage NFS volumes, management server, and connect to hosts on >>>> the >>>>>>>>>>>> Internet? >>>>>>>>>>>>> >>>>>>>>>>>>> Thanks for your help, >>>>>>>>>>>>> -John >>>>>>>>>>>>> >>>>>>>>>>>>> <ssvm-zone1-diagnostics.log> >>>>>>>>>>>>> <vmops-zone1.tar.gz> >>>>>>>>>>>>> <mgmtsvr-zone1-diagnostics.log> >>>>>>>>>>>>> <zone1-interfaces> >>>>>>>>>>>>> <zone1.devcloud.cfg> >>>> >>>>