Rohit,

The routes are not the issue.  Both SSVM VIFs are being bound to the
host-only NIC. VirtualBox explicitly prevents routing between
host-only and NAT NICs in its virtual switch.  Therefore, the routing
configuration below will not allow Internet access if the VIFs are not
bound to the proper physical NIC.

Thanks,
-John




On Dec 15, 2012, at 9:35 PM, Rohit Yadav <rohit.ya...@citrix.com> wrote:

> Hi John,
>
> I think I've the same issue you're facing.
> The systemvms and other instances running on top of xen running on 
> hostonly/nat on vbox (devcloud like vm) won't be able to reach the Internet.
> To fix the issue for dom0, I configured my network to route all traffic 
> through nat and make it the default gateway.
>
> auto xenbr0
> iface xenbr0 inet static
>        bridge_ports eth0
>        address 192.168.56.10
>        netmask 255.255.255.0
>        network 192.168.56.0
>        broadcast 192.168.56.255
>        gateway 192.168.56.1
>        dns_nameservers 8.8.8.8 8.8.4.4
>        post-up route del default gw 192.168.56.1; route add default gw 
> 192.168.56.1 metric 100;
>
> auto xenbr1
> iface xenbr1 inet dhcp
>        bridge_ports eth1
>        dns_nameservers 8.8.8.8 8.8.4.4
>        post-up route add default gw 10.0.3.2
>
> This fixed the problem on dom0, but I'm not sure what to do about the 
> systemvms and instances. May be you can try to swap nics such that nic1 is 
> nat and nic2 is host-only.
> The problem I found using tcpdump was that only arp packets were going 
> through. So, maybe by enabling promiscuous networking mode on xen (just like 
> that was done on vbox) may fix this, but I've never tried that on devcloud.
>
> Regards.
> ________________________________________
> From: John Burwell [jburw...@basho.com]
> Sent: Sunday, December 16, 2012 2:55 AM
> To: cloudstack-dev@incubator.apache.org
> Subject: Re: SSVM Network Configuration Issue
>
> Rohit,
>
> As I stated below, I know which the VIF->PIF->device Xen mapping needs
> to be used for each network.  My question is how do I configure
> CloudStack with that information.
>
> Thanks for your help,
> -John
>
>
>
>
> On Dec 15, 2012, at 2:51 PM, Rohit Yadav <rohit.ya...@citrix.com> wrote:
>
>> About xen bridging, these can help;
>>
>> brctl show all (bridge vif mappings)
>> ip show addr xenbr0 (bridge specific info)
>> brctl showmacs br0 (bridge mac mappings)
>>
>> Wiki:
>> http://wiki.xen.org/wiki/Xen_FAQ_Networking
>> http://wiki.xen.org/wiki/XenNetworking
>>
>> Regards.
>>
>> ________________________________________
>> From: John Burwell [jburw...@basho.com]
>> Sent: Saturday, December 15, 2012 8:35 PM
>> To: cloudstack-dev@incubator.apache.org
>> Subject: Re: SSVM Network Configuration Issue
>>
>> Marcus,
>>
>> That's what I thought.  The Xen physical bridge names are xenbr0 (to
>> eth0) and xenbr1 (to eth1).  Using basic network configuration, I set
>> the Xen network traffic labels for each to the appropriate bridge
>> device name.  I receive errors regarding invalid network device when
>> it attempts to create a VM.    Does anyone else know how determine the
>> mapping of physical devices to CloudStack Xen network traffic labels?
>>
>> Thanks,
>> -John
>>
>> On Dec 15, 2012, at 1:20 AM, Marcus Sorensen <shadow...@gmail.com> wrote:
>>
>>> Vlans in advanced/KVM should only be required for the guest networks. If I
>>> create a bridge on physical eth0, name it 'br0', and a bridge on physical
>>> eth1, naming it 'br1', and then set my management network to label 'br0',
>>> my public network to 'br1', and my guest network to 'br1', it should use
>>> the bridges you asked for when connecting the system VMs for the specified
>>> traffic. I'd just leave the 'vlan' blank when specifying public and
>>> pod(management) IPs. In this scenario, the only place you need to enter
>>> vlans is on the guest, and it should create new tagged interfaces/bridges
>>> on eth1(per your label of br1) as new guest networks are brought online.
>>> This is how my dev VMs are usually set up.
>>>
>>>
>>> On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote:
>>>
>>>> Marcus,
>>>>
>>>> My question, more specifically, is are VLANs required to implement traffic
>>>> labels?  Also, can traffic labels be configured in Basic networking mode or
>>>> do I need to switch my configuration to Advanced?
>>>>
>>>> I am not disagreeing on the how DNS servers should be associated with
>>>> interfaces nor do I think a network operator should be required to make any
>>>> upstream router configuration changes.  I am simply saying that CloudStack
>>>> should not make assumptions about the gateways that have been specified.
>>>> The behavior I experienced of CloudStack attempting to
>>>> "correct" my configuration by injecting another route fails the rule of
>>>> least surprise and is based on incomplete knowledge.  In my opinion,
>>>> CloudStack (or any system of its ilk) should faithfully (or slavishly)
>>>> realize the routes on the system VM as specified.  If the configuration is
>>>> incorrect, networking will fail in an expected manner, and the operator can
>>>> adjust their environment as necessary.   Otherwise, there is an upstream
>>>> router configuration to which CloudStack has no visibility, but with which
>>>> it is completely compatible.  Essentially, I am asking CloudStack to do
>>>> less, assume I know what I am doing, and break in a manner consistent with
>>>> other network applications.
>>>>
>>>> Thanks,
>>>> -John
>>>>
>>>> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote:
>>>>
>>>>> Traffic labels essentially tell the system which physical network to use.
>>>>> So if you've allocated a vlan for a specific traffic type, it will first
>>>>> look at the tag associated with that traffic type, figure out which
>>>>> physical interface goes with that, and then create a tagged interface and
>>>>> bridge also on that physical.
>>>>>
>>>>> I guess we'll just have to disagree, I think the current behavior makes
>>>>> total sense.  To me, internal DNS should always use the management
>>>>> interface, since it's internally facing. There's no sane way to do that
>>>>> other than a static route on the system vm (it seems you're suggesting
>>>> that
>>>>> the network operator force something like this on the upstream router,
>>>>> which seems really strange to require everyone to create static routes on
>>>>> their public network to force specific IPs back into their internal
>>>>> networks, so correct me if I have the wrong impression).  Cloudstack is
>>>>> doing exactly what you tell it to. You told it that 10.0.3.2 should be
>>>>> accessible via your internal network by setting it as your internal DNS.
>>>>> The fact that a broken config doesn't work isn't CloudStack's fault.
>>>>>
>>>>> Note that internal DNS is just the default for the ssvm, public DNS is
>>>>> still offered as a backup, so had you not said that 10.0.3.2 was
>>>> available
>>>>> on your internal network (perhaps offering a dummy internal DNS
>>>>> address or 192.68.56.1),
>>>>> lookups would fall back to public and everything would work as expected
>>>> as
>>>>> well.
>>>>>
>>>>> There is also a global config called 'use.external.dns', but in setting
>>>>> this, restarting the management server, recreating system VMs, I don't
>>>> see
>>>>> a noticeable difference on any of this, so perhaps that would solve your
>>>>> issue as well but it's either broken or doesn't do what I thought it
>>>> would.
>>>>>
>>>>>
>>>>> On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote:
>>>>>
>>>>>> Marcus,
>>>>>>
>>>>>> Are traffic labels independent of VLANs?  I ask because my current XCP
>>>>>> network configuration is bridged, and I am not using Open vSwitch.
>>>>>>
>>>>>> I disagree on the routing issue.  CloudStack should do what's told
>>>> because
>>>>>> it does not have insight into or control of the configuration of the
>>>> routes
>>>>>> in the layers beneath it.  If CloudStack simply did as it was told, it
>>>>>> would fail as expected in a typical networking environment while
>>>> preserving
>>>>>> the flexibility of configuration expected by a network engineer.
>>>>>>
>>>>>> Thanks,
>>>>>> -John
>>>>>>
>>>>>> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com>
>>>> wrote:
>>>>>>
>>>>>>> I can't really tell you for xen, although it might be similar to KVM.
>>>>>>> During setup I would set a traffic label matching the name of the
>>>> bridge,
>>>>>>> for example if my public interface were eth0 and the bridge I had set
>>>> up
>>>>>>> was br0, I'd go to the zone network settings, find public traffic, and
>>>>>> set
>>>>>>> a label on it of "br0". Maybe someone more familiar with the xen setup
>>>>>> can
>>>>>>> help.
>>>>>>>
>>>>>>> On the DNS, it makes sense from the perspective that the ssvm has
>>>> access
>>>>>> to
>>>>>>> your internal networks, thus it uses your internal DNS. Its default
>>>>>> gateway
>>>>>>> is public. So if I have a DNS server on an internal network at
>>>>>>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this
>>>> route
>>>>>>> has to be set in order for the DNS server to be reachable. You would
>>>>>> under
>>>>>>> normal circumstances not want to use a DNS server on public net as your
>>>>>>> internal DNS setting anyway, although I agree that the route insertion
>>>>>>> should have a bit more sanity checking and not set a static route to
>>>> your
>>>>>>> default gateway.
>>>>>>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote:
>>>>>>>
>>>>>>>> Marcus,
>>>>>>>>
>>>>>>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS
>>>>>> for
>>>>>>>> the management network to use it, and the route table in the SSVM is
>>>> now
>>>>>>>> correct.  However, this behavior does not seem correct.  At a minimum,
>>>>>> it
>>>>>>>> violates the rule of least surprise.  CloudStack shouldn't be adding
>>>>>>>> gateways that are not configured.  Therefore, I have entered a
>>>>>> defect[1] to
>>>>>>>> remove the behavior.
>>>>>>>>
>>>>>>>> With the route table fixed, I am now experiencing a new problem.  The
>>>>>>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the
>>>>>> internal
>>>>>>>> NIC (192.168.56.0/24) on the host.  The host-only network
>>>>>> (192.168.56.15)
>>>>>>>> is configured on xenbr0 and the NAT network is configured on xenbr1.
>>>>>> As a
>>>>>>>> reference, the following is the contents of the
>>>> /etc/network/interfaces
>>>>>>>> file and ifconfig from devcloud host:
>>>>>>>>
>>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat
>>>>>>>> /etc/network/interfaces
>>>>>>>> # The loopback network interface
>>>>>>>> auto lo
>>>>>>>> iface lo inet loopback
>>>>>>>>
>>>>>>>> auto eth0
>>>>>>>> iface eth0 inet manual
>>>>>>>>
>>>>>>>> allow-hotplug eth1
>>>>>>>> iface eth1 inet manual
>>>>>>>>
>>>>>>>> # The primary network interface
>>>>>>>> auto xenbr0
>>>>>>>> iface xenbr0 inet static
>>>>>>>> address 192.168.56.15
>>>>>>>> netmask 255.255.255.0
>>>>>>>> network 192.168.56.0
>>>>>>>> broadcast 192.168.56.255
>>>>>>>> dns_nameserver 192.168.56.15
>>>>>>>> bridge_ports eth0
>>>>>>>>
>>>>>>>> auto xenbr1
>>>>>>>> iface xenbr1 inet dhcp
>>>>>>>> bridge_ports eth1
>>>>>>>> dns_nameserver 8.8.8.8 8.8.4.4
>>>>>>>> post-up route add default gw 10.0.3.2
>>>>>>>>
>>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig
>>>>>>>> eth0      Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
>>>>>>>>      inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:777 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:188 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:109977 (109.9 KB)  TX bytes:11900 (11.9 KB)
>>>>>>>>
>>>>>>>> eth1      Link encap:Ethernet  HWaddr 08:00:27:df:00:00
>>>>>>>>      inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:478719 (478.7 KB)  TX bytes:2542459 (2.5 MB)
>>>>>>>>
>>>>>>>> lo        Link encap:Local Loopback
>>>>>>>>      inet addr:127.0.0.1  Mask:255.0.0.0
>>>>>>>>      inet6 addr: ::1/128 Scope:Host
>>>>>>>>      UP LOOPBACK RUNNING  MTU:16436  Metric:1
>>>>>>>>      RX packets:360285 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:0
>>>>>>>>      RX bytes:169128181 (169.1 MB)  TX bytes:169128181 (169.1 MB)
>>>>>>>>
>>>>>>>> vif1.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:6 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:152 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:292 (292.0 B)  TX bytes:9252 (9.2 KB)
>>>>>>>>
>>>>>>>> vif1.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:566 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:44227 (44.2 KB)  TX bytes:173995 (173.9 KB)
>>>>>>>>
>>>>>>>> vif1.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:3 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:838 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:84 (84.0 B)  TX bytes:111361 (111.3 KB)
>>>>>>>>
>>>>>>>> vif4.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:64 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:197 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:10276 (10.2 KB)  TX bytes:18453 (18.4 KB)
>>>>>>>>
>>>>>>>> vif4.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:2051 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:233914 (233.9 KB)  TX bytes:364243 (364.2 KB)
>>>>>>>>
>>>>>>>> vif4.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:3 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:582 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:84 (84.0 B)  TX bytes:74700 (74.7 KB)
>>>>>>>>
>>>>>>>> vif4.3    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
>>>>>>>>      RX packets:0 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:585 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:32
>>>>>>>>      RX bytes:0 (0.0 B)  TX bytes:74826 (74.8 KB)
>>>>>>>>
>>>>>>>> xapi0     Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
>>>>>>>>      inet addr:169.254.0.1  Bcast:169.254.255.255  Mask:255.255.0.0
>>>>>>>>      inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:568 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:0
>>>>>>>>      RX bytes:76284 (76.2 KB)  TX bytes:109085 (109.0 KB)
>>>>>>>>
>>>>>>>> xenbr0    Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
>>>>>>>>      inet addr:192.168.56.15  Bcast:192.168.56.255
>>>>>> Mask:255.255.255.0
>>>>>>>>      inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:4162 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:0
>>>>>>>>      RX bytes:469199 (469.1 KB)  TX bytes:485688 (485.6 KB)
>>>>>>>>
>>>>>>>> xenbr1    Link encap:Ethernet  HWaddr 08:00:27:df:00:00
>>>>>>>>      inet addr:10.0.3.15  Bcast:10.0.3.255  Mask:255.255.255.0
>>>>>>>>      inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:0
>>>>>>>>      RX bytes:404327 (404.3 KB)  TX bytes:2501443 (2.5 MB)
>>>>>>>>
>>>>>>>> These physical NICs on the host translate to the following Xen PIFs:
>>>>>>>>
>>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list
>>>>>>>> uuid ( RO)                  : 207413c9-5058-7a40-6c96-2dab21057f30
>>>>>>>>            device ( RO): eth1
>>>>>>>> currently-attached ( RO): true
>>>>>>>>              VLAN ( RO): -1
>>>>>>>>      network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72
>>>>>>>>
>>>>>>>>
>>>>>>>> uuid ( RO)                  : c0274787-e768-506f-3191-f0ac17b0c72b
>>>>>>>>            device ( RO): eth0
>>>>>>>> currently-attached ( RO): true
>>>>>>>>              VLAN ( RO): -1
>>>>>>>>      network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
>>>>>>>>
>>>>>>>> The following is the ifconfig from the SSVM:
>>>>>>>>
>>>>>>>> root@s-5-TEST:~# ifconfig
>>>>>>>> eth0      Link encap:Ethernet  HWaddr 0e:00:a9:fe:03:8b
>>>>>>>>      inet addr:169.254.3.139  Bcast:169.254.255.255
>>>>>> Mask:255.255.0.0
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:235 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:92 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:21966 (21.4 KiB)  TX bytes:16404 (16.0 KiB)
>>>>>>>>      Interrupt:8
>>>>>>>>
>>>>>>>> eth1      Link encap:Ethernet  HWaddr 06:bc:62:00:00:05
>>>>>>>>      inet addr:192.168.56.104  Bcast:192.168.56.255
>>>>>>>> Mask:255.255.255.0
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:2532 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:341242 (333.2 KiB)  TX bytes:272183 (265.8 KiB)
>>>>>>>>      Interrupt:10
>>>>>>>>
>>>>>>>> eth2      Link encap:Ethernet  HWaddr 06:12:72:00:00:37
>>>>>>>>      inet addr:10.0.3.204  Bcast:10.0.3.255  Mask:255.255.255.0
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:600 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:3 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:68648 (67.0 KiB)  TX bytes:126 (126.0 B)
>>>>>>>>      Interrupt:11
>>>>>>>>
>>>>>>>> eth3      Link encap:Ethernet  HWaddr 06:25:e2:00:00:15
>>>>>>>>      inet addr:192.168.56.120  Bcast:192.168.56.255
>>>>>>>> Mask:255.255.255.0
>>>>>>>>      UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
>>>>>>>>      RX packets:603 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:1000
>>>>>>>>      RX bytes:68732 (67.1 KiB)  TX bytes:0 (0.0 B)
>>>>>>>>      Interrupt:12
>>>>>>>>
>>>>>>>> lo        Link encap:Local Loopback
>>>>>>>>      inet addr:127.0.0.1  Mask:255.0.0.0
>>>>>>>>      UP LOOPBACK RUNNING  MTU:16436  Metric:1
>>>>>>>>      RX packets:61 errors:0 dropped:0 overruns:0 frame:0
>>>>>>>>      TX packets:61 errors:0 dropped:0 overruns:0 carrier:0
>>>>>>>>      collisions:0 txqueuelen:0
>>>>>>>>      RX bytes:5300 (5.1 KiB)  TX bytes:5300 (5.1 KiB)
>>>>>>>>
>>>>>>>> Finally, the following are the vif params for the eth2 device on the
>>>>>> SSVM
>>>>>>>> depicting its connection to eth0 instead of eth1:
>>>>>>>>
>>>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243
>>>>>>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210
>>>>>>>> uuid ( RO)                        :
>>>> be44bb30-5700-b461-760e-10fe93079210
>>>>>>>>                 vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7
>>>>>>>>           vm-name-label ( RO): s-5-TEST
>>>>>>>>      allowed-operations (SRO): attach; unplug_force; unplug
>>>>>>>>      current-operations (SRO):
>>>>>>>>                  device ( RO): 2
>>>>>>>>                     MAC ( RO): 06:12:72:00:00:37
>>>>>>>>       MAC-autogenerated ( RO): false
>>>>>>>>                     MTU ( RO): 1500
>>>>>>>>      currently-attached ( RO): true
>>>>>>>>      qos_algorithm_type ( RW): ratelimit
>>>>>>>>    qos_algorithm_params (MRW): kbps: 25600
>>>>>>>> qos_supported_algorithms (SRO):
>>>>>>>>            other-config (MRW): nicira-iface-id:
>>>>>>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id:
>>>>>>>> 7958d91f-e52d-a25d-718c-7f831ae701d7
>>>>>>>>            network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
>>>>>>>>      network-name-label ( RO): Pool-wide network associated with
>>>>>> eth0
>>>>>>>>             io_read_kbs ( RO): 0.007
>>>>>>>>            io_write_kbs ( RO): 0.000
>>>>>>>>
>>>>>>>> How do I configure CloudStack such that the guest network NIC on the
>>>> VM
>>>>>>>> will be connected to correct physical NIC?
>>>>>>>>
>>>>>>>> Thanks for your help,
>>>>>>>> -John
>>>>>>>>
>>>>>>>>
>>>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590
>>>>>>>>
>>>>>>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com>
>>>>>> wrote:
>>>>>>>>
>>>>>>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the
>>>> use
>>>>>> of
>>>>>>>>> management network to route to 10.0.3.2 for DNS. that's where the
>>>> route
>>>>>>>> is
>>>>>>>>> coming from. you will want to use something on your management net
>>>> for
>>>>>>>>> internal DNS, or something other than that router.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com>
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> Anthony,
>>>>>>>>>>
>>>>>>>>>> I apologize for forgetting to response to the part of your answer
>>>> the
>>>>>>>>>> first part of the question.  I had set the management.network.cidr
>>>> and
>>>>>>>> host
>>>>>>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively.
>>>>>>>> Please
>>>>>>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my
>>>>>> original
>>>>>>>>>> email for the actual setting, as well as, the network configurations
>>>>>>>> used
>>>>>>>>>> when this problem occurs.
>>>>>>>>>>
>>>>>>>>>> Thanks,
>>>>>>>>>> -John
>>>>>>>>>>
>>>>>>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com>
>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hi join,
>>>>>>>>>>>
>>>>>>>>>>> Try following,
>>>>>>>>>>>
>>>>>>>>>>> Set global configuration management.network.cidr to your management
>>>>>>>>>> server CIDR, if this configuration is not available in UI, you can
>>>>>>>> change
>>>>>>>>>> it in DB directly.
>>>>>>>>>>>
>>>>>>>>>>> Restart management,
>>>>>>>>>>> Stop/Start SSVM and CPVM.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> And could you post "cat /proc/cmdline" in SSVM?
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Anthony
>>>>>>>>>>>
>>>>>>>>>>>> -----Original Message-----
>>>>>>>>>>>> From: John Burwell [mailto:jburw...@basho.com]
>>>>>>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM
>>>>>>>>>>>> To: cloudstack-dev@incubator.apache.org
>>>>>>>>>>>> Subject: Re: SSVM Network Configuration Issue
>>>>>>>>>>>>
>>>>>>>>>>>> All,
>>>>>>>>>>>>
>>>>>>>>>>>> I was wondering if anyone else is experiencing this problem when
>>>>>> using
>>>>>>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT
>>>>>>>>>>>> adapter.  One aspect of this issue that seems interesting is that
>>>>>>>>>>>> following route table from the SSVM:
>>>>>>>>>>>>
>>>>>>>>>>>> root@s-5-TEST:~# route
>>>>>>>>>>>> Kernel IP routing table
>>>>>>>>>>>> Destination     Gateway         Genmask         Flags Metric Ref
>>>>>>>> Use
>>>>>>>>>>>> Iface
>>>>>>>>>>>> 10.0.3.2        192.168.56.1    255.255.255.255 UGH   0      0
>>>>>>>> 0
>>>>>>>>>>>> eth1
>>>>>>>>>>>> 10.0.3.0        *               255.255.255.0   U     0      0
>>>>>>>> 0
>>>>>>>>>>>> eth2
>>>>>>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
>>>>>>>> 0
>>>>>>>>>>>> eth1
>>>>>>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
>>>>>>>> 0
>>>>>>>>>>>> eth3
>>>>>>>>>>>> link-local      *               255.255.0.0     U     0      0
>>>>>>>> 0
>>>>>>>>>>>> eth0
>>>>>>>>>>>> default         10.0.3.2        0.0.0.0         UG    0      0
>>>>>>>> 0
>>>>>>>>>>>> eth2
>>>>>>>>>>>>
>>>>>>>>>>>> In particular, the gateways for the management and guest networks
>>>> do
>>>>>>>>>>>> not match to the configuration provided to the management server
>>>>>> (i.e.
>>>>>>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and
>>>>>> 192.168.56.1
>>>>>>>> is
>>>>>>>>>>>> the gateway for the 192.168.56.0/24 network).  With this
>>>>>>>> configuration,
>>>>>>>>>>>> the SSVM has a socket connection to the management server, but is
>>>> in
>>>>>>>>>>>> alert state.  Finally, when I remove the host-only NIC and use
>>>> only
>>>>>> a
>>>>>>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to
>>>>>>>>>>>> believe that the segregated network configuration is at the root
>>>> of
>>>>>>>> the
>>>>>>>>>>>> problem.
>>>>>>>>>>>>
>>>>>>>>>>>> Until I can get the networking on the SSVM configured, I am unable
>>>>>> to
>>>>>>>>>>>> complete the testing of the S3-backed Secondary Storage
>>>> enhancement.
>>>>>>>>>>>>
>>>>>>>>>>>> Thank you for your help,
>>>>>>>>>>>> -John
>>>>>>>>>>>>
>>>>>>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com>
>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> All,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox
>>>>>>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1.  I have configured the
>>>>>>>> base
>>>>>>>>>>>> management server VM (zone1) to serve as both the zone1, as well
>>>> as,
>>>>>>>>>>>> the management server (running MySql) with eth0 as a host-only
>>>>>> adapter
>>>>>>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see
>>>> the
>>>>>>>>>>>> attached zone1-interfaces file for the exact network configuration
>>>>>> on
>>>>>>>>>>>> the VM).  The management and guest networks are configured as
>>>>>> follows:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Zone 1
>>>>>>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?)
>>>>>>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8
>>>>>>>>>>>>> Zone 2
>>>>>>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?)
>>>>>>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8
>>>>>>>>>>>>>
>>>>>>>>>>>>> The management server deploys and starts without error.  I then
>>>>>>>>>>>> populate the configuration it using the attached Marvin
>>>>>> configuration
>>>>>>>>>>>> file (zone1.devcloud.cfg) and restart the management server in
>>>> order
>>>>>>>> to
>>>>>>>>>>>> allow the global configuration option changes to take effect.
>>>>>>>>>>>> Following the restart, the CPVM and SSVM start without error.
>>>>>>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable
>>>>>> to
>>>>>>>>>>>> connect outbound through the guest network (very important for my
>>>>>>>> tests
>>>>>>>>>>>> because I am testing S3-backed secondary storage).
>>>>>>>>>>>>>
>>>>>>>>>>>>> From the diagnostic checks I have performed on the management
>>>>>> server
>>>>>>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting
>>>>>>>> back
>>>>>>>>>>>> to the management server.  I have attached a set of diagnostic
>>>>>>>>>>>> information from the management server
>>>>>> (mgmtsvr-zone1-diagnostics.log)
>>>>>>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the
>>>>>> results
>>>>>>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other
>>>>>>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the
>>>>>>>> SSVM).
>>>>>>>>>>>> Finally, I have attached the vmops log from the management server
>>>>>>>>>>>> (vmops-zone1.log).
>>>>>>>>>>>>>
>>>>>>>>>>>>> What changes need to be made to management server configuration
>>>> in
>>>>>>>>>>>> order to start up an SSVM that can communicate with the secondary
>>>>>>>>>>>> storage NFS volumes, management server, and connect to hosts on
>>>> the
>>>>>>>>>>>> Internet?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks for your help,
>>>>>>>>>>>>> -John
>>>>>>>>>>>>>
>>>>>>>>>>>>> <ssvm-zone1-diagnostics.log>
>>>>>>>>>>>>> <vmops-zone1.tar.gz>
>>>>>>>>>>>>> <mgmtsvr-zone1-diagnostics.log>
>>>>>>>>>>>>> <zone1-interfaces>
>>>>>>>>>>>>> <zone1.devcloud.cfg>
>>>>
>>>>

Reply via email to