Marcus, That's what I thought. The Xen physical bridge names are xenbr0 (to eth0) and xenbr1 (to eth1). Using basic network configuration, I set the Xen network traffic labels for each to the appropriate bridge device name. I receive errors regarding invalid network device when it attempts to create a VM. Does anyone else know how determine the mapping of physical devices to CloudStack Xen network traffic labels?
Thanks, -John On Dec 15, 2012, at 1:20 AM, Marcus Sorensen <shadow...@gmail.com> wrote: > Vlans in advanced/KVM should only be required for the guest networks. If I > create a bridge on physical eth0, name it 'br0', and a bridge on physical > eth1, naming it 'br1', and then set my management network to label 'br0', > my public network to 'br1', and my guest network to 'br1', it should use > the bridges you asked for when connecting the system VMs for the specified > traffic. I'd just leave the 'vlan' blank when specifying public and > pod(management) IPs. In this scenario, the only place you need to enter > vlans is on the guest, and it should create new tagged interfaces/bridges > on eth1(per your label of br1) as new guest networks are brought online. > This is how my dev VMs are usually set up. > > > On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote: > >> Marcus, >> >> My question, more specifically, is are VLANs required to implement traffic >> labels? Also, can traffic labels be configured in Basic networking mode or >> do I need to switch my configuration to Advanced? >> >> I am not disagreeing on the how DNS servers should be associated with >> interfaces nor do I think a network operator should be required to make any >> upstream router configuration changes. I am simply saying that CloudStack >> should not make assumptions about the gateways that have been specified. >> The behavior I experienced of CloudStack attempting to >> "correct" my configuration by injecting another route fails the rule of >> least surprise and is based on incomplete knowledge. In my opinion, >> CloudStack (or any system of its ilk) should faithfully (or slavishly) >> realize the routes on the system VM as specified. If the configuration is >> incorrect, networking will fail in an expected manner, and the operator can >> adjust their environment as necessary. Otherwise, there is an upstream >> router configuration to which CloudStack has no visibility, but with which >> it is completely compatible. Essentially, I am asking CloudStack to do >> less, assume I know what I am doing, and break in a manner consistent with >> other network applications. >> >> Thanks, >> -John >> >> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote: >> >>> Traffic labels essentially tell the system which physical network to use. >>> So if you've allocated a vlan for a specific traffic type, it will first >>> look at the tag associated with that traffic type, figure out which >>> physical interface goes with that, and then create a tagged interface and >>> bridge also on that physical. >>> >>> I guess we'll just have to disagree, I think the current behavior makes >>> total sense. To me, internal DNS should always use the management >>> interface, since it's internally facing. There's no sane way to do that >>> other than a static route on the system vm (it seems you're suggesting >> that >>> the network operator force something like this on the upstream router, >>> which seems really strange to require everyone to create static routes on >>> their public network to force specific IPs back into their internal >>> networks, so correct me if I have the wrong impression). Cloudstack is >>> doing exactly what you tell it to. You told it that 10.0.3.2 should be >>> accessible via your internal network by setting it as your internal DNS. >>> The fact that a broken config doesn't work isn't CloudStack's fault. >>> >>> Note that internal DNS is just the default for the ssvm, public DNS is >>> still offered as a backup, so had you not said that 10.0.3.2 was >> available >>> on your internal network (perhaps offering a dummy internal DNS >>> address or 192.68.56.1), >>> lookups would fall back to public and everything would work as expected >> as >>> well. >>> >>> There is also a global config called 'use.external.dns', but in setting >>> this, restarting the management server, recreating system VMs, I don't >> see >>> a noticeable difference on any of this, so perhaps that would solve your >>> issue as well but it's either broken or doesn't do what I thought it >> would. >>> >>> >>> On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote: >>> >>>> Marcus, >>>> >>>> Are traffic labels independent of VLANs? I ask because my current XCP >>>> network configuration is bridged, and I am not using Open vSwitch. >>>> >>>> I disagree on the routing issue. CloudStack should do what's told >> because >>>> it does not have insight into or control of the configuration of the >> routes >>>> in the layers beneath it. If CloudStack simply did as it was told, it >>>> would fail as expected in a typical networking environment while >> preserving >>>> the flexibility of configuration expected by a network engineer. >>>> >>>> Thanks, >>>> -John >>>> >>>> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com> >> wrote: >>>> >>>>> I can't really tell you for xen, although it might be similar to KVM. >>>>> During setup I would set a traffic label matching the name of the >> bridge, >>>>> for example if my public interface were eth0 and the bridge I had set >> up >>>>> was br0, I'd go to the zone network settings, find public traffic, and >>>> set >>>>> a label on it of "br0". Maybe someone more familiar with the xen setup >>>> can >>>>> help. >>>>> >>>>> On the DNS, it makes sense from the perspective that the ssvm has >> access >>>> to >>>>> your internal networks, thus it uses your internal DNS. Its default >>>> gateway >>>>> is public. So if I have a DNS server on an internal network at >>>>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this >> route >>>>> has to be set in order for the DNS server to be reachable. You would >>>> under >>>>> normal circumstances not want to use a DNS server on public net as your >>>>> internal DNS setting anyway, although I agree that the route insertion >>>>> should have a bit more sanity checking and not set a static route to >> your >>>>> default gateway. >>>>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote: >>>>> >>>>>> Marcus, >>>>>> >>>>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS >>>> for >>>>>> the management network to use it, and the route table in the SSVM is >> now >>>>>> correct. However, this behavior does not seem correct. At a minimum, >>>> it >>>>>> violates the rule of least surprise. CloudStack shouldn't be adding >>>>>> gateways that are not configured. Therefore, I have entered a >>>> defect[1] to >>>>>> remove the behavior. >>>>>> >>>>>> With the route table fixed, I am now experiencing a new problem. The >>>>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the >>>> internal >>>>>> NIC (192.168.56.0/24) on the host. The host-only network >>>> (192.168.56.15) >>>>>> is configured on xenbr0 and the NAT network is configured on xenbr1. >>>> As a >>>>>> reference, the following is the contents of the >> /etc/network/interfaces >>>>>> file and ifconfig from devcloud host: >>>>>> >>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat >>>>>> /etc/network/interfaces >>>>>> # The loopback network interface >>>>>> auto lo >>>>>> iface lo inet loopback >>>>>> >>>>>> auto eth0 >>>>>> iface eth0 inet manual >>>>>> >>>>>> allow-hotplug eth1 >>>>>> iface eth1 inet manual >>>>>> >>>>>> # The primary network interface >>>>>> auto xenbr0 >>>>>> iface xenbr0 inet static >>>>>> address 192.168.56.15 >>>>>> netmask 255.255.255.0 >>>>>> network 192.168.56.0 >>>>>> broadcast 192.168.56.255 >>>>>> dns_nameserver 192.168.56.15 >>>>>> bridge_ports eth0 >>>>>> >>>>>> auto xenbr1 >>>>>> iface xenbr1 inet dhcp >>>>>> bridge_ports eth1 >>>>>> dns_nameserver 8.8.8.8 8.8.4.4 >>>>>> post-up route add default gw 10.0.3.2 >>>>>> >>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig >>>>>> eth0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:777 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:188 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:109977 (109.9 KB) TX bytes:11900 (11.9 KB) >>>>>> >>>>>> eth1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:478719 (478.7 KB) TX bytes:2542459 (2.5 MB) >>>>>> >>>>>> lo Link encap:Local Loopback >>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>> inet6 addr: ::1/128 Scope:Host >>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>> RX packets:360285 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:0 >>>>>> RX bytes:169128181 (169.1 MB) TX bytes:169128181 (169.1 MB) >>>>>> >>>>>> vif1.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:6 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:152 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:292 (292.0 B) TX bytes:9252 (9.2 KB) >>>>>> >>>>>> vif1.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:566 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:44227 (44.2 KB) TX bytes:173995 (173.9 KB) >>>>>> >>>>>> vif1.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:838 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:84 (84.0 B) TX bytes:111361 (111.3 KB) >>>>>> >>>>>> vif4.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:64 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:197 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:10276 (10.2 KB) TX bytes:18453 (18.4 KB) >>>>>> >>>>>> vif4.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:2051 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:233914 (233.9 KB) TX bytes:364243 (364.2 KB) >>>>>> >>>>>> vif4.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:582 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:84 (84.0 B) TX bytes:74700 (74.7 KB) >>>>>> >>>>>> vif4.3 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>> RX packets:0 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:585 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:32 >>>>>> RX bytes:0 (0.0 B) TX bytes:74826 (74.8 KB) >>>>>> >>>>>> xapi0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>> inet addr:169.254.0.1 Bcast:169.254.255.255 Mask:255.255.0.0 >>>>>> inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:568 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:0 >>>>>> RX bytes:76284 (76.2 KB) TX bytes:109085 (109.0 KB) >>>>>> >>>>>> xenbr0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>> inet addr:192.168.56.15 Bcast:192.168.56.255 >>>> Mask:255.255.255.0 >>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:4162 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:0 >>>>>> RX bytes:469199 (469.1 KB) TX bytes:485688 (485.6 KB) >>>>>> >>>>>> xenbr1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>> inet addr:10.0.3.15 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:0 >>>>>> RX bytes:404327 (404.3 KB) TX bytes:2501443 (2.5 MB) >>>>>> >>>>>> These physical NICs on the host translate to the following Xen PIFs: >>>>>> >>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list >>>>>> uuid ( RO) : 207413c9-5058-7a40-6c96-2dab21057f30 >>>>>> device ( RO): eth1 >>>>>> currently-attached ( RO): true >>>>>> VLAN ( RO): -1 >>>>>> network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72 >>>>>> >>>>>> >>>>>> uuid ( RO) : c0274787-e768-506f-3191-f0ac17b0c72b >>>>>> device ( RO): eth0 >>>>>> currently-attached ( RO): true >>>>>> VLAN ( RO): -1 >>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>> >>>>>> The following is the ifconfig from the SSVM: >>>>>> >>>>>> root@s-5-TEST:~# ifconfig >>>>>> eth0 Link encap:Ethernet HWaddr 0e:00:a9:fe:03:8b >>>>>> inet addr:169.254.3.139 Bcast:169.254.255.255 >>>> Mask:255.255.0.0 >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:235 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:92 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:21966 (21.4 KiB) TX bytes:16404 (16.0 KiB) >>>>>> Interrupt:8 >>>>>> >>>>>> eth1 Link encap:Ethernet HWaddr 06:bc:62:00:00:05 >>>>>> inet addr:192.168.56.104 Bcast:192.168.56.255 >>>>>> Mask:255.255.255.0 >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:2532 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:341242 (333.2 KiB) TX bytes:272183 (265.8 KiB) >>>>>> Interrupt:10 >>>>>> >>>>>> eth2 Link encap:Ethernet HWaddr 06:12:72:00:00:37 >>>>>> inet addr:10.0.3.204 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:600 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:3 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:68648 (67.0 KiB) TX bytes:126 (126.0 B) >>>>>> Interrupt:11 >>>>>> >>>>>> eth3 Link encap:Ethernet HWaddr 06:25:e2:00:00:15 >>>>>> inet addr:192.168.56.120 Bcast:192.168.56.255 >>>>>> Mask:255.255.255.0 >>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>> RX packets:603 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:1000 >>>>>> RX bytes:68732 (67.1 KiB) TX bytes:0 (0.0 B) >>>>>> Interrupt:12 >>>>>> >>>>>> lo Link encap:Local Loopback >>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>> RX packets:61 errors:0 dropped:0 overruns:0 frame:0 >>>>>> TX packets:61 errors:0 dropped:0 overruns:0 carrier:0 >>>>>> collisions:0 txqueuelen:0 >>>>>> RX bytes:5300 (5.1 KiB) TX bytes:5300 (5.1 KiB) >>>>>> >>>>>> Finally, the following are the vif params for the eth2 device on the >>>> SSVM >>>>>> depicting its connection to eth0 instead of eth1: >>>>>> >>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243 >>>>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210 >>>>>> uuid ( RO) : >> be44bb30-5700-b461-760e-10fe93079210 >>>>>> vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>> vm-name-label ( RO): s-5-TEST >>>>>> allowed-operations (SRO): attach; unplug_force; unplug >>>>>> current-operations (SRO): >>>>>> device ( RO): 2 >>>>>> MAC ( RO): 06:12:72:00:00:37 >>>>>> MAC-autogenerated ( RO): false >>>>>> MTU ( RO): 1500 >>>>>> currently-attached ( RO): true >>>>>> qos_algorithm_type ( RW): ratelimit >>>>>> qos_algorithm_params (MRW): kbps: 25600 >>>>>> qos_supported_algorithms (SRO): >>>>>> other-config (MRW): nicira-iface-id: >>>>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id: >>>>>> 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>> network-name-label ( RO): Pool-wide network associated with >>>> eth0 >>>>>> io_read_kbs ( RO): 0.007 >>>>>> io_write_kbs ( RO): 0.000 >>>>>> >>>>>> How do I configure CloudStack such that the guest network NIC on the >> VM >>>>>> will be connected to correct physical NIC? >>>>>> >>>>>> Thanks for your help, >>>>>> -John >>>>>> >>>>>> >>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590 >>>>>> >>>>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com> >>>> wrote: >>>>>> >>>>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the >> use >>>> of >>>>>>> management network to route to 10.0.3.2 for DNS. that's where the >> route >>>>>> is >>>>>>> coming from. you will want to use something on your management net >> for >>>>>>> internal DNS, or something other than that router. >>>>>>> >>>>>>> >>>>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com> >>>>>> wrote: >>>>>>> >>>>>>>> Anthony, >>>>>>>> >>>>>>>> I apologize for forgetting to response to the part of your answer >> the >>>>>>>> first part of the question. I had set the management.network.cidr >> and >>>>>> host >>>>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively. >>>>>> Please >>>>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my >>>> original >>>>>>>> email for the actual setting, as well as, the network configurations >>>>>> used >>>>>>>> when this problem occurs. >>>>>>>> >>>>>>>> Thanks, >>>>>>>> -John >>>>>>>> >>>>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com> >> wrote: >>>>>>>> >>>>>>>>> Hi join, >>>>>>>>> >>>>>>>>> Try following, >>>>>>>>> >>>>>>>>> Set global configuration management.network.cidr to your management >>>>>>>> server CIDR, if this configuration is not available in UI, you can >>>>>> change >>>>>>>> it in DB directly. >>>>>>>>> >>>>>>>>> Restart management, >>>>>>>>> Stop/Start SSVM and CPVM. >>>>>>>>> >>>>>>>>> >>>>>>>>> And could you post "cat /proc/cmdline" in SSVM? >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Anthony >>>>>>>>> >>>>>>>>>> -----Original Message----- >>>>>>>>>> From: John Burwell [mailto:jburw...@basho.com] >>>>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM >>>>>>>>>> To: cloudstack-dev@incubator.apache.org >>>>>>>>>> Subject: Re: SSVM Network Configuration Issue >>>>>>>>>> >>>>>>>>>> All, >>>>>>>>>> >>>>>>>>>> I was wondering if anyone else is experiencing this problem when >>>> using >>>>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT >>>>>>>>>> adapter. One aspect of this issue that seems interesting is that >>>>>>>>>> following route table from the SSVM: >>>>>>>>>> >>>>>>>>>> root@s-5-TEST:~# route >>>>>>>>>> Kernel IP routing table >>>>>>>>>> Destination Gateway Genmask Flags Metric Ref >>>>>> Use >>>>>>>>>> Iface >>>>>>>>>> 10.0.3.2 192.168.56.1 255.255.255.255 UGH 0 0 >>>>>> 0 >>>>>>>>>> eth1 >>>>>>>>>> 10.0.3.0 * 255.255.255.0 U 0 0 >>>>>> 0 >>>>>>>>>> eth2 >>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>> 0 >>>>>>>>>> eth1 >>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>> 0 >>>>>>>>>> eth3 >>>>>>>>>> link-local * 255.255.0.0 U 0 0 >>>>>> 0 >>>>>>>>>> eth0 >>>>>>>>>> default 10.0.3.2 0.0.0.0 UG 0 0 >>>>>> 0 >>>>>>>>>> eth2 >>>>>>>>>> >>>>>>>>>> In particular, the gateways for the management and guest networks >> do >>>>>>>>>> not match to the configuration provided to the management server >>>> (i.e. >>>>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and >>>> 192.168.56.1 >>>>>> is >>>>>>>>>> the gateway for the 192.168.56.0/24 network). With this >>>>>> configuration, >>>>>>>>>> the SSVM has a socket connection to the management server, but is >> in >>>>>>>>>> alert state. Finally, when I remove the host-only NIC and use >> only >>>> a >>>>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to >>>>>>>>>> believe that the segregated network configuration is at the root >> of >>>>>> the >>>>>>>>>> problem. >>>>>>>>>> >>>>>>>>>> Until I can get the networking on the SSVM configured, I am unable >>>> to >>>>>>>>>> complete the testing of the S3-backed Secondary Storage >> enhancement. >>>>>>>>>> >>>>>>>>>> Thank you for your help, >>>>>>>>>> -John >>>>>>>>>> >>>>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com> >>>> wrote: >>>>>>>>>> >>>>>>>>>>> All, >>>>>>>>>>> >>>>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox >>>>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1. I have configured the >>>>>> base >>>>>>>>>> management server VM (zone1) to serve as both the zone1, as well >> as, >>>>>>>>>> the management server (running MySql) with eth0 as a host-only >>>> adapter >>>>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see >> the >>>>>>>>>> attached zone1-interfaces file for the exact network configuration >>>> on >>>>>>>>>> the VM). The management and guest networks are configured as >>>> follows: >>>>>>>>>>> >>>>>>>>>>> Zone 1 >>>>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?) >>>>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>> Zone 2 >>>>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?) >>>>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>> >>>>>>>>>>> The management server deploys and starts without error. I then >>>>>>>>>> populate the configuration it using the attached Marvin >>>> configuration >>>>>>>>>> file (zone1.devcloud.cfg) and restart the management server in >> order >>>>>> to >>>>>>>>>> allow the global configuration option changes to take effect. >>>>>>>>>> Following the restart, the CPVM and SSVM start without error. >>>>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable >>>> to >>>>>>>>>> connect outbound through the guest network (very important for my >>>>>> tests >>>>>>>>>> because I am testing S3-backed secondary storage). >>>>>>>>>>> >>>>>>>>>>> From the diagnostic checks I have performed on the management >>>> server >>>>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting >>>>>> back >>>>>>>>>> to the management server. I have attached a set of diagnostic >>>>>>>>>> information from the management server >>>> (mgmtsvr-zone1-diagnostics.log) >>>>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the >>>> results >>>>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other >>>>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the >>>>>> SSVM). >>>>>>>>>> Finally, I have attached the vmops log from the management server >>>>>>>>>> (vmops-zone1.log). >>>>>>>>>>> >>>>>>>>>>> What changes need to be made to management server configuration >> in >>>>>>>>>> order to start up an SSVM that can communicate with the secondary >>>>>>>>>> storage NFS volumes, management server, and connect to hosts on >> the >>>>>>>>>> Internet? >>>>>>>>>>> >>>>>>>>>>> Thanks for your help, >>>>>>>>>>> -John >>>>>>>>>>> >>>>>>>>>>> <ssvm-zone1-diagnostics.log> >>>>>>>>>>> <vmops-zone1.tar.gz> >>>>>>>>>>> <mgmtsvr-zone1-diagnostics.log> >>>>>>>>>>> <zone1-interfaces> >>>>>>>>>>> <zone1.devcloud.cfg> >> >>