Rohit, As I stated below, I know which the VIF->PIF->device Xen mapping needs to be used for each network. My question is how do I configure CloudStack with that information.
Thanks for your help, -John On Dec 15, 2012, at 2:51 PM, Rohit Yadav <rohit.ya...@citrix.com> wrote: > About xen bridging, these can help; > > brctl show all (bridge vif mappings) > ip show addr xenbr0 (bridge specific info) > brctl showmacs br0 (bridge mac mappings) > > Wiki: > http://wiki.xen.org/wiki/Xen_FAQ_Networking > http://wiki.xen.org/wiki/XenNetworking > > Regards. > > ________________________________________ > From: John Burwell [jburw...@basho.com] > Sent: Saturday, December 15, 2012 8:35 PM > To: cloudstack-dev@incubator.apache.org > Subject: Re: SSVM Network Configuration Issue > > Marcus, > > That's what I thought. The Xen physical bridge names are xenbr0 (to > eth0) and xenbr1 (to eth1). Using basic network configuration, I set > the Xen network traffic labels for each to the appropriate bridge > device name. I receive errors regarding invalid network device when > it attempts to create a VM. Does anyone else know how determine the > mapping of physical devices to CloudStack Xen network traffic labels? > > Thanks, > -John > > On Dec 15, 2012, at 1:20 AM, Marcus Sorensen <shadow...@gmail.com> wrote: > >> Vlans in advanced/KVM should only be required for the guest networks. If I >> create a bridge on physical eth0, name it 'br0', and a bridge on physical >> eth1, naming it 'br1', and then set my management network to label 'br0', >> my public network to 'br1', and my guest network to 'br1', it should use >> the bridges you asked for when connecting the system VMs for the specified >> traffic. I'd just leave the 'vlan' blank when specifying public and >> pod(management) IPs. In this scenario, the only place you need to enter >> vlans is on the guest, and it should create new tagged interfaces/bridges >> on eth1(per your label of br1) as new guest networks are brought online. >> This is how my dev VMs are usually set up. >> >> >> On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote: >> >>> Marcus, >>> >>> My question, more specifically, is are VLANs required to implement traffic >>> labels? Also, can traffic labels be configured in Basic networking mode or >>> do I need to switch my configuration to Advanced? >>> >>> I am not disagreeing on the how DNS servers should be associated with >>> interfaces nor do I think a network operator should be required to make any >>> upstream router configuration changes. I am simply saying that CloudStack >>> should not make assumptions about the gateways that have been specified. >>> The behavior I experienced of CloudStack attempting to >>> "correct" my configuration by injecting another route fails the rule of >>> least surprise and is based on incomplete knowledge. In my opinion, >>> CloudStack (or any system of its ilk) should faithfully (or slavishly) >>> realize the routes on the system VM as specified. If the configuration is >>> incorrect, networking will fail in an expected manner, and the operator can >>> adjust their environment as necessary. Otherwise, there is an upstream >>> router configuration to which CloudStack has no visibility, but with which >>> it is completely compatible. Essentially, I am asking CloudStack to do >>> less, assume I know what I am doing, and break in a manner consistent with >>> other network applications. >>> >>> Thanks, >>> -John >>> >>> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote: >>> >>>> Traffic labels essentially tell the system which physical network to use. >>>> So if you've allocated a vlan for a specific traffic type, it will first >>>> look at the tag associated with that traffic type, figure out which >>>> physical interface goes with that, and then create a tagged interface and >>>> bridge also on that physical. >>>> >>>> I guess we'll just have to disagree, I think the current behavior makes >>>> total sense. To me, internal DNS should always use the management >>>> interface, since it's internally facing. There's no sane way to do that >>>> other than a static route on the system vm (it seems you're suggesting >>> that >>>> the network operator force something like this on the upstream router, >>>> which seems really strange to require everyone to create static routes on >>>> their public network to force specific IPs back into their internal >>>> networks, so correct me if I have the wrong impression). Cloudstack is >>>> doing exactly what you tell it to. You told it that 10.0.3.2 should be >>>> accessible via your internal network by setting it as your internal DNS. >>>> The fact that a broken config doesn't work isn't CloudStack's fault. >>>> >>>> Note that internal DNS is just the default for the ssvm, public DNS is >>>> still offered as a backup, so had you not said that 10.0.3.2 was >>> available >>>> on your internal network (perhaps offering a dummy internal DNS >>>> address or 192.68.56.1), >>>> lookups would fall back to public and everything would work as expected >>> as >>>> well. >>>> >>>> There is also a global config called 'use.external.dns', but in setting >>>> this, restarting the management server, recreating system VMs, I don't >>> see >>>> a noticeable difference on any of this, so perhaps that would solve your >>>> issue as well but it's either broken or doesn't do what I thought it >>> would. >>>> >>>> >>>> On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote: >>>> >>>>> Marcus, >>>>> >>>>> Are traffic labels independent of VLANs? I ask because my current XCP >>>>> network configuration is bridged, and I am not using Open vSwitch. >>>>> >>>>> I disagree on the routing issue. CloudStack should do what's told >>> because >>>>> it does not have insight into or control of the configuration of the >>> routes >>>>> in the layers beneath it. If CloudStack simply did as it was told, it >>>>> would fail as expected in a typical networking environment while >>> preserving >>>>> the flexibility of configuration expected by a network engineer. >>>>> >>>>> Thanks, >>>>> -John >>>>> >>>>> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com> >>> wrote: >>>>> >>>>>> I can't really tell you for xen, although it might be similar to KVM. >>>>>> During setup I would set a traffic label matching the name of the >>> bridge, >>>>>> for example if my public interface were eth0 and the bridge I had set >>> up >>>>>> was br0, I'd go to the zone network settings, find public traffic, and >>>>> set >>>>>> a label on it of "br0". Maybe someone more familiar with the xen setup >>>>> can >>>>>> help. >>>>>> >>>>>> On the DNS, it makes sense from the perspective that the ssvm has >>> access >>>>> to >>>>>> your internal networks, thus it uses your internal DNS. Its default >>>>> gateway >>>>>> is public. So if I have a DNS server on an internal network at >>>>>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this >>> route >>>>>> has to be set in order for the DNS server to be reachable. You would >>>>> under >>>>>> normal circumstances not want to use a DNS server on public net as your >>>>>> internal DNS setting anyway, although I agree that the route insertion >>>>>> should have a bit more sanity checking and not set a static route to >>> your >>>>>> default gateway. >>>>>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote: >>>>>> >>>>>>> Marcus, >>>>>>> >>>>>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS >>>>> for >>>>>>> the management network to use it, and the route table in the SSVM is >>> now >>>>>>> correct. However, this behavior does not seem correct. At a minimum, >>>>> it >>>>>>> violates the rule of least surprise. CloudStack shouldn't be adding >>>>>>> gateways that are not configured. Therefore, I have entered a >>>>> defect[1] to >>>>>>> remove the behavior. >>>>>>> >>>>>>> With the route table fixed, I am now experiencing a new problem. The >>>>>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the >>>>> internal >>>>>>> NIC (192.168.56.0/24) on the host. The host-only network >>>>> (192.168.56.15) >>>>>>> is configured on xenbr0 and the NAT network is configured on xenbr1. >>>>> As a >>>>>>> reference, the following is the contents of the >>> /etc/network/interfaces >>>>>>> file and ifconfig from devcloud host: >>>>>>> >>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat >>>>>>> /etc/network/interfaces >>>>>>> # The loopback network interface >>>>>>> auto lo >>>>>>> iface lo inet loopback >>>>>>> >>>>>>> auto eth0 >>>>>>> iface eth0 inet manual >>>>>>> >>>>>>> allow-hotplug eth1 >>>>>>> iface eth1 inet manual >>>>>>> >>>>>>> # The primary network interface >>>>>>> auto xenbr0 >>>>>>> iface xenbr0 inet static >>>>>>> address 192.168.56.15 >>>>>>> netmask 255.255.255.0 >>>>>>> network 192.168.56.0 >>>>>>> broadcast 192.168.56.255 >>>>>>> dns_nameserver 192.168.56.15 >>>>>>> bridge_ports eth0 >>>>>>> >>>>>>> auto xenbr1 >>>>>>> iface xenbr1 inet dhcp >>>>>>> bridge_ports eth1 >>>>>>> dns_nameserver 8.8.8.8 8.8.4.4 >>>>>>> post-up route add default gw 10.0.3.2 >>>>>>> >>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig >>>>>>> eth0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:777 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:188 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:109977 (109.9 KB) TX bytes:11900 (11.9 KB) >>>>>>> >>>>>>> eth1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:478719 (478.7 KB) TX bytes:2542459 (2.5 MB) >>>>>>> >>>>>>> lo Link encap:Local Loopback >>>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>>> inet6 addr: ::1/128 Scope:Host >>>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>>> RX packets:360285 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:0 >>>>>>> RX bytes:169128181 (169.1 MB) TX bytes:169128181 (169.1 MB) >>>>>>> >>>>>>> vif1.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:6 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:152 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:292 (292.0 B) TX bytes:9252 (9.2 KB) >>>>>>> >>>>>>> vif1.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:566 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:44227 (44.2 KB) TX bytes:173995 (173.9 KB) >>>>>>> >>>>>>> vif1.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:838 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:84 (84.0 B) TX bytes:111361 (111.3 KB) >>>>>>> >>>>>>> vif4.0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:64 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:197 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:10276 (10.2 KB) TX bytes:18453 (18.4 KB) >>>>>>> >>>>>>> vif4.1 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:2051 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:233914 (233.9 KB) TX bytes:364243 (364.2 KB) >>>>>>> >>>>>>> vif4.2 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:3 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:582 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:84 (84.0 B) TX bytes:74700 (74.7 KB) >>>>>>> >>>>>>> vif4.3 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link >>>>>>> UP BROADCAST RUNNING NOARP PROMISC MTU:1500 Metric:1 >>>>>>> RX packets:0 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:585 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:32 >>>>>>> RX bytes:0 (0.0 B) TX bytes:74826 (74.8 KB) >>>>>>> >>>>>>> xapi0 Link encap:Ethernet HWaddr fe:ff:ff:ff:ff:ff >>>>>>> inet addr:169.254.0.1 Bcast:169.254.255.255 Mask:255.255.0.0 >>>>>>> inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:568 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:0 >>>>>>> RX bytes:76284 (76.2 KB) TX bytes:109085 (109.0 KB) >>>>>>> >>>>>>> xenbr0 Link encap:Ethernet HWaddr 08:00:27:7e:74:9c >>>>>>> inet addr:192.168.56.15 Bcast:192.168.56.255 >>>>> Mask:255.255.255.0 >>>>>>> inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:4162 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:0 >>>>>>> RX bytes:469199 (469.1 KB) TX bytes:485688 (485.6 KB) >>>>>>> >>>>>>> xenbr1 Link encap:Ethernet HWaddr 08:00:27:df:00:00 >>>>>>> inet addr:10.0.3.15 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>>> inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:4129 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:0 >>>>>>> RX bytes:404327 (404.3 KB) TX bytes:2501443 (2.5 MB) >>>>>>> >>>>>>> These physical NICs on the host translate to the following Xen PIFs: >>>>>>> >>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list >>>>>>> uuid ( RO) : 207413c9-5058-7a40-6c96-2dab21057f30 >>>>>>> device ( RO): eth1 >>>>>>> currently-attached ( RO): true >>>>>>> VLAN ( RO): -1 >>>>>>> network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72 >>>>>>> >>>>>>> >>>>>>> uuid ( RO) : c0274787-e768-506f-3191-f0ac17b0c72b >>>>>>> device ( RO): eth0 >>>>>>> currently-attached ( RO): true >>>>>>> VLAN ( RO): -1 >>>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>>> >>>>>>> The following is the ifconfig from the SSVM: >>>>>>> >>>>>>> root@s-5-TEST:~# ifconfig >>>>>>> eth0 Link encap:Ethernet HWaddr 0e:00:a9:fe:03:8b >>>>>>> inet addr:169.254.3.139 Bcast:169.254.255.255 >>>>> Mask:255.255.0.0 >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:235 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:92 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:21966 (21.4 KiB) TX bytes:16404 (16.0 KiB) >>>>>>> Interrupt:8 >>>>>>> >>>>>>> eth1 Link encap:Ethernet HWaddr 06:bc:62:00:00:05 >>>>>>> inet addr:192.168.56.104 Bcast:192.168.56.255 >>>>>>> Mask:255.255.255.0 >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:2532 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:341242 (333.2 KiB) TX bytes:272183 (265.8 KiB) >>>>>>> Interrupt:10 >>>>>>> >>>>>>> eth2 Link encap:Ethernet HWaddr 06:12:72:00:00:37 >>>>>>> inet addr:10.0.3.204 Bcast:10.0.3.255 Mask:255.255.255.0 >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:600 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:3 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:68648 (67.0 KiB) TX bytes:126 (126.0 B) >>>>>>> Interrupt:11 >>>>>>> >>>>>>> eth3 Link encap:Ethernet HWaddr 06:25:e2:00:00:15 >>>>>>> inet addr:192.168.56.120 Bcast:192.168.56.255 >>>>>>> Mask:255.255.255.0 >>>>>>> UP BROADCAST RUNNING MULTICAST MTU:1500 Metric:1 >>>>>>> RX packets:603 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:0 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:1000 >>>>>>> RX bytes:68732 (67.1 KiB) TX bytes:0 (0.0 B) >>>>>>> Interrupt:12 >>>>>>> >>>>>>> lo Link encap:Local Loopback >>>>>>> inet addr:127.0.0.1 Mask:255.0.0.0 >>>>>>> UP LOOPBACK RUNNING MTU:16436 Metric:1 >>>>>>> RX packets:61 errors:0 dropped:0 overruns:0 frame:0 >>>>>>> TX packets:61 errors:0 dropped:0 overruns:0 carrier:0 >>>>>>> collisions:0 txqueuelen:0 >>>>>>> RX bytes:5300 (5.1 KiB) TX bytes:5300 (5.1 KiB) >>>>>>> >>>>>>> Finally, the following are the vif params for the eth2 device on the >>>>> SSVM >>>>>>> depicting its connection to eth0 instead of eth1: >>>>>>> >>>>>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243 >>>>>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210 >>>>>>> uuid ( RO) : >>> be44bb30-5700-b461-760e-10fe93079210 >>>>>>> vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>>> vm-name-label ( RO): s-5-TEST >>>>>>> allowed-operations (SRO): attach; unplug_force; unplug >>>>>>> current-operations (SRO): >>>>>>> device ( RO): 2 >>>>>>> MAC ( RO): 06:12:72:00:00:37 >>>>>>> MAC-autogenerated ( RO): false >>>>>>> MTU ( RO): 1500 >>>>>>> currently-attached ( RO): true >>>>>>> qos_algorithm_type ( RW): ratelimit >>>>>>> qos_algorithm_params (MRW): kbps: 25600 >>>>>>> qos_supported_algorithms (SRO): >>>>>>> other-config (MRW): nicira-iface-id: >>>>>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id: >>>>>>> 7958d91f-e52d-a25d-718c-7f831ae701d7 >>>>>>> network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a >>>>>>> network-name-label ( RO): Pool-wide network associated with >>>>> eth0 >>>>>>> io_read_kbs ( RO): 0.007 >>>>>>> io_write_kbs ( RO): 0.000 >>>>>>> >>>>>>> How do I configure CloudStack such that the guest network NIC on the >>> VM >>>>>>> will be connected to correct physical NIC? >>>>>>> >>>>>>> Thanks for your help, >>>>>>> -John >>>>>>> >>>>>>> >>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590 >>>>>>> >>>>>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com> >>>>> wrote: >>>>>>> >>>>>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the >>> use >>>>> of >>>>>>>> management network to route to 10.0.3.2 for DNS. that's where the >>> route >>>>>>> is >>>>>>>> coming from. you will want to use something on your management net >>> for >>>>>>>> internal DNS, or something other than that router. >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com> >>>>>>> wrote: >>>>>>>> >>>>>>>>> Anthony, >>>>>>>>> >>>>>>>>> I apologize for forgetting to response to the part of your answer >>> the >>>>>>>>> first part of the question. I had set the management.network.cidr >>> and >>>>>>> host >>>>>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively. >>>>>>> Please >>>>>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my >>>>> original >>>>>>>>> email for the actual setting, as well as, the network configurations >>>>>>> used >>>>>>>>> when this problem occurs. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> -John >>>>>>>>> >>>>>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com> >>> wrote: >>>>>>>>> >>>>>>>>>> Hi join, >>>>>>>>>> >>>>>>>>>> Try following, >>>>>>>>>> >>>>>>>>>> Set global configuration management.network.cidr to your management >>>>>>>>> server CIDR, if this configuration is not available in UI, you can >>>>>>> change >>>>>>>>> it in DB directly. >>>>>>>>>> >>>>>>>>>> Restart management, >>>>>>>>>> Stop/Start SSVM and CPVM. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> And could you post "cat /proc/cmdline" in SSVM? >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> Anthony >>>>>>>>>> >>>>>>>>>>> -----Original Message----- >>>>>>>>>>> From: John Burwell [mailto:jburw...@basho.com] >>>>>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM >>>>>>>>>>> To: cloudstack-dev@incubator.apache.org >>>>>>>>>>> Subject: Re: SSVM Network Configuration Issue >>>>>>>>>>> >>>>>>>>>>> All, >>>>>>>>>>> >>>>>>>>>>> I was wondering if anyone else is experiencing this problem when >>>>> using >>>>>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT >>>>>>>>>>> adapter. One aspect of this issue that seems interesting is that >>>>>>>>>>> following route table from the SSVM: >>>>>>>>>>> >>>>>>>>>>> root@s-5-TEST:~# route >>>>>>>>>>> Kernel IP routing table >>>>>>>>>>> Destination Gateway Genmask Flags Metric Ref >>>>>>> Use >>>>>>>>>>> Iface >>>>>>>>>>> 10.0.3.2 192.168.56.1 255.255.255.255 UGH 0 0 >>>>>>> 0 >>>>>>>>>>> eth1 >>>>>>>>>>> 10.0.3.0 * 255.255.255.0 U 0 0 >>>>>>> 0 >>>>>>>>>>> eth2 >>>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>>> 0 >>>>>>>>>>> eth1 >>>>>>>>>>> 192.168.56.0 * 255.255.255.0 U 0 0 >>>>>>> 0 >>>>>>>>>>> eth3 >>>>>>>>>>> link-local * 255.255.0.0 U 0 0 >>>>>>> 0 >>>>>>>>>>> eth0 >>>>>>>>>>> default 10.0.3.2 0.0.0.0 UG 0 0 >>>>>>> 0 >>>>>>>>>>> eth2 >>>>>>>>>>> >>>>>>>>>>> In particular, the gateways for the management and guest networks >>> do >>>>>>>>>>> not match to the configuration provided to the management server >>>>> (i.e. >>>>>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and >>>>> 192.168.56.1 >>>>>>> is >>>>>>>>>>> the gateway for the 192.168.56.0/24 network). With this >>>>>>> configuration, >>>>>>>>>>> the SSVM has a socket connection to the management server, but is >>> in >>>>>>>>>>> alert state. Finally, when I remove the host-only NIC and use >>> only >>>>> a >>>>>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to >>>>>>>>>>> believe that the segregated network configuration is at the root >>> of >>>>>>> the >>>>>>>>>>> problem. >>>>>>>>>>> >>>>>>>>>>> Until I can get the networking on the SSVM configured, I am unable >>>>> to >>>>>>>>>>> complete the testing of the S3-backed Secondary Storage >>> enhancement. >>>>>>>>>>> >>>>>>>>>>> Thank you for your help, >>>>>>>>>>> -John >>>>>>>>>>> >>>>>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com> >>>>> wrote: >>>>>>>>>>> >>>>>>>>>>>> All, >>>>>>>>>>>> >>>>>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox >>>>>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1. I have configured the >>>>>>> base >>>>>>>>>>> management server VM (zone1) to serve as both the zone1, as well >>> as, >>>>>>>>>>> the management server (running MySql) with eth0 as a host-only >>>>> adapter >>>>>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see >>> the >>>>>>>>>>> attached zone1-interfaces file for the exact network configuration >>>>> on >>>>>>>>>>> the VM). The management and guest networks are configured as >>>>> follows: >>>>>>>>>>>> >>>>>>>>>>>> Zone 1 >>>>>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?) >>>>>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>>> Zone 2 >>>>>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?) >>>>>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8 >>>>>>>>>>>> >>>>>>>>>>>> The management server deploys and starts without error. I then >>>>>>>>>>> populate the configuration it using the attached Marvin >>>>> configuration >>>>>>>>>>> file (zone1.devcloud.cfg) and restart the management server in >>> order >>>>>>> to >>>>>>>>>>> allow the global configuration option changes to take effect. >>>>>>>>>>> Following the restart, the CPVM and SSVM start without error. >>>>>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable >>>>> to >>>>>>>>>>> connect outbound through the guest network (very important for my >>>>>>> tests >>>>>>>>>>> because I am testing S3-backed secondary storage). >>>>>>>>>>>> >>>>>>>>>>>> From the diagnostic checks I have performed on the management >>>>> server >>>>>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting >>>>>>> back >>>>>>>>>>> to the management server. I have attached a set of diagnostic >>>>>>>>>>> information from the management server >>>>> (mgmtsvr-zone1-diagnostics.log) >>>>>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the >>>>> results >>>>>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other >>>>>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the >>>>>>> SSVM). >>>>>>>>>>> Finally, I have attached the vmops log from the management server >>>>>>>>>>> (vmops-zone1.log). >>>>>>>>>>>> >>>>>>>>>>>> What changes need to be made to management server configuration >>> in >>>>>>>>>>> order to start up an SSVM that can communicate with the secondary >>>>>>>>>>> storage NFS volumes, management server, and connect to hosts on >>> the >>>>>>>>>>> Internet? >>>>>>>>>>>> >>>>>>>>>>>> Thanks for your help, >>>>>>>>>>>> -John >>>>>>>>>>>> >>>>>>>>>>>> <ssvm-zone1-diagnostics.log> >>>>>>>>>>>> <vmops-zone1.tar.gz> >>>>>>>>>>>> <mgmtsvr-zone1-diagnostics.log> >>>>>>>>>>>> <zone1-interfaces> >>>>>>>>>>>> <zone1.devcloud.cfg> >>> >>>