Vlans in advanced/KVM should only be required for the guest networks. If I
create a bridge on physical eth0, name it 'br0', and a bridge on physical
eth1, naming it 'br1', and then set my management network to label 'br0',
my public network to 'br1', and my guest network to 'br1', it should use
the bridges you asked for when connecting the system VMs for the specified
traffic. I'd just leave the 'vlan' blank when specifying public and
pod(management) IPs. In this scenario, the only place you need to enter
vlans is on the guest, and it should create new tagged interfaces/bridges
on eth1(per your label of br1) as new guest networks are brought online.
This is how my dev VMs are usually set up.


On Thu, Dec 6, 2012 at 11:03 AM, John Burwell <jburw...@basho.com> wrote:

> Marcus,
>
> My question, more specifically, is are VLANs required to implement traffic
> labels?  Also, can traffic labels be configured in Basic networking mode or
> do I need to switch my configuration to Advanced?
>
> I am not disagreeing on the how DNS servers should be associated with
> interfaces nor do I think a network operator should be required to make any
> upstream router configuration changes.  I am simply saying that CloudStack
> should not make assumptions about the gateways that have been specified.
>  The behavior I experienced of CloudStack attempting to
> "correct" my configuration by injecting another route fails the rule of
> least surprise and is based on incomplete knowledge.  In my opinion,
> CloudStack (or any system of its ilk) should faithfully (or slavishly)
> realize the routes on the system VM as specified.  If the configuration is
> incorrect, networking will fail in an expected manner, and the operator can
> adjust their environment as necessary.   Otherwise, there is an upstream
> router configuration to which CloudStack has no visibility, but with which
> it is completely compatible.  Essentially, I am asking CloudStack to do
> less, assume I know what I am doing, and break in a manner consistent with
> other network applications.
>
> Thanks,
> -John
>
> On Dec 6, 2012, at 12:30 PM, Marcus Sorensen <shadow...@gmail.com> wrote:
>
> > Traffic labels essentially tell the system which physical network to use.
> > So if you've allocated a vlan for a specific traffic type, it will first
> > look at the tag associated with that traffic type, figure out which
> > physical interface goes with that, and then create a tagged interface and
> > bridge also on that physical.
> >
> > I guess we'll just have to disagree, I think the current behavior makes
> > total sense.  To me, internal DNS should always use the management
> > interface, since it's internally facing. There's no sane way to do that
> > other than a static route on the system vm (it seems you're suggesting
> that
> > the network operator force something like this on the upstream router,
> > which seems really strange to require everyone to create static routes on
> > their public network to force specific IPs back into their internal
> > networks, so correct me if I have the wrong impression).  Cloudstack is
> > doing exactly what you tell it to. You told it that 10.0.3.2 should be
> > accessible via your internal network by setting it as your internal DNS.
> > The fact that a broken config doesn't work isn't CloudStack's fault.
> >
> > Note that internal DNS is just the default for the ssvm, public DNS is
> > still offered as a backup, so had you not said that 10.0.3.2 was
> available
> > on your internal network (perhaps offering a dummy internal DNS
> > address or 192.68.56.1),
> > lookups would fall back to public and everything would work as expected
> as
> > well.
> >
> > There is also a global config called 'use.external.dns', but in setting
> > this, restarting the management server, recreating system VMs, I don't
> see
> > a noticeable difference on any of this, so perhaps that would solve your
> > issue as well but it's either broken or doesn't do what I thought it
> would.
> >
> >
> > On Thu, Dec 6, 2012 at 8:39 AM, John Burwell <jburw...@basho.com> wrote:
> >
> >> Marcus,
> >>
> >> Are traffic labels independent of VLANs?  I ask because my current XCP
> >> network configuration is bridged, and I am not using Open vSwitch.
> >>
> >> I disagree on the routing issue.  CloudStack should do what's told
> because
> >> it does not have insight into or control of the configuration of the
> routes
> >> in the layers beneath it.  If CloudStack simply did as it was told, it
> >> would fail as expected in a typical networking environment while
> preserving
> >> the flexibility of configuration expected by a network engineer.
> >>
> >> Thanks,
> >> -John
> >>
> >> On Dec 6, 2012, at 10:35 AM, Marcus Sorensen <shadow...@gmail.com>
> wrote:
> >>
> >>> I can't really tell you for xen, although it might be similar to KVM.
> >>> During setup I would set a traffic label matching the name of the
> bridge,
> >>> for example if my public interface were eth0 and the bridge I had set
> up
> >>> was br0, I'd go to the zone network settings, find public traffic, and
> >> set
> >>> a label on it of "br0". Maybe someone more familiar with the xen setup
> >> can
> >>> help.
> >>>
> >>> On the DNS, it makes sense from the perspective that the ssvm has
> access
> >> to
> >>> your internal networks, thus it uses your internal DNS. Its default
> >> gateway
> >>> is public. So if I have a DNS server on an internal network at
> >>> 10.30.20.10/24, and my management network on 192.168.10.0/24, this
> route
> >>> has to be set in order for the DNS server to be reachable. You would
> >> under
> >>> normal circumstances not want to use a DNS server on public net as your
> >>> internal DNS setting anyway, although I agree that the route insertion
> >>> should have a bit more sanity checking and not set a static route to
> your
> >>> default gateway.
> >>> On Dec 6, 2012 6:31 AM, "John Burwell" <jburw...@basho.com> wrote:
> >>>
> >>>> Marcus,
> >>>>
> >>>> I setup a small PowerDNS recursor on 192.168.56.15, configured the DNS
> >> for
> >>>> the management network to use it, and the route table in the SSVM is
> now
> >>>> correct.  However, this behavior does not seem correct.  At a minimum,
> >> it
> >>>> violates the rule of least surprise.  CloudStack shouldn't be adding
> >>>> gateways that are not configured.  Therefore, I have entered a
> >> defect[1] to
> >>>> remove the behavior.
> >>>>
> >>>> With the route table fixed, I am now experiencing a new problem.  The
> >>>> external NIC (10.0.3.0/24) on the SSVM is being connected to the
> >> internal
> >>>> NIC (192.168.56.0/24) on the host.  The host-only network
> >> (192.168.56.15)
> >>>> is configured on xenbr0 and the NAT network is configured on xenbr1.
> >> As a
> >>>> reference, the following is the contents of the
> /etc/network/interfaces
> >>>> file and ifconfig from devcloud host:
> >>>>
> >>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# cat
> >>>> /etc/network/interfaces
> >>>> # The loopback network interface
> >>>> auto lo
> >>>> iface lo inet loopback
> >>>>
> >>>> auto eth0
> >>>> iface eth0 inet manual
> >>>>
> >>>> allow-hotplug eth1
> >>>> iface eth1 inet manual
> >>>>
> >>>> # The primary network interface
> >>>> auto xenbr0
> >>>> iface xenbr0 inet static
> >>>> address 192.168.56.15
> >>>> netmask 255.255.255.0
> >>>> network 192.168.56.0
> >>>> broadcast 192.168.56.255
> >>>> dns_nameserver 192.168.56.15
> >>>> bridge_ports eth0
> >>>>
> >>>> auto xenbr1
> >>>> iface xenbr1 inet dhcp
> >>>> bridge_ports eth1
> >>>> dns_nameserver 8.8.8.8 8.8.4.4
> >>>> post-up route add default gw 10.0.3.2
> >>>>
> >>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# ifconfig
> >>>> eth0      Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
> >>>>         inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:777 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:188 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:109977 (109.9 KB)  TX bytes:11900 (11.9 KB)
> >>>>
> >>>> eth1      Link encap:Ethernet  HWaddr 08:00:27:df:00:00
> >>>>         inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:3910 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:478719 (478.7 KB)  TX bytes:2542459 (2.5 MB)
> >>>>
> >>>> lo        Link encap:Local Loopback
> >>>>         inet addr:127.0.0.1  Mask:255.0.0.0
> >>>>         inet6 addr: ::1/128 Scope:Host
> >>>>         UP LOOPBACK RUNNING  MTU:16436  Metric:1
> >>>>         RX packets:360285 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:360285 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:0
> >>>>         RX bytes:169128181 (169.1 MB)  TX bytes:169128181 (169.1 MB)
> >>>>
> >>>> vif1.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:6 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:152 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:292 (292.0 B)  TX bytes:9252 (9.2 KB)
> >>>>
> >>>> vif1.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:566 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:1405 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:44227 (44.2 KB)  TX bytes:173995 (173.9 KB)
> >>>>
> >>>> vif1.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:3 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:838 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:84 (84.0 B)  TX bytes:111361 (111.3 KB)
> >>>>
> >>>> vif4.0    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:64 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:197 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:10276 (10.2 KB)  TX bytes:18453 (18.4 KB)
> >>>>
> >>>> vif4.1    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:2051 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:2446 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:233914 (233.9 KB)  TX bytes:364243 (364.2 KB)
> >>>>
> >>>> vif4.2    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:3 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:582 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:84 (84.0 B)  TX bytes:74700 (74.7 KB)
> >>>>
> >>>> vif4.3    Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet6 addr: fe80::fcff:ffff:feff:ffff/64 Scope:Link
> >>>>         UP BROADCAST RUNNING NOARP PROMISC  MTU:1500  Metric:1
> >>>>         RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:585 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:32
> >>>>         RX bytes:0 (0.0 B)  TX bytes:74826 (74.8 KB)
> >>>>
> >>>> xapi0     Link encap:Ethernet  HWaddr fe:ff:ff:ff:ff:ff
> >>>>         inet addr:169.254.0.1  Bcast:169.254.255.255  Mask:255.255.0.0
> >>>>         inet6 addr: fe80::c870:1aff:fec2:22b/64 Scope:Link
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:568 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:1132 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:0
> >>>>         RX bytes:76284 (76.2 KB)  TX bytes:109085 (109.0 KB)
> >>>>
> >>>> xenbr0    Link encap:Ethernet  HWaddr 08:00:27:7e:74:9c
> >>>>         inet addr:192.168.56.15  Bcast:192.168.56.255
> >> Mask:255.255.255.0
> >>>>         inet6 addr: fe80::a00:27ff:fe7e:749c/64 Scope:Link
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:4162 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:3281 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:0
> >>>>         RX bytes:469199 (469.1 KB)  TX bytes:485688 (485.6 KB)
> >>>>
> >>>> xenbr1    Link encap:Ethernet  HWaddr 08:00:27:df:00:00
> >>>>         inet addr:10.0.3.15  Bcast:10.0.3.255  Mask:255.255.255.0
> >>>>         inet6 addr: fe80::a00:27ff:fedf:0/64 Scope:Link
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:4129 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:3114 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:0
> >>>>         RX bytes:404327 (404.3 KB)  TX bytes:2501443 (2.5 MB)
> >>>>
> >>>> These physical NICs on the host translate to the following Xen PIFs:
> >>>>
> >>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# xe pif-list
> >>>> uuid ( RO)                  : 207413c9-5058-7a40-6c96-2dab21057f30
> >>>>               device ( RO): eth1
> >>>>   currently-attached ( RO): true
> >>>>                 VLAN ( RO): -1
> >>>>         network-uuid ( RO): 1679ddb1-5a21-b827-ab07-c16275d5ce72
> >>>>
> >>>>
> >>>> uuid ( RO)                  : c0274787-e768-506f-3191-f0ac17b0c72b
> >>>>               device ( RO): eth0
> >>>>   currently-attached ( RO): true
> >>>>                 VLAN ( RO): -1
> >>>>         network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
> >>>>
> >>>> The following is the ifconfig from the SSVM:
> >>>>
> >>>> root@s-5-TEST:~# ifconfig
> >>>> eth0      Link encap:Ethernet  HWaddr 0e:00:a9:fe:03:8b
> >>>>         inet addr:169.254.3.139  Bcast:169.254.255.255
> >> Mask:255.255.0.0
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:235 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:92 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:21966 (21.4 KiB)  TX bytes:16404 (16.0 KiB)
> >>>>         Interrupt:8
> >>>>
> >>>> eth1      Link encap:Ethernet  HWaddr 06:bc:62:00:00:05
> >>>>         inet addr:192.168.56.104  Bcast:192.168.56.255
> >>>> Mask:255.255.255.0
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:2532 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:2127 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:341242 (333.2 KiB)  TX bytes:272183 (265.8 KiB)
> >>>>         Interrupt:10
> >>>>
> >>>> eth2      Link encap:Ethernet  HWaddr 06:12:72:00:00:37
> >>>>         inet addr:10.0.3.204  Bcast:10.0.3.255  Mask:255.255.255.0
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:600 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:3 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:68648 (67.0 KiB)  TX bytes:126 (126.0 B)
> >>>>         Interrupt:11
> >>>>
> >>>> eth3      Link encap:Ethernet  HWaddr 06:25:e2:00:00:15
> >>>>         inet addr:192.168.56.120  Bcast:192.168.56.255
> >>>> Mask:255.255.255.0
> >>>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>>         RX packets:603 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:1000
> >>>>         RX bytes:68732 (67.1 KiB)  TX bytes:0 (0.0 B)
> >>>>         Interrupt:12
> >>>>
> >>>> lo        Link encap:Local Loopback
> >>>>         inet addr:127.0.0.1  Mask:255.0.0.0
> >>>>         UP LOOPBACK RUNNING  MTU:16436  Metric:1
> >>>>         RX packets:61 errors:0 dropped:0 overruns:0 frame:0
> >>>>         TX packets:61 errors:0 dropped:0 overruns:0 carrier:0
> >>>>         collisions:0 txqueuelen:0
> >>>>         RX bytes:5300 (5.1 KiB)  TX bytes:5300 (5.1 KiB)
> >>>>
> >>>> Finally, the following are the vif params for the eth2 device on the
> >> SSVM
> >>>> depicting its connection to eth0 instead of eth1:
> >>>>
> >>>> root@zone1:/opt/cloudstack/apache-tomcat-6.0.32/bin# !1243
> >>>> xe vif-param-list uuid=be44bb30-5700-b461-760e-10fe93079210
> >>>> uuid ( RO)                        :
> be44bb30-5700-b461-760e-10fe93079210
> >>>>                    vm-uuid ( RO): 7958d91f-e52d-a25d-718c-7f831ae701d7
> >>>>              vm-name-label ( RO): s-5-TEST
> >>>>         allowed-operations (SRO): attach; unplug_force; unplug
> >>>>         current-operations (SRO):
> >>>>                     device ( RO): 2
> >>>>                        MAC ( RO): 06:12:72:00:00:37
> >>>>          MAC-autogenerated ( RO): false
> >>>>                        MTU ( RO): 1500
> >>>>         currently-attached ( RO): true
> >>>>         qos_algorithm_type ( RW): ratelimit
> >>>>       qos_algorithm_params (MRW): kbps: 25600
> >>>>   qos_supported_algorithms (SRO):
> >>>>               other-config (MRW): nicira-iface-id:
> >>>> 3d68b9f8-98d1-4ac7-92d8-fb57cb8b0adc; nicira-vm-id:
> >>>> 7958d91f-e52d-a25d-718c-7f831ae701d7
> >>>>               network-uuid ( RO): 8ee927b1-a35d-ac10-4471-d7a6a475839a
> >>>>         network-name-label ( RO): Pool-wide network associated with
> >> eth0
> >>>>                io_read_kbs ( RO): 0.007
> >>>>               io_write_kbs ( RO): 0.000
> >>>>
> >>>> How do I configure CloudStack such that the guest network NIC on the
> VM
> >>>> will be connected to correct physical NIC?
> >>>>
> >>>> Thanks for your help,
> >>>> -John
> >>>>
> >>>>
> >>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-590
> >>>>
> >>>> On Dec 5, 2012, at 2:47 PM, Marcus Sorensen <shadow...@gmail.com>
> >> wrote:
> >>>>
> >>>>> Yes, see your cmdline. internaldns1=10.0.3.2, so it is forcing the
> use
> >> of
> >>>>> management network to route to 10.0.3.2 for DNS. that's where the
> route
> >>>> is
> >>>>> coming from. you will want to use something on your management net
> for
> >>>>> internal DNS, or something other than that router.
> >>>>>
> >>>>>
> >>>>> On Wed, Dec 5, 2012 at 11:59 AM, John Burwell <jburw...@basho.com>
> >>>> wrote:
> >>>>>
> >>>>>> Anthony,
> >>>>>>
> >>>>>> I apologize for forgetting to response to the part of your answer
> the
> >>>>>> first part of the question.  I had set the management.network.cidr
> and
> >>>> host
> >>>>>> global settings to 192.168.0.0/24 and 192.168.56.18 respectively.
> >>>> Please
> >>>>>> see the zone1.devcloud.cfg Marvin configuration attached to my
> >> original
> >>>>>> email for the actual setting, as well as, the network configurations
> >>>> used
> >>>>>> when this problem occurs.
> >>>>>>
> >>>>>> Thanks,
> >>>>>> -John
> >>>>>>
> >>>>>> On Dec 5, 2012, at 12:46 PM, Anthony Xu <xuefei...@citrix.com>
> wrote:
> >>>>>>
> >>>>>>> Hi join,
> >>>>>>>
> >>>>>>> Try following,
> >>>>>>>
> >>>>>>> Set global configuration management.network.cidr to your management
> >>>>>> server CIDR, if this configuration is not available in UI, you can
> >>>> change
> >>>>>> it in DB directly.
> >>>>>>>
> >>>>>>> Restart management,
> >>>>>>> Stop/Start SSVM and CPVM.
> >>>>>>>
> >>>>>>>
> >>>>>>> And could you post "cat /proc/cmdline" in SSVM?
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> Anthony
> >>>>>>>
> >>>>>>>> -----Original Message-----
> >>>>>>>> From: John Burwell [mailto:jburw...@basho.com]
> >>>>>>>> Sent: Wednesday, December 05, 2012 9:11 AM
> >>>>>>>> To: cloudstack-dev@incubator.apache.org
> >>>>>>>> Subject: Re: SSVM Network Configuration Issue
> >>>>>>>>
> >>>>>>>> All,
> >>>>>>>>
> >>>>>>>> I was wondering if anyone else is experiencing this problem when
> >> using
> >>>>>>>> secondary storage on a devcloud-style VM with a host-only and NAT
> >>>>>>>> adapter.  One aspect of this issue that seems interesting is that
> >>>>>>>> following route table from the SSVM:
> >>>>>>>>
> >>>>>>>> root@s-5-TEST:~# route
> >>>>>>>> Kernel IP routing table
> >>>>>>>> Destination     Gateway         Genmask         Flags Metric Ref
> >>>> Use
> >>>>>>>> Iface
> >>>>>>>> 10.0.3.2        192.168.56.1    255.255.255.255 UGH   0      0
> >>>> 0
> >>>>>>>> eth1
> >>>>>>>> 10.0.3.0        *               255.255.255.0   U     0      0
> >>>> 0
> >>>>>>>> eth2
> >>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
> >>>> 0
> >>>>>>>> eth1
> >>>>>>>> 192.168.56.0    *               255.255.255.0   U     0      0
> >>>> 0
> >>>>>>>> eth3
> >>>>>>>> link-local      *               255.255.0.0     U     0      0
> >>>> 0
> >>>>>>>> eth0
> >>>>>>>> default         10.0.3.2        0.0.0.0         UG    0      0
> >>>> 0
> >>>>>>>> eth2
> >>>>>>>>
> >>>>>>>> In particular, the gateways for the management and guest networks
> do
> >>>>>>>> not match to the configuration provided to the management server
> >> (i.e.
> >>>>>>>> 10.0.3.2 is the gateway for the 10.0.3.0/24 network and
> >> 192.168.56.1
> >>>> is
> >>>>>>>> the gateway for the 192.168.56.0/24 network).  With this
> >>>> configuration,
> >>>>>>>> the SSVM has a socket connection to the management server, but is
> in
> >>>>>>>> alert state.  Finally, when I remove the host-only NIC and use
> only
> >> a
> >>>>>>>> NAT adapter the SSVM's networking works as expecting leading me to
> >>>>>>>> believe that the segregated network configuration is at the root
> of
> >>>> the
> >>>>>>>> problem.
> >>>>>>>>
> >>>>>>>> Until I can get the networking on the SSVM configured, I am unable
> >> to
> >>>>>>>> complete the testing of the S3-backed Secondary Storage
> enhancement.
> >>>>>>>>
> >>>>>>>> Thank you for your help,
> >>>>>>>> -John
> >>>>>>>>
> >>>>>>>> On Dec 3, 2012, at 4:46 PM, John Burwell <jburw...@basho.com>
> >> wrote:
> >>>>>>>>
> >>>>>>>>> All,
> >>>>>>>>>
> >>>>>>>>> I am setting up a multi-zone devcloud configuration on VirtualBox
> >>>>>>>> 4.2.4 using the Ubuntu 12.04.1 and Xen 4.1.  I have configured the
> >>>> base
> >>>>>>>> management server VM (zone1) to serve as both the zone1, as well
> as,
> >>>>>>>> the management server (running MySql) with eth0 as a host-only
> >> adapter
> >>>>>>>> and a static IP of 192.168.56.15 and eth1 as a NAT adapter (see
> the
> >>>>>>>> attached zone1-interfaces file for the exact network configuration
> >> on
> >>>>>>>> the VM).  The management and guest networks are configured as
> >> follows:
> >>>>>>>>>
> >>>>>>>>> Zone 1
> >>>>>>>>> Management: 192.168.56.100-149 gw 192.168.56.1 dns 10.0.3.2 (?)
> >>>>>>>>> Guest: 10.0.3.200-10.0.3.220 gw 10.0.3.2 dns 8.8.8.8
> >>>>>>>>> Zone 2
> >>>>>>>>> Management: 192.168.56.150-200 gw 192.68.56.1 dns 10.0.3.2 (?)
> >>>>>>>>> Guest: 10.0.3.221-240 gw 10.0.3.2 dns 8.8.8.8
> >>>>>>>>>
> >>>>>>>>> The management server deploys and starts without error.  I then
> >>>>>>>> populate the configuration it using the attached Marvin
> >> configuration
> >>>>>>>> file (zone1.devcloud.cfg) and restart the management server in
> order
> >>>> to
> >>>>>>>> allow the global configuration option changes to take effect.
> >>>>>>>> Following the restart, the CPVM and SSVM start without error.
> >>>>>>>> Unfortunately, they drop into alert status, and the SSVM is unable
> >> to
> >>>>>>>> connect outbound through the guest network (very important for my
> >>>> tests
> >>>>>>>> because I am testing S3-backed secondary storage).
> >>>>>>>>>
> >>>>>>>>> From the diagnostic checks I have performed on the management
> >> server
> >>>>>>>> and the SSVM, it appears that the daemon on the SSVM is connecting
> >>>> back
> >>>>>>>> to the management server.  I have attached a set of diagnostic
> >>>>>>>> information from the management server
> >> (mgmtsvr-zone1-diagnostics.log)
> >>>>>>>> and SSVM server (ssvm-zone1-diagnostics.log) that includes the
> >> results
> >>>>>>>> of ifconfig, route, netstat and ping checks, as well as, other
> >>>>>>>> information (e.g. the contents of /var/cache/cloud/cmdline on the
> >>>> SSVM).
> >>>>>>>> Finally, I have attached the vmops log from the management server
> >>>>>>>> (vmops-zone1.log).
> >>>>>>>>>
> >>>>>>>>> What changes need to be made to management server configuration
> in
> >>>>>>>> order to start up an SSVM that can communicate with the secondary
> >>>>>>>> storage NFS volumes, management server, and connect to hosts on
> the
> >>>>>>>> Internet?
> >>>>>>>>>
> >>>>>>>>> Thanks for your help,
> >>>>>>>>> -John
> >>>>>>>>>
> >>>>>>>>> <ssvm-zone1-diagnostics.log>
> >>>>>>>>> <vmops-zone1.tar.gz>
> >>>>>>>>> <mgmtsvr-zone1-diagnostics.log>
> >>>>>>>>> <zone1-interfaces>
> >>>>>>>>> <zone1.devcloud.cfg>
> >>>>>>>
> >>>>>>
> >>>>>>
> >>>>
> >>>>
> >>
> >>
>
>

Reply via email to