Tnx Remi!

Brilliant advice!

чт, 5 окт. 2017 г. в 15:38, Remi Bergsma <rberg...@schubergphilis.com>:

> Hi,
>
> We solved this problem by splitting the network in an underlay and overlay
> network. The underlay is the physical network, including the management
> traffic and storage from hypervisors and such. The simpler, the better. In
> the overlay there’s your services layer, for example the guest networks for
> your clients. Since you’re already using vxlan that shouldn’t be hard. In
> our setup, each POD is a rack with TOR (Top-of-rack routing) which means a
> L2 domain stays within a rack. If that goes wrong, only one rack aka POD
> has issues and not the rest. We’ve many PODs in several zones. The overlay
> makes tunnels (we use Nicira but vxlan, ovn or nuage can do the same) and
> those can be created over L3 interconnected PODs. Effectively, this gives
> you L2 (overlay) over L3 (underlay). Storage wise we have both cluster-wide
> (in the POD) and zone-wide (basically a POD with just storage).
>
> VMs in a given guest network can run in any of the PODs and still have a
> L2 connection between then, even though the actual physical network is L3.
> This is one of the great benefits of SDN (Software defined networking).
> It’s pretty much the best of both worlds. We scaled this quite a bit and
> it’s rock solid.
>
> Regards,
> Remi
>
>
> On 05/10/2017, 13:57, "Rafael Weingärtner" <raf...@autonomiccs.com.br>
> wrote:
>
>     Exactly; the management IPs is defined per POD already, the public you
>     could work out dedicating domains per POD, and then you can dedicate a
>     pool of IPs for each domain. The guest networking problem is solved if
>     you force users from let´s say same domain to say in the same Pod.
>
>     The other approach as you said would be a zone per POD.
>
>     Please keep us posted with your tests, your findings may be valuable to
>     spot improvement in ACS design and help others with more complex
>     deployments.
>
>
>     On 10/5/2017 6:51 AM, Andrija Panic wrote:
>     > Thanks Rafael,
>     >
>     > yes that is my expectation also (same broadcast domain for Guest
> network),
>     > so it doesn't really solve my problem (identical thing is expected
> for
>     > Public Network, at least, if not other networks also)
>     > Other options seems to be zones per each X racks...
>     >
>     > Will see.
>     >
>     > Thanks
>     >
>     > On 4 October 2017 at 22:25, Rafael Weingärtner <
> raf...@autonomiccs.com.br>
>     > wrote:
>     >
>     >> I think this can cause problems, if not properly managed. Unless you
>     >> concentrate Domains/Users in Pods. Otherwise, you might end up with
> some
>     >> VMs of the same user/domain/project in different pods, and if they
> are all
>     >> in the same VPC for instance, we would expect them to be in the same
>     >> broadcast domain.
>     >>
>     >> I think to apply what you want, it may require some designing and
> testing,
>     >> but it feels feasible with ACS.
>     >>
>     >>
>     >> On 10/4/2017 5:19 PM, Andrija Panic wrote:
>     >>
>     >>> Anyone?  I know I'm trying to squeeze some free paid consulting
> here :),
>     >>> but trying to understand if PODs makes sense in this situation....
>     >>>
>     >>> Thx
>     >>>
>     >>> On 2 October 2017 at 10:21, Andrija Panic <andrija.pa...@gmail.com
> >
>     >>> wrote:
>     >>>
>     >>> Hi guys,
>     >>>> Sorry for long post below...
>     >>>>
>     >>>> I was wondering if someone could bring some light for me for
> multiple
>     >>>> PODs
>     >>>> networking design (L2 vs L3) - idea is to make smaller L2
> broadcast
>     >>>> domains
>     >>>> (any other reason?)
>     >>>>
>     >>>> We might decide to transition from current single pod, single
> cluster
>     >>>> (single zone) to multiple PODs design (or not...) - we will
> eventually
>     >>>> grow
>     >>>> to over 50 racks worth of KVM hosts (1000+ hosts) so Im trying to
>     >>>> understand best options to avoid having insanely huge L2 broadcast
>     >>>> domains...
>     >>>>
>     >>>> Mgmt network is routed between pods, that is clear.
>     >>>>
>     >>>> We have dedicated primary storage network and Secondary Storage
> networks
>     >>>> (vlan interfaces configured locally on all KVM hosts, providing
> direct L2
>     >>>> connection obviously, not shared with mgmt.network), and same for
> Public
>     >>>> and Guest networks... (Advanced networking in zone, Vxlan used as
>     >>>> isolation)
>     >>>>
>     >>>> Now with multiple PODs, since Public Network and Guest network is
> defined
>     >>>> per Zone level (not POD level), and currently same zone-wide
> setup for
>     >>>> Primary Storage... what would be the best way to make this
> traffic stay
>     >>>> inside PODs as much as possible and is this possible at all?
> Perhaps I
>     >>>> would need to look into multiple zones, not PODs.
>     >>>>
>     >>>> My humble conclusion, based on having all dedicated networks, is
> that I
>     >>>> need to strech (L2 attach as vlan interface) primary and secondary
>     >>>> storage
>     >>>> network across all racks/PODs, and also need to strech Guest vlan
> (that
>     >>>> carry all Guest VXLAN tunnels), and again same for Public
> Network...and
>     >>>> this again makes huge broadcast domains and doesn't solve my
> issue...
>     >>>> Don't see other option in my head to make networking work across
> PODs.
>     >>>>
>     >>>> Any suggestion is most welcome (and if of any use as info - we
> dont plan
>     >>>> for any Xen, VmWare etc, will stay purely with KVM).
>     >>>>
>     >>>> Thanks
>     >>>> Andrija
>     >>>>
>     >>>>
>     >>>
>     >> --
>     >> Rafael Weingärtner
>     >>
>     >>
>     >
>
>     --
>     Rafael Weingärtner
>
>
>
>

Reply via email to