Tnx Remi! Brilliant advice!
чт, 5 окт. 2017 г. в 15:38, Remi Bergsma <rberg...@schubergphilis.com>: > Hi, > > We solved this problem by splitting the network in an underlay and overlay > network. The underlay is the physical network, including the management > traffic and storage from hypervisors and such. The simpler, the better. In > the overlay there’s your services layer, for example the guest networks for > your clients. Since you’re already using vxlan that shouldn’t be hard. In > our setup, each POD is a rack with TOR (Top-of-rack routing) which means a > L2 domain stays within a rack. If that goes wrong, only one rack aka POD > has issues and not the rest. We’ve many PODs in several zones. The overlay > makes tunnels (we use Nicira but vxlan, ovn or nuage can do the same) and > those can be created over L3 interconnected PODs. Effectively, this gives > you L2 (overlay) over L3 (underlay). Storage wise we have both cluster-wide > (in the POD) and zone-wide (basically a POD with just storage). > > VMs in a given guest network can run in any of the PODs and still have a > L2 connection between then, even though the actual physical network is L3. > This is one of the great benefits of SDN (Software defined networking). > It’s pretty much the best of both worlds. We scaled this quite a bit and > it’s rock solid. > > Regards, > Remi > > > On 05/10/2017, 13:57, "Rafael Weingärtner" <raf...@autonomiccs.com.br> > wrote: > > Exactly; the management IPs is defined per POD already, the public you > could work out dedicating domains per POD, and then you can dedicate a > pool of IPs for each domain. The guest networking problem is solved if > you force users from let´s say same domain to say in the same Pod. > > The other approach as you said would be a zone per POD. > > Please keep us posted with your tests, your findings may be valuable to > spot improvement in ACS design and help others with more complex > deployments. > > > On 10/5/2017 6:51 AM, Andrija Panic wrote: > > Thanks Rafael, > > > > yes that is my expectation also (same broadcast domain for Guest > network), > > so it doesn't really solve my problem (identical thing is expected > for > > Public Network, at least, if not other networks also) > > Other options seems to be zones per each X racks... > > > > Will see. > > > > Thanks > > > > On 4 October 2017 at 22:25, Rafael Weingärtner < > raf...@autonomiccs.com.br> > > wrote: > > > >> I think this can cause problems, if not properly managed. Unless you > >> concentrate Domains/Users in Pods. Otherwise, you might end up with > some > >> VMs of the same user/domain/project in different pods, and if they > are all > >> in the same VPC for instance, we would expect them to be in the same > >> broadcast domain. > >> > >> I think to apply what you want, it may require some designing and > testing, > >> but it feels feasible with ACS. > >> > >> > >> On 10/4/2017 5:19 PM, Andrija Panic wrote: > >> > >>> Anyone? I know I'm trying to squeeze some free paid consulting > here :), > >>> but trying to understand if PODs makes sense in this situation.... > >>> > >>> Thx > >>> > >>> On 2 October 2017 at 10:21, Andrija Panic <andrija.pa...@gmail.com > > > >>> wrote: > >>> > >>> Hi guys, > >>>> Sorry for long post below... > >>>> > >>>> I was wondering if someone could bring some light for me for > multiple > >>>> PODs > >>>> networking design (L2 vs L3) - idea is to make smaller L2 > broadcast > >>>> domains > >>>> (any other reason?) > >>>> > >>>> We might decide to transition from current single pod, single > cluster > >>>> (single zone) to multiple PODs design (or not...) - we will > eventually > >>>> grow > >>>> to over 50 racks worth of KVM hosts (1000+ hosts) so Im trying to > >>>> understand best options to avoid having insanely huge L2 broadcast > >>>> domains... > >>>> > >>>> Mgmt network is routed between pods, that is clear. > >>>> > >>>> We have dedicated primary storage network and Secondary Storage > networks > >>>> (vlan interfaces configured locally on all KVM hosts, providing > direct L2 > >>>> connection obviously, not shared with mgmt.network), and same for > Public > >>>> and Guest networks... (Advanced networking in zone, Vxlan used as > >>>> isolation) > >>>> > >>>> Now with multiple PODs, since Public Network and Guest network is > defined > >>>> per Zone level (not POD level), and currently same zone-wide > setup for > >>>> Primary Storage... what would be the best way to make this > traffic stay > >>>> inside PODs as much as possible and is this possible at all? > Perhaps I > >>>> would need to look into multiple zones, not PODs. > >>>> > >>>> My humble conclusion, based on having all dedicated networks, is > that I > >>>> need to strech (L2 attach as vlan interface) primary and secondary > >>>> storage > >>>> network across all racks/PODs, and also need to strech Guest vlan > (that > >>>> carry all Guest VXLAN tunnels), and again same for Public > Network...and > >>>> this again makes huge broadcast domains and doesn't solve my > issue... > >>>> Don't see other option in my head to make networking work across > PODs. > >>>> > >>>> Any suggestion is most welcome (and if of any use as info - we > dont plan > >>>> for any Xen, VmWare etc, will stay purely with KVM). > >>>> > >>>> Thanks > >>>> Andrija > >>>> > >>>> > >>> > >> -- > >> Rafael Weingärtner > >> > >> > > > > -- > Rafael Weingärtner > > > >