Ok but why direct routing instead of NAT? If the public IPs are _only_ on LVS there is no point to use LVS-DR.
LVS has the public IPs and redirects to the private IPs, this _must_ work. Did you try NAT? Or at least can you give it a shot? -- Regards, Sébastien Han. On Fri, Feb 15, 2013 at 3:55 PM, Samuel Winchenbach <swinc...@gmail.com> wrote: > Sure... I have undone these settings but I saved a copy: > > two hosts: > test1 eth0: 10.21.0.1/16 eth1: 130.x.x.x/24 > test2 eth0: 10.21.0.2/16 eth1: 130.x.x.x/24 > > VIP: 10.21.21.1 (just for testing, later I would add a 130.x.x.x/24 VIP for > public APIs > > k > eystone is bound to 10.21.0.1 on test1 and 10.21.0.2 on test2 > > > > in /etc/sysctl.conf: > net.ipv4.conf.all.arp_ignore = 1 > net.ipv4.conf.eth0.arp_ignore = 1 > net.ipv4.conf.all.arp_announce = 2 > net.ipv4.conf.eth0.arp_announce = 2 > > root# sysctl -p > > in /etc/sysctl.conf: > > checktimeout= > 3 > > > checkinterval= > 5 > > > autoreload= > yes > > > logfile="/var/log/ldirectord.log" > > quiescent=no > > virtual=10.21.21.1:5000 > > real=10.2 > 1 > .0.1:5000 gate > > real=10.2 > 1 > .0.2:5000 gate > > scheduler= > w > rr > protocol=tcp > checktype=connect > checkport=5000 > > virtual=10.21.21.1: > 35357 > > real=10.2 > 1 > .0.1: > 35357 > gate > > real=10.2 > 1 > .0.2: > 35357 > gate > > scheduler= > w > rr > protocol=tcp > checktype=connect > checkport=35357 > > > crm shell: > > > primitive > p_openstack_ > ip ocf:heartbeat:IPaddr2 \ > > > op monitor interval="60" timeout="20" \ > > > params ip= > "10.21.21.1 > " > cidr_netmask=" > 16 > " > lvs_support="true" > > p > rimitive > p_openstack_ip_lo > ocf:heartbeat:IPaddr2 \ > > > op monitor interval="60" timeout="20" \ > > > params ip=" > 10.21.21.1 > " nic="lo" > cidr_netmask="32" > > primitive > p_openstack_ > lvs ocf:heartbeat:ldirectord \ > > > op monitor interval="20" timeout="10" > > group > g_openstack_ > ip > _ > lvs > p_openstack_ > ip > p_openstack_ > lvs > > clone > c_openstack_ip_lo > > p_openstack_ip_lo > meta interleave="true" > > colocation > co_openstack_lo_never_lvs > -inf: c > _openstack_ip_lo > > g_openstack_ip_lvs > > Thanks for taking a look at this. > > Sam > > > > > On Fri, Feb 15, 2013 at 3:54 AM, Sébastien Han <han.sebast...@gmail.com> > wrote: >> >> Hum I don't see the problem, it's possible to load-balance VIPs with LVS, >> there are just IPs... Can I see your conf? >> >> -- >> Regards, >> Sébastien Han. >> >> >> On Thu, Feb 14, 2013 at 8:34 PM, Samuel Winchenbach <swinc...@gmail.com> >> wrote: >>> >>> W >>> ell, I think I will have to go with one ip per service and forget about >>> load balancing. It seems as though with LVS routing requests internally >>> through the VIP is difficult (impossible?) at least with LVS-DR. It seems >>> like a shame not to be able to distribute the work among the controller >>> nodes. >>> >>> >>> On Thu, Feb 14, 2013 at 9:50 AM, Samuel Winchenbach <swinc...@gmail.com> >>> wrote: >>>> >>>> Hi Sébastien, >>>> >>>> I have two hosts with public interfaces with a number (~8) compute nodes >>>> behind them. I am trying to set the two public nodes in for HA and load >>>> balancing, I plan to run all the openstack services on these two nodes in >>>> Active/Active where possible. I currently have MySQL and RabbitMQ setup >>>> in >>>> pacemaker with a drbd backend. >>>> >>>> That is a quick summary. If there is anything else I can answer about >>>> my setup please let me know. >>>> >>>> Thanks, >>>> Sam >>>> >>>> >>>> On Thu, Feb 14, 2013 at 9:26 AM, Sébastien Han <han.sebast...@gmail.com> >>>> wrote: >>>>> >>>>> Well I don't know your setup, if you use LB for API service or if you >>>>> use an active/passive pacemaker but at the end it's not that much IPs I >>>>> guess. I dare to say that Keepalived sounds outdated to me... >>>>> >>>>> If you use pacemaker and want to have the same IP for all the resources >>>>> simply create a resource group with all the openstack service inside it >>>>> (it's ugly but if it's what you want :)). Give me more info about your >>>>> setup >>>>> and we can go further in the discussion :). >>>>> >>>>> -- >>>>> Regards, >>>>> Sébastien Han. >>>>> >>>>> >>>>> On Thu, Feb 14, 2013 at 3:15 PM, Samuel Winchenbach >>>>> <swinc...@gmail.com> wrote: >>>>>> >>>>>> T >>>>>> he only real problem is that it would consume a lot of IP addresses >>>>>> when exposing the public interfaces. I _think_ I may have the solution >>>>>> in >>>>>> your blog actually: >>>>>> http://www.sebastien-han.fr/blog/2012/10/19/highly-available-lvs/ >>>>>> and >>>>>> http://clusterlabs.org/wiki/Using_ldirectord >>>>>> >>>>>> I am trying to weigh the pros and cons of this method vs >>>>>> keepalived/haproxy and just biting the bullet and using one IP per >>>>>> service. >>>>>> >>>>>> >>>>>> On Thu, Feb 14, 2013 at 4:17 AM, Sébastien Han >>>>>> <han.sebast...@gmail.com> wrote: >>>>>>> >>>>>>> What's the problem to have one IP on service pool basis? >>>>>>> >>>>>>> -- >>>>>>> Regards, >>>>>>> Sébastien Han. >>>>>>> >>>>>>> >>>>>>> On Wed, Feb 13, 2013 at 8:45 PM, Samuel Winchenbach >>>>>>> <swinc...@gmail.com> wrote: >>>>>>>> >>>>>>>> What if the VIP is created on a different host than keystone is >>>>>>>> started on? It seems like you either need to set >>>>>>>> net.ipv4.ip_nonlocal_bind >>>>>>>> = 1 or create a colocation in pacemaker (which would either require all >>>>>>>> services to be on the same host, or have an ip-per-service). >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Wed, Feb 13, 2013 at 2:28 PM, Razique Mahroua >>>>>>>> <razique.mahr...@gmail.com> wrote: >>>>>>>>> >>>>>>>>> There we go >>>>>>>>> https://review.openstack.org/#/c/21581/ >>>>>>>>> >>>>>>>>> Razique Mahroua - Nuage & Co >>>>>>>>> razique.mahr...@gmail.com >>>>>>>>> Tel : +33 9 72 37 94 15 >>>>>>>>> >>>>>>>>> >>>>>>>>> Le 13 févr. 2013 à 20:15, Razique Mahroua >>>>>>>>> <razique.mahr...@gmail.com> a écrit : >>>>>>>>> >>>>>>>>> I'm currently updating that part of the documentation - indeed it >>>>>>>>> states that two IPs are used, but in fact, you end up with only one >>>>>>>>> VIP for >>>>>>>>> the API service. >>>>>>>>> I'll send the patch tonight >>>>>>>>> >>>>>>>>> Razique Mahroua - Nuage & Co >>>>>>>>> razique.mahr...@gmail.com >>>>>>>>> Tel : +33 9 72 37 94 15 >>>>>>>>> >>>>>>>>> <NUAGECO-LOGO-Fblan_petit.jpg> >>>>>>>>> >>>>>>>>> Le 13 févr. 2013 à 20:05, Samuel Winchenbach <swinc...@gmail.com> a >>>>>>>>> écrit : >>>>>>>>> >>>>>>>>> In that documentation it looks like each openstack service gets it >>>>>>>>> own IP (keystone is being assigned 192.168.42.103 and glance is >>>>>>>>> getting >>>>>>>>> 192.168.42.104). >>>>>>>>> >>>>>>>>> I might be missing something too because in the section titled >>>>>>>>> "Configure the VIP" it create a primitive called "p_api-ip" (or >>>>>>>>> p_ip_api if >>>>>>>>> you read the text above it) and then in "Adding Keystone resource to >>>>>>>>> Pacemaker" it creates a group with "p_ip_keystone"??? >>>>>>>>> >>>>>>>>> >>>>>>>>> Stranger yet, "Configuring OpenStack Services to use High Available >>>>>>>>> Glance API" says: "For Nova, for example, if your Glance API service >>>>>>>>> IP >>>>>>>>> address is 192.168.42.104 as in the configuration explained here, you >>>>>>>>> would >>>>>>>>> use the following line in your nova.conf file : glance_api_servers = >>>>>>>>> 192.168.42.103" But, in the step before it set: "registry_host = >>>>>>>>> 192.168.42.104"? >>>>>>>>> >>>>>>>>> So I am not sure which ip you would connect to here... >>>>>>>>> >>>>>>>>> Sam >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Wed, Feb 13, 2013 at 1:29 PM, JuanFra Rodriguez Cardoso >>>>>>>>> <juanfra.rodriguez.card...@gmail.com> wrote: >>>>>>>>>> >>>>>>>>>> Hi Samuel: >>>>>>>>>> >>>>>>>>>> Yes, it's possible with pacemaker. Look at >>>>>>>>>> http://docs.openstack.org/trunk/openstack-ha/content/ch-intro.html. >>>>>>>>>> >>>>>>>>>> Regards, >>>>>>>>>> JuanFra >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> 2013/2/13 Samuel Winchenbach <swinc...@gmail.com> >>>>>>>>>>> >>>>>>>>>>> Hi All, >>>>>>>>>>> >>>>>>>>>>> I currently have a HA OpenStack cluster running where the >>>>>>>>>>> OpenStack services are kept alive with a combination of haproxy and >>>>>>>>>>> keepalived. >>>>>>>>>>> >>>>>>>>>>> Is it possible to configure pacemaker so that all the OpenStack >>>>>>>>>>> services are served by the same IP? With keepalived I have a >>>>>>>>>>> virtual ip >>>>>>>>>>> that can move from server to server and haproxy sends the request >>>>>>>>>>> to a >>>>>>>>>>> machine that has a "live" service. This allows one (public) ip to >>>>>>>>>>> handle >>>>>>>>>>> all incoming requests. I believe it is the combination of >>>>>>>>>>> VRRP/IPVS that >>>>>>>>>>> allows this. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Is it possible to do something similar with pacemaker? I really >>>>>>>>>>> don't want to have an IP for each service, and I don't want to make >>>>>>>>>>> it a >>>>>>>>>>> requirement that all OpenStack services must be running on the same >>>>>>>>>>> server. >>>>>>>>>>> >>>>>>>>>>> Thanks... I hope this question is clear, I feel like I sort of >>>>>>>>>>> butchered the wording a bit. >>>>>>>>>>> >>>>>>>>>>> Sam >>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>>>> Post to : openstack@lists.launchpad.net >>>>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>>> Post to : openstack@lists.launchpad.net >>>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> _______________________________________________ >>>>>>>> Mailing list: https://launchpad.net/~openstack >>>>>>>> Post to : openstack@lists.launchpad.net >>>>>>>> Unsubscribe : https://launchpad.net/~openstack >>>>>>>> More help : https://help.launchpad.net/ListHelp >>>>>>>> >>>>>>> >>>>>> >>>>> >>>> >>> >> > _______________________________________________ Mailing list: https://launchpad.net/~openstack Post to : openstack@lists.launchpad.net Unsubscribe : https://launchpad.net/~openstack More help : https://help.launchpad.net/ListHelp