I don't want to have a single load balancer because I want to avoid a single 
point of failure. And we'll be pushing enough data that it would be a huge 
bottleneck.

A failed node will not receive new requests, but when the requests that were 
sent to it fail I'd like to retry those automatically instead of having errors 
bubble up to our application.
On Thursday, April 7, 2011 at 8:17 PM, Wilson MacGyver wrote: 
> use haproxy to ping each of the riak node. and remove nodes that
> aren't responding is
> what we did.
> 
> we use a single haproxy instead of 1 per application server.
> 
> On Thu, Apr 7, 2011 at 9:47 PM, Greg Nelson <gro...@dropcam.com> wrote:
> > Hello,
> > I have a simple three node cluster that I have been using for testing and
> > benchmarking Riak. Lately I've been simulating various failure scenarios --
> > like a node going down, disk going bad, etc.
> > My application talks to Riak through an haproxy instance running locally on
> > each application server. It's configured to round-robin over the nodes in
> > the cluster for both HTTP and PBC interfaces, and uses the HTTP /ping health
> > check. I assume this is a rather typical setup.
> 
> -- 
> Omnem crede diem tibi diluxisse supremum.
> 
_______________________________________________
riak-users mailing list
riak-users@lists.basho.com
http://lists.basho.com/mailman/listinfo/riak-users_lists.basho.com

Reply via email to