Do you open all these nodes one by one on every Security Group in each
region every time you add a node or did you manage to automate it somehow ?


2013/6/5 Dan Kogan <d...@iqtell.com>

> Hi,****
>
> ** **
>
> We are using a very similar configuration.  From our experience, Cassandra
> nodes in the same DC need access over both public and private IP on the
> storage port (7000/7001).  Nodes from other DC will need access over public
> IP on the storage port.****
>
> All Cassandra nodes also need access over the public IP on the Thrift port
> (9160).****
>
> ** **
>
> Dan****
>
> ** **
>
> *From:* Alain RODRIGUEZ [mailto:arodr...@gmail.com]
> *Sent:* Wednesday, June 05, 2013 9:49 AM
> *To:* user@cassandra.apache.org
> *Subject:* Looking for a fully working AWS multi DC configuration.****
>
> ** **
>
> Hi,****
>
> ** **
>
> We use to work on a single DC (EC2Snitch / SimpleStrategy). For latency
> reason we had top open a new DC in the US (us-east). We run C* 1.2.2. We
> don't use VPC.****
>
> ** **
>
> Now we use:****
>
> - 2 DC (eu-west, us-east)****
>
> - EC2MultiRegionSnitch / NTS****
>
> - public IPs as broadcast_address and seeds****
>
> - private IPs as listen_address****
>
> ** **
>
> Yet we are experimenting some troubles (node can't reach itself, Could not
> start register mbean in JMX...), mainly because of the use of public IPs
> and the AWS inter-region communication.****
>
> ** **
>
> If someone has successfully setup this kind of cluster, I would like to
> know, if our configuration is correct and if I am missing something.****
>
> ** **
>
> I also would like to know what ports I have to open and either where I
> have to open them from.****
>
> ** **
>
> Any insight would be greatly appreciated.****
>

Reply via email to