Use VPC Peering rather than VPN, More reliable. On Tue, Aug 11, 2015 at 5:14 PM, Prem Yadav <ipremya...@gmail.com> wrote:
> 1) There are ways to connect two VPCs using VPN. > 2) About the connectivity using public IP. Can you ping the one public ip > from another one in a different region. > If ping works, please check port connectivity using telnet. You can start > a temp server on a port using netcat. If connectivity fails, you need to > looks into your routing tables to allow connectivity on the public ip > addresses. > > On Tue, Aug 11, 2015 at 7:51 PM, Asher Newcomer <asher...@gmail.com> > wrote: > >> X-post w/ SO: link >> <https://stackoverflow.com/questions/31949043/cassandra-change-from-single-region-ec2-to-multi-region> >> >> I have (had) a working 4 node Cassandra cluster setup in an EC2 VPC. >> Setup was as follows: >> >> 172.18.100.110 - seed - DC1 / RAC1 >> >> 172.18.100.111 - DC1 / RAC1 >> >> 172.18.100.112 - seed - DC1 / RAC2 >> >> 172.18.100.113 - DC1 / RAC2 >> >> All of the above nodes are in East-1D, and I have configured it using the >> GossipingPropertyFileSnitch (I would rather not use the EC2 specific >> snitches). >> >> listen_address & broadcast_address were both set to the node's private IP. >> >> I then wanted to expand the cluster into a new region (us-west). Because >> cross-region private IP communication is not supported in EC2, I attempted >> to change the settings to have the nodes communicate through their public >> IPs. >> >> listen_address remained set to private IP >> broadcast_address was changed to the public IP >> seeds_list IPs were changed to the appropriate public IPs >> >> I restarted the nodes one by one expecting them to simply 'work', but now >> they only see themselves and not the other nodes. >> >> nodetool status consistently returns: >> >> Datacenter: DC1 >> =============== >> Status=Up/Down >> |/ State=Normal/Leaving/Joining/Moving >> -- Address Load Tokens Owns Host ID Rack >> DN 172.18.100.112 ? 256 ? 968aaa8a-32b7-4493-9747-3df1c3784164 r1 >> DN 172.18.100.113 ? 256 ? 8e03643c-9db8-4906-aabc-0a8f4f5c087d r1 >> UN [public IP of local node] 75.91 GB 256 ? >> 6fdcc85d-6c78-46f2-b41f-abfe1c86ac69 RAC1 >> DN 172.18.100.110 ? 256 ? fb7b78a8-d1cc-46fe-ab18-f0d3075cb426 r1 >> >> On each individual node, the other nodes seem 'stuck' using the private >> IP addresses. >> >> *How do I force the nodes to look for each other at their public >> addresses?* >> >> I have fully opened the EC2 security group/firewall as a test to rule out >> any problems there - and it hasn't helped. >> >> Any ideas most appreciated. >> > >