On Wed, Sep 25, 2013 at 12:41 PM, Skye Book <skye.b...@gmail.com> wrote:
> I have a three node cluster using the EC2 Multi-Region Snitch currently > operating only in US-EAST. On having a node go down this morning, I > started a new node with an identical configuration, except for the seed > list, the listen address and the rpc address. The new node comes up and > creates its own cluster rather than joining the pre-existing ring. I've > tried creating a node both *before* ad *after* using `nodetool remove` > for the bad node, each time with the same result. > What version of Cassandra? This particular confusing behavior is fixed upstream, in a version you should not deploy to production yet. Take some solace, however, that you may be the last Cassandra administrator to die for a broken code path! https://issues.apache.org/jira/browse/CASSANDRA-5768 Does anyone have any suggestions for where to look that might put me on the > right track? > It must be that your seed list is wrong in some way, or your node state is wrong. If you're trying to bootstrap a node, note that you can't bootstrap a node when it is in its own seed list. If you have installed Cassandra via debian package, there is a possibility that your node has started before you explicitly started it. If so, it might have invalid node state. Have you tried wiping the data directory and trying again? What is your seed list? Are you sure the new node can reach the seeds on the network layer? =Rob