Sorry that only applies our you're using NTS. You're right that simple strategy won't work very well in this case. To migrate you'll likely need to do a DC migration to ensuite no downtime, as replica placement will change even if RF stays the same.
On 15 Sep. 2017 08:26, "kurt greaves" <k...@instaclustr.com> wrote: > If you have racks configured and lose nodes you should replace the node > with one from the same rack. You then need to repair, and definitely don't > decommission until you do. > > Also 40 nodes with 256 vnodes is not a fun time for repair. > > On 15 Sep. 2017 03:36, "Dominik Petrovic" <dominik.petro...@mail.ru.invalid> > wrote: > >> @jeff, >> I'm using 3 availability zones, during the life of the cluster we lost >> nodes, retired others and we end up having some of the data >> written/replicated on a single availability zone. We saw it with nodetool >> getendpoints. >> Regards >> >> >> Thursday, September 14, 2017 9:23 AM -07:00 from Jeff Jirsa < >> jji...@gmail.com>: >> >> With one datacenter/region, what did you discover in an outage you think >> you'll solve with network topology strategy? It should be equivalent for a >> single D.C. >> >> -- >> Jeff Jirsa >> >> >> On Sep 14, 2017, at 8:47 AM, Dominik Petrovic < >> dominik.petro...@mail.ru.INVALID> wrote: >> >> Thank you for the replies! >> >> @jeff my current cluster details are: >> 1 datacenter >> 40 nodes, with vnodes=256 >> RF=3 >> What is your advice? is it a production cluster, so I need to be very >> careful about it. >> Regards >> >> >> Thu, 14 Sep 2017 -2:47:52 -0700 from Jeff Jirsa <jji...@gmail.com>: >> >> The token distribution isn't going to change - the way Cassandra maps >> replicas will change. >> >> How many data centers/regions will you have when you're done? What's your >> RF now? You definitely need to run repair before you ALTER, but you've got >> a bit of a race here between the repairs and the ALTER, which you MAY be >> able to work around if we know more about your cluster. >> >> How many nodes >> How many regions >> How many replicas per region when you're done? >> >> >> >> >> -- >> Jeff Jirsa >> >> >> On Sep 13, 2017, at 2:04 PM, Dominik Petrovic < >> dominik.petro...@mail.ru.INVALID> wrote: >> >> Dear community, >> I'd like to receive additional info on how to modify a keyspace >> replication strategy. >> >> My Cassandra cluster is on AWS, Cassandra 2.1.15 using vnodes, the >> cluster's snitch is configured to Ec2Snitch, but the keyspace the >> developers created has replication class SimpleStrategy = 3. >> >> During an outage last week we realized the discrepancy between the >> configuration and we would now fix the issue using NetworkTopologyStrategy. >> >> What are the suggested steps to perform? >> For Cassandra 2.1 I found only this doc: http://docs.datastax.com/ >> en/cassandra/2.1/cassandra/operations/opsChangeKSStrategy.html >> that does not mention anything about repairing the cluster >> >> For Cassandra 3 I found this other doc: https://docs.datastax.com >> /en/cassandra/3.0/cassandra/operations/opsChangeKSStrategy.html >> That involves also the cluster repair operation. >> >> On a test cluster I tried the steps for Cassandra 2.1 but the token >> distribution in the ring didn't change so I'm assuming that wasn't the >> right think to do. >> I also perform a nodetool repair -pr but nothing changed as well. >> Some advice? >> >> -- >> Dominik Petrovic >> >> >> >> -- >> Dominik Petrovic >> >> >> >> -- >> Dominik Petrovic >> >