Hello Joshua Thank you very much for the replies. I will go through the tickets you sent to understand the process.
Regards, Pradeep On Mon, Aug 27, 2018 at 6:15 AM, rajasekhar kommineni <rajaco...@gmail.com> wrote: > Hi Pradeep, > > For changing the snitch you have decommission and add the node with new > switch and updated properties files. > > Thanks, > > > On Aug 26, 2018, at 2:15 PM, Joshua Galbraith <jgalbra...@newrelic.com. > INVALID> wrote: > > Pradeep, > > Here are some related tickets that may also be helpful in understanding > the current behavior of these options. > > * https://issues.apache.org/jira/browse/CASSANDRA-5897 > * https://issues.apache.org/jira/browse/CASSANDRA-9474 > * https://issues.apache.org/jira/browse/CASSANDRA-10243 > * https://issues.apache.org/jira/browse/CASSANDRA-10242 > > On Sun, Aug 26, 2018 at 1:20 PM, Joshua Galbraith <jgalbra...@newrelic.com > > wrote: > >> Pradeep, >> >> That being said, I haven't experimented with -Dcassandra.ignore_dc=true >> -Dcassandra.ignore_rack=true before. >> >> The description here may be helpful: >> https://github.com/apache/cassandra/blob/trunk/NEWS.txt#L685-L693 >> >> I would spin up a small test cluster with data you don't care about and >> verify that your above assumptions are correct there first. >> >> On Sun, Aug 26, 2018 at 1:09 PM, Joshua Galbraith < >> jgalbra...@newrelic.com> wrote: >> >>> Pradeep. >>> >>> Right, so from that documentation is sounds like you actually have to >>> stop all nodes in the cluster at once and bring them back up one at a time. >>> A rolling restart won't work here. >>> >>> On Sun, Aug 26, 2018 at 11:46 AM, Pradeep Chhetri <prad...@stashaway.com >>> > wrote: >>> >>>> Hi Joshua, >>>> >>>> Thank you for the reply. Sorry i forgot to mention that I already went >>>> through that documentation. There are few missing things regarding which I >>>> have few questions: >>>> >>>> 1) One thing which isn't mentioned there is that cassandra fails to >>>> restart when we change the datacenter name *or* rack name of a node. >>>> So whether should i first rolling restart cassandra with flag >>>> "-Dcassandra.ignore_dc=true -Dcassandra.ignore_rack=true", then run >>>> sequential repair and then cleanup and then rolling restart cassandra >>>> without that flag. >>>> >>>> 2) Should i not allow any read/write operation from applications during >>>> the time when sequential repair is running. >>>> >>>> Regards, >>>> Pradeep >>>> >>>> On Mon, Aug 27, 2018 at 12:19 AM, Joshua Galbraith < >>>> jgalbra...@newrelic.com.invalid> wrote: >>>> >>>>> Pradeep, it sounds like what you're proposing counts as a topology >>>>> change because you are changing the datacenter name and rack name. >>>>> >>>>> Please refer to the documentation here about what to do in that >>>>> situation: >>>>> https://docs.datastax.com/en/cassandra/3.0/cassandra/operati >>>>> ons/opsSwitchSnitch.html >>>>> >>>>> In particular: >>>>> >>>>> Simply altering the snitch and replication to move some nodes to a new >>>>>> datacenter will result in data being replicated incorrectly. >>>>> >>>>> >>>>> Topology changes may occur when the replicas are placed in different >>>>>> places by the new snitch. Specifically, the replication strategy places >>>>>> the >>>>>> replicas based on the information provided by the new snitch. >>>>> >>>>> >>>>> If the topology of the network has changed, but no datacenters are >>>>>> added: >>>>>> a. Shut down all the nodes, then restart them. >>>>>> b. Run a sequential repair and nodetool cleanup on each node. >>>>> >>>>> >>>>> On Sun, Aug 26, 2018 at 11:14 AM, Pradeep Chhetri < >>>>> prad...@stashaway.com> wrote: >>>>> >>>>>> Hello everyone, >>>>>> >>>>>> Since i didn't hear from anyone, just want to describe my question >>>>>> again: >>>>>> >>>>>> Am i correct in understanding that i need to do following steps to >>>>>> migrate data from SimpleSnitch to GPFS changing datacenter name and rack >>>>>> name to AWS region and Availability zone respectively >>>>>> >>>>>> 1) Update the rack and datacenter fields in >>>>>> cassandra-rackdc.properties file and rolling restart cassandra with this >>>>>> flag "-Dcassandra.ignore_dc=true -Dcassandra.ignore_rack=true" >>>>>> >>>>>> 2) Run nodetool repair --sequential and nodetool cleanup. >>>>>> >>>>>> 3) Rolling restart cassandra removing the flag >>>>>> "-Dcassandra.ignore_dc=true -Dcassandra.ignore_rack=true" >>>>>> >>>>>> Regards, >>>>>> Pradeep >>>>>> >>>>>> On Thu, Aug 23, 2018 at 10:53 PM, Pradeep Chhetri < >>>>>> prad...@stashaway.com> wrote: >>>>>> >>>>>>> Hello, >>>>>>> >>>>>>> I am currently running a 3.11.2 cluster in SimpleSnitch hence the >>>>>>> datacenter is datacenter1 and rack is rack1 for all nodes on AWS. I >>>>>>> want to >>>>>>> switch to GPFS by changing the rack name to the availability-zone name >>>>>>> and >>>>>>> datacenter name to region name. >>>>>>> >>>>>>> When I try to restart individual nodes by changing those values, it >>>>>>> failed to start throwing the error about dc and rack name mismatch but >>>>>>> gives me an option to set ignore_dc and ignore_rack to true to bypass >>>>>>> it. >>>>>>> >>>>>>> I am not sure if it is safe to set those two flags to true and if >>>>>>> there is any drawback now or in future when i add a new datacenter to >>>>>>> the >>>>>>> cluster. I went through the documentation on Switching Snitches but >>>>>>> didn't >>>>>>> get much explanation. >>>>>>> >>>>>>> Regards, >>>>>>> Pradeep >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> -- >>>>> *Joshua Galbraith *| Lead Software Engineer | New Relic >>>>> >>>> >>>> >>> >>> >>> -- >>> *Joshua Galbraith *| Lead Software Engineer | New Relic >>> >> >> >> >> -- >> *Joshua Galbraith *| Lead Software Engineer | New Relic >> > > > > -- > *Joshua Galbraith *| Lead Software Engineer | New Relic > > >