Make sure the new cluster has a different cluster name, and avoid copying system.peers if you can avoid it. Doing so risks merging your new cluster and old cluster if they’re able to reach each other.
-- Jeff Jirsa > On Jan 11, 2018, at 1:41 AM, Pradeep Chhetri <prad...@stashaway.com> wrote: > > Thank you very much Jean. Since i don't have any constraints, as you said, i > will try copying the complete keyspace system node by node first and will do > nodetool refresh and see if it works. > > > >> On Thu, Jan 11, 2018 at 3:21 PM, Jean Carlo <jean.jeancar...@gmail.com> >> wrote: >> Hello, >> >> Basically, every node has to have the same token range. So yes you have to >> play with initial_token having the same numbers of tokens per node like the >> cluster source. To save time and if you dont have any constraints about the >> name of the cluster etc. you can just copy and paste the complete keyspace >> system node by node. >> >> So you will have the same cluster( cluster name, confs, etc) >> >> >> Saludos >> >> Jean Carlo >> >> "The best way to predict the future is to invent it" Alan Kay >> >>> On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri <prad...@stashaway.com> >>> wrote: >>> Hello Jean, >>> >>> I am running cassandra 3.11.1. >>> >>> Since i dont have much cassandra operations experience yet, I have a >>> follow-up question - how can i ensure the same token ranges distribution ? >>> Do i need to set initial_token configuration for each cassandra node ? >>> >>> Thank you for the quick response. >>> >>> >>> >>> >>> >>>> On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo <jean.jeancar...@gmail.com> >>>> wrote: >>>> Hello Pradeep, >>>> >>>> Actually the key here is to know if your cluster has the same token ranges >>>> distribution. So it is not only the same size but also the same tokens >>>> match node by node, from cluster source to cluster destination. In that >>>> case, you can use nodetool refresh.So after copy all your sstable node by >>>> node, it would be enough to make nodetool refresh in every node to restore >>>> your data. You can also restart casandra instead of doing nodetool >>>> refresh. It will help you to avoid the compactions after refreshing. >>>> >>>> >>>> Saludos >>>> >>>> Jean Carlo >>>> >>>> "The best way to predict the future is to invent it" Alan Kay >>>> >>>>> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri <prad...@stashaway.com> >>>>> wrote: >>>>> Hello everyone, >>>>> >>>>> We are running cassandra cluster inside containers over Kubernetes. We >>>>> have a requirement where we need to restore a fresh new cluster with >>>>> existing snapshot on weekly basis. >>>>> >>>>> Currently, while doing it manually. i need to copy the snapshot folder >>>>> inside container and then run sstableloader utility to load those tables. >>>>> >>>>> Since the source and destination cluster size is equal, I was thinking if >>>>> there are some easy way to just copy and paste the complete data >>>>> directory by mapping the nodes one to one. >>>>> >>>>> Since i wasn't able to find documentation around other backup >>>>> restoration methods apart from nodetool snapshot and sstableloader, I >>>>> haven't explored much. I recently came across this project - >>>>> https://github.com/Netflix/Priam but tried it yet. >>>>> >>>>> Would be very happy if i can get some ideas around various ways of >>>>> backup/restoration while running inside containers. >>>>> >>>>> Thank you >>>> >>> >> >