Hello Jeff,

Thank you for the reply.

One doubt - If i copy the /var/lib/cassandra one to one from source cluster
to destination cluster nodes and change the cluster name in configuration
and delete system.peers table and restart each cassandra node, do you think
the cluster will come up properly. Although the clusters are in different
VPC in AWS but still i want to make sure that they get different
clustername.

Regards.

On Thu, Jan 11, 2018 at 9:37 PM, Jeff Jirsa <jji...@gmail.com> wrote:

> Make sure the new cluster has a different cluster name, and avoid copying
> system.peers if you can avoid it. Doing so risks merging your new cluster
> and old cluster if they’re able to reach each other.
>
> --
> Jeff Jirsa
>
>
> On Jan 11, 2018, at 1:41 AM, Pradeep Chhetri <prad...@stashaway.com>
> wrote:
>
> Thank you very much Jean. Since i don't have any constraints, as you said,
> i will try copying the complete keyspace system node by node first and will
> do nodetool refresh and see if it works.
>
>
>
> On Thu, Jan 11, 2018 at 3:21 PM, Jean Carlo <jean.jeancar...@gmail.com>
> wrote:
>
>> Hello,
>>
>> Basically, every node has to have the same token range. So yes you have
>> to play with initial_token having the same numbers of tokens per node like
>> the cluster source. To save time and if you dont have any constraints about
>> the name of the cluster etc. you can just copy and paste the complete
>> keyspace system node by node.
>>
>> So you will have the same cluster( cluster name, confs, etc)
>>
>>
>> Saludos
>>
>> Jean Carlo
>>
>> "The best way to predict the future is to invent it" Alan Kay
>>
>> On Thu, Jan 11, 2018 at 10:28 AM, Pradeep Chhetri <prad...@stashaway.com>
>> wrote:
>>
>>> Hello Jean,
>>>
>>> I am running cassandra 3.11.1.
>>>
>>> Since i dont have much cassandra operations experience yet, I have a
>>> follow-up question - how can i ensure the same token ranges distribution ?
>>> Do i need to set initial_token configuration for each cassandra node ?
>>>
>>> Thank you for the quick response.
>>>
>>>
>>>
>>>
>>>
>>> On Thu, Jan 11, 2018 at 3:04 PM, Jean Carlo <jean.jeancar...@gmail.com>
>>> wrote:
>>>
>>>> Hello Pradeep,
>>>>
>>>> Actually the key here is to know if your cluster has the same token
>>>> ranges distribution. So it is not only the same size but also the same
>>>> tokens match node by node, from cluster source to cluster destination. In
>>>> that case, you can use nodetool refresh.So after copy all your sstable node
>>>> by node, it would be enough to make nodetool refresh in every node to
>>>> restore your data. You can also restart casandra instead of doing nodetool
>>>> refresh. It will help you to avoid the compactions after refreshing.
>>>>
>>>>
>>>> Saludos
>>>>
>>>> Jean Carlo
>>>>
>>>> "The best way to predict the future is to invent it" Alan Kay
>>>>
>>>> On Thu, Jan 11, 2018 at 9:58 AM, Pradeep Chhetri <prad...@stashaway.com
>>>> > wrote:
>>>>
>>>>> Hello everyone,
>>>>>
>>>>> We are running cassandra cluster inside containers over Kubernetes. We
>>>>> have a requirement where we need to restore a fresh new cluster with
>>>>> existing snapshot on weekly basis.
>>>>>
>>>>> Currently, while doing it manually. i need to copy the snapshot folder
>>>>> inside container and then run sstableloader utility to load those tables.
>>>>>
>>>>> Since the source and destination cluster size is equal, I was thinking
>>>>> if there are some easy way to just copy and paste the complete data
>>>>> directory by mapping the nodes one to one.
>>>>>
>>>>> Since i wasn't able to find documentation around other  backup
>>>>> restoration methods apart from nodetool snapshot and sstableloader, I
>>>>> haven't explored much. I recently came across this project -
>>>>> https://github.com/Netflix/Priam but tried it yet.
>>>>>
>>>>> Would be very happy if i can get some ideas around various ways of
>>>>> backup/restoration while running inside containers.
>>>>>
>>>>> Thank you
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to