Run the repair with -pr option on each node which will repair only the
parition range.

nodetool repair -pr
On Wed, Oct 31, 2018 at 7:04 PM Surbhi Gupta <surbhi.gupt...@gmail.com> wrote:
>
> Nodetool repair will take way more time than nodetool rebuild.
> How much data u have in your original data center?
> Repair should be run to make the data consistent in case of node down more 
> than hintedhandoff period and dropped mutations.
> But as a thumb rule ,generally we run repair using opscenter (if using 
> Datastax) most of the times.
>
> So in your case run “nodetool rebuild <original data enter>” on all the nodes 
> in new data center.
> For making the rebuild process fast, increase three parameters, compaction 
> throughput , stream throughput and interdcstream  throughput.
>
> Thanks
> Surbhi
> On Tue, Oct 30, 2018 at 11:29 PM Akshay Bhardwaj 
> <akshay.bhardwaj1...@gmail.com> wrote:
>>
>> Hi Jonathan,
>>
>> That makes sense. Thank you for the explanation.
>>
>> Another quick question, as the cluster is still operative and the data for 
>> the past 2 weeks (since updating replication factor) is present in both the 
>> data centres, should I run "nodetool rebuild" or "nodetool repair"?
>>
>> I read that nodetool rebuild is faster and is useful till the new data 
>> centre is empty and no partition keys are present. So when is the good time 
>> to use either of the commands and what impact can it have on the data centre 
>> operations?
>>
>> Thanks and Regards
>>
>> Akshay Bhardwaj
>> +91-97111-33849
>>
>>
>> On Wed, Oct 31, 2018 at 2:34 AM Jonathan Haddad <j...@jonhaddad.com> wrote:
>>>
>>> You need to run "nodetool rebuild -- <existing-dc-name>" on each node in 
>>> the new DC to get the old data to replicate.  It doesn't do it 
>>> automatically because Cassandra has no way of knowing if you're done adding 
>>> nodes and if it were to migrate automatically, it could cause a lot of 
>>> problems. Imagine streaming 100 nodes data to 3 nodes in the new DC, not 
>>> fun.
>>>
>>> On Tue, Oct 30, 2018 at 1:59 PM Akshay Bhardwaj 
>>> <akshay.bhardwaj1...@gmail.com> wrote:
>>>>
>>>> Hi Experts,
>>>>
>>>> I previously had 1 Cassandra data centre in AWS Singapore region with 5 
>>>> nodes, with my keyspace's replication factor as 3 in Network topology.
>>>>
>>>> After this cluster has been running smoothly for 4 months (500 GB of data 
>>>> on each node's disk), I added 2nd data centre in AWS Mumbai region with 
>>>> yet again 5 nodes in Network topology.
>>>>
>>>> After updating my keyspace's replication factor to 
>>>> {"AWS_Sgp":3,"AWS_Mum":3}, my expectation was that the data present in Sgp 
>>>> region will immediately start replicating on the Mum region's nodes. 
>>>> However even after 2 weeks I do not see historical data to be replicated, 
>>>> but new data being written on Sgp region is present in Mum region as well.
>>>>
>>>> Any help or suggestions to debug this issue will be highly appreciated.
>>>>
>>>> Regards
>>>> Akshay Bhardwaj
>>>> +91-97111-33849
>>>>
>>>>
>>>
>>>
>>> --
>>> Jon Haddad
>>> http://www.rustyrazorblade.com
>>> twitter: rustyrazorblade
>>>
>>>
>>
>>


-- 
Best Regards,
Kiran.M.K.

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org

Reply via email to