I would check for stored hints in /var/lib/cassandra/data/system

Putting nodes in different racks can make placement tricky so…
Are you running a multi DC setup ? Are you using the NTS ? What is the RF 
setting ? What setting do you have for the Snitch ? What is the full node 
assignments. 

Cheers


-----------------
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 19/07/2012, at 6:00 PM, Mariusz Dymarek wrote:

> Hi again,
> we have now moved all nodes to correct position in ring, but we can see 
> higher load on 2 nodes, than on other nodes:
> ...
> node01-05   rack1  Up  Normal  244.65 GB  6,67% 
> 102084710076281539039012382229530463432
> node02-13   rack2  Up  Normal  240.26 GB  6,67% 
> 107756082858297180096735292353393266961
> node01-13   rack1  Up  Normal  243.75 GB  6,67% 
> 113427455640312821154458202477256070485
> node02-05   rack2  Up  Normal  249.31 GB  6,67% 
> 119098828422328462212181112601118874004
> node01-14   rack1  Up  Normal  244.95 GB  6,67% 
> 124770201204344103269904022724981677533
> node02-14   rack2  Up  Normal  392.7 GB   6,67% 
> 130441573986359744327626932848844481058
> node01-06   rack1  Up  Normal  249.3 GB   6,67% 
> 136112946768375385385349842972707284576
> node02-15   rack2  Up  Normal  286.82 GB  6,67% 
> 141784319550391026443072753096570088106
> node01-15   rack1  Up  Normal  245.21 GB  6,67% 
> 147455692332406667500795663220432891630
> node02-06   rack2  Up  Normal  244.9 GB   6,67% 
> 153127065114422308558518573344295695148
> ...
> 
> Node:
> * node01-15  = >  286.82 GB
> * node02-14  = >  392.7 GB
> 
> average load on all other nodes is around 245 GB, nodetool cleanup command 
> was invoked on problematic nodes after move operation...
> Why this has happen?
> And how can we balance cluster?
> On 06.07.2012 20:15, aaron morton wrote:
>> If you have the time yes I would wait for the bootstrap to finish. It
>> will make you life easier.
>> 
>> good luck.
>> 
>> 
>> -----------------
>> Aaron Morton
>> Freelance Developer
>> @aaronmorton
>> http://www.thelastpickle.com
>> 
>> On 6/07/2012, at 7:12 PM, Mariusz Dymarek wrote:
>> 
>>> Hi,
>>> we`re in the middle of extending our cluster from 10 to 30 nodes,
>>> we`re running cassandra 1.1.1...
>>> We`ve generated initial tokens for new nodes:
>>> "0": 0, # existing: node01-01
>>> "1": 5671372782015641057722910123862803524, # new: node02-07
>>> "2": 11342745564031282115445820247725607048, # new: node01-07
>>> "3": 17014118346046923173168730371588410572, # existing: node02-01
>>> "4": 22685491128062564230891640495451214097, # new: node01-08
>>> "5": 28356863910078205288614550619314017621, # new: node02-08
>>> "6": 34028236692093846346337460743176821145, # existing: node01-02
>>> "7": 39699609474109487404060370867039624669, # new: node02-09
>>> "8": 45370982256125128461783280990902428194, # new: node01-09
>>> "9": 51042355038140769519506191114765231718, # existing: node02-02
>>> "10": 56713727820156410577229101238628035242, # new: node01-10
>>> "11": 62385100602172051634952011362490838766, # new: node02-10
>>> "12": 68056473384187692692674921486353642291, # existing: node01-03
>>> "13": 73727846166203333750397831610216445815, # new: node02-11
>>> "14": 79399218948218974808120741734079249339, # new: node01-11
>>> "15": 85070591730234615865843651857942052864, # existing: node02-03
>>> "16": 90741964512250256923566561981804856388, # new: node01-12
>>> "17": 96413337294265897981289472105667659912, # new: node02-12
>>> "18": 102084710076281539039012382229530463436, # existing: node01-05
>>> "19": 107756082858297180096735292353393266961, # new: node02-13
>>> "20": 113427455640312821154458202477256070485, # new: node01-13
>>> "21": 119098828422328462212181112601118874009, # existing: node02-05
>>> "22": 124770201204344103269904022724981677533, # new: node01-14
>>> "23": 130441573986359744327626932848844481058, # new: node02-14
>>> "24": 136112946768375385385349842972707284582, # existing: node01-06
>>> "25": 141784319550391026443072753096570088106, # new: node02-15
>>> "26": 147455692332406667500795663220432891630, # new: node01-15
>>> "27": 153127065114422308558518573344295695155, # existing: node02-06
>>> "28": 158798437896437949616241483468158498679, # new: node01-16
>>> "29": 164469810678453590673964393592021302203 # new: node02-16
>>> then we`ve started to boostrap new nodes,
>>> but due to copy and paste mistake:
>>> * node node01-14 was started with
>>> 130441573986359744327626932848844481058 as initial token(so node01-14
>>> has initial_token, what should belong to node02-14), it
>>> should have 124770201204344103269904022724981677533 as initial_token
>>> * node node02-14 was started with
>>> 136112946768375385385349842972707284582 as initial token, so it has
>>> token from existing node01-06....
>>> 
>>> However we`ve used other program for generating previous
>>> initial_tokens and actual token of node01-06 in ring is
>>> 136112946768375385385349842972707284576.
>>> Summing up: we have currently this situation in ring:
>>> 
>>> node02-05 rack2 Up Normal 596.31 GB 6.67%
>>> 119098828422328462212181112601118874004
>>> node01-14 rack1 Up Joining 242.92 KB 0.00%
>>> 130441573986359744327626932848844481058
>>> node01-06 rack1 Up Normal 585.5 GB 13.33%
>>> 136112946768375385385349842972707284576
>>> node02-14 rack2 Up Joining 113.17 KB 0.00%
>>> 136112946768375385385349842972707284582
>>> node02-15 rack2 Up Joining 178.05 KB 0.00%
>>> 141784319550391026443072753096570088106
>>> node01-15 rack1 Up Joining 191.7 GB 0.00%
>>> 147455692332406667500795663220432891630
>>> node02-06 rack2 Up Normal 597.69 GB 20.00%
>>> 153127065114422308558518573344295695148
>>> 
>>> 
>>> We would like to get back to our original configuration.
>>> Is it safe to wait for finishing bootstraping of all new nodes and
>>> after that invoke:
>>> * nodetool -h node01-14 move 124770201204344103269904022724981677533
>>> * nodetool -h node02-14 move 130441573986359744327626932848844481058
>>> We should probably run nodetool cleanup on several nodes after that...
>>> Regards
>>> Dymarek Mariusz
>> 
> 

Reply via email to