mmm, has repair completed on all nodes ? 

> Also, while it was digging around, I noticed that we do a LOT of reads 
> immediately after writes, and almost every read from the first DC was 
> bringing a read-repair along with it. 
What CL are you using ? 

Cheers

-----------------
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 20/03/2012, at 7:39 AM, Caleb Rackliffe wrote:

> Hey Aaron,
> 
> I've run cleanup jobs across all 15 nodes, and after that, I still have about 
> a 24 million to 15 million key ratio between the data centers.  The first DC 
> is a few months older than the second, and it also began its life before 
> 1.0.7 was out, whereas the second started at 1.0.7.  I wonder if running and 
> upgradesstables would be interesting?
> 
> Also, while it was digging around, I noticed that we do a LOT of reads 
> immediately after writes, and almost every read from the first DC was 
> bringing a read-repair along with it.  (Possibly because the distant DC had 
> not yet received certain mutations?)  I ended up turning RR off entirely, 
> since I've got HH in place to handle short-duration failures :)
> 
> Caleb Rackliffe | Software Developer  
> M 949.981.0159 | ca...@steelhouse.com
> <EB2FF764-478C-4966-9B0A-E7B76D6AD7DC[21].png>
> 
> From: aaron morton <aa...@thelastpickle.com>
> Reply-To: "user@cassandra.apache.org" <user@cassandra.apache.org>
> Date: Mon, 19 Mar 2012 13:34:38 -0400
> To: "user@cassandra.apache.org" <user@cassandra.apache.org>
> Subject: Re: Token Ring Gaps in a 2 DC Setup
> 
>>  I've also run repair on a few nodes in both data centers, but the sizes are 
>> still vastly different.
> If repair is completing on all the nodes then the data is fully distributed. 
> 
> If you want to dig around…
> 
> Take a look at the data files on disk. Do the nodes in DC 1 have some larger, 
> older, data files ? These may be waiting for compaction to catch up them. 
> 
> If you have done any toke moves, did you run cleanup afterwards ? 
> 
> 
> Cheers
> 
> -----------------
> Aaron Morton
> Freelance Developer
> @aaronmorton
> http://www.thelastpickle.com
> 
> On 18/03/2012, at 8:35 PM, Caleb Rackliffe wrote:
> 
>> More detail…
>> 
>> I'm running 1.0.7 on these boxes, and the keyspace readout from the CLI 
>> looks like this:
>> 
>> create keyspace Users
>>   with placement_strategy = 'NetworkTopologyStrategy'
>>   and strategy_options = {DC2 : 1, DC1 : 2}
>>   and durable_writes = true;
>> 
>> Thanks!
>> 
>> Caleb Rackliffe | Software Developer 
>> M 949.981.0159 | ca...@steelhouse.com
>> 
>> From: Caleb Rackliffe <ca...@steelhouse.com>
>> Date: Sun, 18 Mar 2012 02:47:05 -0400
>> To: "user@cassandra.apache.org" <user@cassandra.apache.org>
>> Subject: Token Ring Gaps in a 2 DC Setup
>> 
>> Hi Everyone,
>> 
>> I have a cluster using NetworkTopologyStrategy that looks like this:
>> 
>> 10.41.116.22     DC1         RAC1         Up     Normal  13.21 GB        
>> 10.00%  0                                           
>> 10.54.149.202   DC2         RAC1         Up     Normal  6.98 GB            
>> 0.00%   1                                           
>> 10.41.116.20     DC1         RAC2         Up     Normal  12.75 GB        
>> 10.00%  17014118300000000000000000000000000000      
>> 10.41.116.16     DC1         RAC3         Up     Normal  12.62 GB        
>> 10.00%  34028236700000000000000000000000000000      
>> 10.54.149.203   DC2         RAC1         Up     Normal  6.7 GB              
>> 0.00%   34028236700000000000000000000000000001      
>> 10.41.116.18     DC1         RAC4         Up     Normal  10.8 GB          
>> 10.00%  51042355000000000000000000000000000000      
>> 10.41.116.14     DC1         RAC5         Up     Normal  10.27 GB        
>> 10.00%  68056473400000000000000000000000000000      
>> 10.54.149.204   DC2         RAC1         Up     Normal  6.7 GB             
>> 0.00%   68056473400000000000000000000000000001      
>> 10.41.116.12     DC1         RAC6         Up     Normal  10.58 GB        
>> 10.00%  85070591700000000000000000000000000000      
>> 10.41.116.10     DC1         RAC7         Up     Normal  10.89 GB        
>> 10.00%  102084710000000000000000000000000000000     
>> 10.54.149.205   DC2         RAC1         Up     Normal  7.51 GB           
>> 0.00%   102084710000000000000000000000000000001     
>> 10.41.116.8       DC1         RAC8          Up     Normal  10.48 GB        
>> 10.00%  119098828000000000000000000000000000000     
>> 10.41.116.24     DC1         RAC9         Up     Normal  10.89 GB        
>> 10.00%  136112947000000000000000000000000000000     
>> 10.54.149.206   DC2         RAC1         Up     Normal  6.37 GB           
>> 0.00%   136112947000000000000000000000000000001     
>> 10.41.116.26     DC1         RAC10       Up     Normal  11.17 GB        
>> 10.00%  153127065000000000000000000000000000000
>> 
>> There are two data centers, one with 10 nodes/2 replicas and one with 5 
>> nodes/1 replica.  What I've attempted to do with my token assignments is 
>> have each node in the smaller DC handle 20% of the keyspace, and this would 
>> mean that I should see roughly equal usage on all 15 boxes.  It just doesn't 
>> seem to be happening that way, though.  It looks like the "1 replica" nodes 
>> are carrying about half the data the "2 replica" nodes are.  It's almost as 
>> if those nodes are only handling 10% of the keyspace instead of 20%.
>> 
>> Does anybody have any suggestions as to what might be going on?  I've run 
>> nodetool getendpoints against a bunch of keys, and I always get back three 
>> nodes, so I'm pretty confused.  I've also run repair on a few nodes in both 
>> data centers, but the sizes are still vastly different.
>> 
>> Thanks!
>> 
>> Caleb Rackliffe | Software Developer 
>> M 949.981.0159 | ca...@steelhouse.com
> 

Reply via email to