For Data Stax Enterprise specific questions try the support forums 
http://www.datastax.com/support-forums/

Cheers

-----------------
Aaron Morton
Freelance Cassandra Developer
New Zealand

@aaronmorton
http://www.thelastpickle.com

On 31/01/2013, at 8:27 AM, S C <as...@outlook.com> wrote:

> I am using DseDelegateSnitch
> 
> Thanks,
> SC
> From: aa...@thelastpickle.com
> Subject: Re: cluster issues
> Date: Tue, 29 Jan 2013 20:15:45 +1300
> To: user@cassandra.apache.org
> 
>       • We can always be proactive in keeping the time sync. But, Is there 
> any way to recover from a time drift (in a reactive manner)? Since it was a 
> lab environment, I dropped the KS (deleted data directory)
> There is a way to remove future dated columns, but it not for the faint 
> hearted. 
> 
> Basically:
> 1) Drop the gc_grace_seconds to 0
> 2) Delete the column with a timestamp way in the future, so it is guaranteed 
> to be higher than the value you want to delete. 
> 3) Flush the CF
> 4) Compact all the SSTables that contain the row. The easiest way to do that 
> is a major compaction, but we normally advise not to do that because it 
> creates one big file. You can also do a user defined compaction. 
> 
>       • Are there any other scenarios that would lead a cluster look like 
> below? Note:Actual topology of the cluster - ONE Cassandra node and TWO 
> Analytic nodes.
>       •
> What snitch are you using?
> If you have the property file snitch do all nodes have the same configuration 
> ?
> 
> There is a lot of sickness there. If possible I would scrub and start again. 
> 
> Cheers
>  
> -----------------
> Aaron Morton
> Freelance Cassandra Developer
> New Zealand
> 
> @aaronmorton
> http://www.thelastpickle.com
> 
> On 29/01/2013, at 6:29 AM, S C <as...@outlook.com> wrote:
> 
> One of our node in a 3 node cluster drifted by ~ 20-25 seconds. While I 
> figured this pretty quickly, I had few questions that am looking for some 
> answers.
> 
>       • We can always be proactive in keeping the time sync. But, Is there 
> any way to recover from a time drift (in a reactive manner)? Since it was a 
> lab environment, I dropped the KS (deleted data directory).
>       • Are there any other scenarios that would lead a cluster look like 
> below?Note:Actual topology of the cluster - ONE Cassandra node and TWO 
> Analytic nodes.
> 
> 
> On 192.168.2.100
> Address         DC          Rack        Status State   Load            Owns   
>              Token                                       
>                                                                               
>              113427455640312821154458202477256070485     
> 192.168.2.100  Cassandra   rack1       Up     Normal  601.34 MB       33.33%  
>             0                                           
> 192.168.2.101  Analytics   rack1       Down   Normal  149.75 MB       33.33%  
>             56713727820156410577229101238628035242      
> 192.168.2.102  Analytics   rack1       Down   Normal  ?               33.33%  
>             113427455640312821154458202477256070485   
> 
> On 192.168.2.101
> Address         DC          Rack        Status State   Load            Owns   
>              Token                                       
>                                                                               
>              113427455640312821154458202477256070485     
> 192.168.2.100  Analytics   rack1       Down   Normal  ?               33.33%  
>             0                                          
> 192.168.2.101  Analytics   rack1       Up     Normal  158.59 MB       33.33%  
>             56713727820156410577229101238628035242      
> 192.168.2.102  Analytics   rack1       Down   Normal  ?               33.33%  
>             113427455640312821154458202477256070485    
> 
> On 192.168.2.102
> Address         DC          Rack        Status State   Load            Owns   
>              Token                                       
>                                                                               
>              113427455640312821154458202477256070485     
> 192.168.2.100  Analytics   rack1       Down   Normal  ?               33.33%  
>             0                                          
> 192.168.2.101  Analytics   rack1       Down   Normal  ?               33.33%  
>             56713727820156410577229101238628035242      
> 192.168.2.102  Analytics   rack1       Up     Normal  117.02 MB       33.33%  
>             113427455640312821154458202477256070485     
> 
> 
> Appreciate your valuable inputs.
> 
> Thanks,
> SC

Reply via email to