On Sep 14, 2011, at 2:08 PM, Brandon Williams wrote:

> On Wed, Sep 14, 2011 at 8:54 AM, Ryan Hadley <r...@sgizmo.com> wrote:
>> Hi,
>> 
>> So, here's the backstory:
>> 
>> We were running Cassandra 0.7.4 and at one point in time had a node in the 
>> ring at 10.84.73.18. We removed this node from the ring successfully in 
>> 0.7.4. It stopped showing in the nodetool ring command. But occasionally 
>> we'd still get weird log entries about failing to write/read to IP 
>> 10.84.73.18.
>> 
>> We upgraded to Cassandra 0.8.4. Now, nodetool ring shows this old node:
>> 
>> 10.84.73.18     datacenter1 rack1       Down   Leaving ?               6.71% 
>>   32695837177645752437561450928649262701
>> 
>> So I started a nodetool removetoken on 
>> 32695837177645752437561450928649262701 last Friday. It's still going strong 
>> this morning, on day 5:
>> 
>> ./bin/nodetool -h 10.84.73.47 -p 8080 removetoken status
>> RemovalStatus: Removing token (32695837177645752437561450928649262701). 
>> Waiting for replication confirmation from 
>> [/10.84.73.49,/10.84.73.48,/10.84.73.51].
>> 
>> Should I just be patient? Or is something really weird with this node?
> 
> 5 days seems excessive unless there is a very large amount of data per
> node.  I would check nodetool netstats, and if the streams don't look
> active issue a 'removetoken force' against 10.84.73.47 and accept that
> you may possibly need to run repair to restore the replica count.
> 
> -Brandon

Hi Brandon,

Thanks for the reply. Quick question though:

1. We write all data to this ring with a TTL of 30 days
2. This node hasn't been in the ring for at least 90 days, more like 120 days 
since it's been in the ring.

So, if I nodetool removetoken forced it, would I still have to be concerned 
about running a repair?

Also, after this node is removed, I'm going to rebalance with nodetool move. 
Would that remove the repair requirement too?

Thanks-
Ryan

Reply via email to