It depends on what you really use which CL for your operations.
Your RF is 2, so if you read/write with CL=ALL, your r/w will be
always consistent. If your read is CL=ONE, you have chance to read old
data anytime, decommission is not matter. CL=QUORUM on RF=2 is
semantically identical with CL=ALL.

maki

2011/5/13 Ryan Hadley <r...@sgizmo.com>:
> Hi,
>
> I'm running Cassandra (0.7.4) on a 4 node ring.  It was a 3 node ring, but we 
> ended up expanding it to 4... So then I followed the many suggestions to 
> rebalance the ring.  I found a script that suggested I use:
>
> # ~/nodes_calc.py
> How many nodes are in your cluster? 4
> node 0: 0
> node 1: 42535295865117307932921825928971026432
> node 2: 85070591730234615865843651857942052864
> node 3: 127605887595351923798765477786913079296
>
> So I started to migrate each node to those tokens.
>
> I have my replication factor set to 2, so I guess I was expecting to be able 
> to continue to use this ring without issues.  But it seems that the node 
> still accepts writes while it's decommissioning?  I say this because if I 
> interrupt the decommission by stopping Cassandra and starting it again, it 
> appears to run through several commit logs.  And as soon as it's through with 
> those commit logs, I no longer get consistency issues.
>
> The issue I'm seeing is that writes to this ring will succeed, but it's 
> possible for a subsequent read to return an older object.  For several 
> minutes even.
>
> I'm not sure if I did something wrong... learning as I go here and this list 
> archive has been very useful.  But, is there anyway I can rebalance the node 
> and get better consistency?
>
> Thanks,
> Ryan

Reply via email to