possibilities include

1) you're using something other than rackunwarepartitioner, which is
the only one that behaves the way you describe
2) you've moved nodes around w/o running cleanup afterwards

On Sun, Aug 22, 2010 at 10:09 PM, Zhong Li <z...@voxeo.com> wrote:
> Today, I checked all nodes data and logs, there are very few nodes reported
> connections up/down. I found some data on each nodes which I don't
> understand.
>
> The ReplicationFactor is 2, write Consistency Level is one.  Example, the
> ring like Node1(Token1)->Node2(Token2)->Node3(Token3)->....... Node1 has
> token1, suppose all data with key as token1 should be  on Node1 and Node2,
> but why I can find some Node1/Node2 data on Node3 also? I dumped the data on
> Node3 to my local, red them and found some Node1/Node2's data on the Node3
>  and those data should be deleted.
>
> Why Node3 has Node1/Node2's data?
>
> Thanks.
>
>
> On Aug 18, 2010, at 10:44 AM, Jonathan Ellis wrote:
>
>> HH would handle it if it were a FD false positive, but if a node
>> actually does go down then it can miss writes before HH kicks in.
>>
>> On Wed, Aug 18, 2010 at 9:30 AM, Raj N <raj.cassan...@gmail.com> wrote:
>>>
>>> Guys,
>>>    Correct me if I am wrong. The whole problem is because a node missed
>>> an
>>> update when it was down. Shouldn’t HintedHandoff take care of this case?
>>>
>>> Thanks
>>> -Raj
>>>
>>> -----Original Message-----
>>> From: Jonathan Ellis [mailto:jbel...@gmail.com]
>>> Sent: Wednesday, August 18, 2010 9:22 AM
>>> To: user@cassandra.apache.org
>>> Subject: Re: data deleted came back after 9 days.
>>>
>>> Actually, tombstones are read repaired too -- as long as they are not
>>> expired.  But nodetool repair is much less error-prone than relying on RR
>>> and your memory of what deletes you issued.
>>>
>>> Either way, you'd need to increase GCGraceSeconds first to make the
>>> tombstones un-expired first.
>>>
>>> On Wed, Aug 18, 2010 at 12:43 AM, Benjamin Black <b...@b3k.us> wrote:
>>>>
>>>> On Tue, Aug 17, 2010 at 7:49 PM, Zhong Li <z...@voxeo.com> wrote:
>>>>>
>>>>> Those data were inserted one node, then deleted on a remote node in
>>>>> less than 2 seconds. So it is very possible some node lost tombstone
>>>>> when connection lost.
>>>>> My question, is a ConstencyLevel.ALL read can retrieve lost tombstone
>>>>> back instead of repair?
>>>>>
>>>>
>>>> No.  Read repair does not replay operations.  You must run nodetool
>>>> repair.
>>>>
>>>>
>>>> b
>>>>
>>>
>>>
>>>
>>> --
>>> Jonathan Ellis
>>> Project Chair, Apache Cassandra
>>> co-founder of Riptano, the source for professional Cassandra support
>>> http://riptano.com
>>>
>>
>>
>>
>> --
>> Jonathan Ellis
>> Project Chair, Apache Cassandra
>> co-founder of Riptano, the source for professional Cassandra support
>> http://riptano.com
>
>



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of Riptano, the source for professional Cassandra support
http://riptano.com

Reply via email to