> All our writes/deletes are done with CL.QUORUM.
> Our reads are done with CL.ONE. Although the reads that confirmed the old 
> data were done with CL.QUORUM.
mmmm

> According to 
> https://svn.apache.org/viewvc/cassandra/branches/cassandra-0.6/CHANGES.txt 
> 0.6.6 has the same patch
> for (CASSANDRA-1074) as 0.7 and so I assumed that minor compactions in 0.6.6 
> and up also purged tombstones.
My bad. As you were. 

After the repair did the un-deleted data remain un-deleted ? Are you back to a 
stable situation ? 

Without a lot more detail I am at a bit of a loss. 

I know it's painful but migrating to 1.0 *really* will make your life so much 
easier and faster. At some point you may hit a bug or a problem in 0.6 and the 
solution may be to upgrade, quickly.

Cheers

-----------------
Aaron Morton
Freelance Developer
@aaronmorton
http://www.thelastpickle.com

On 6/03/2012, at 11:13 PM, Stefan Reek wrote:

> Hi Aaron,
> 
> Thanks for the quick reply.
> All our writes/deletes are done with CL.QUORUM.
> Our reads are done with CL.ONE. Although the reads that confirmed the old 
> data were done with CL.QUORUM.
> According to 
> https://svn.apache.org/viewvc/cassandra/branches/cassandra-0.6/CHANGES.txt 
> 0.6.6 has the same patch
> for (CASSANDRA-1074) as 0.7 and so I assumed that minor compactions in 0.6.6 
> and up also purged tombstones.
> The only suspicious thing I noticed was that after adding the fourth node 
> repairs became extremely slow and heavy.
> Running it degraded the performance of the whole cluster and the new node 
> even went OOM when running it.
> 
> Cheers,
> 
> Stefan
> 
> On 03/06/2012 10:51 AM, aaron morton wrote:
>> 
>>> After we added a fourth node, keeping RF=3, some old data appeared in the 
>>> database.
>> What CL are you working at ? (Should not matter too much with repair 
>> working, just asking)
>> 
>> 
>>> We don't run compact on the nodes explicitly as I understand that running 
>>> repair will trigger a
>>> major compaction. I'm not entirely sure if it does so, but in any case the 
>>> tombstones will be removed by a minor
>>> compaction.
>> In 0.6.x tombstones were only purged during a major / manual compaction. 
>> Purging during minor compaction came in during 0.7
>> https://github.com/apache/cassandra/blob/trunk/CHANGES.txt#L1467
>> 
>>> Can anyone think of any reason why the old data reappeared?
>> It sounds like you are doing things correctly. The complicating factor is 
>> 0.6 is so very old. 
>> 
>> 
>> If I wanted to poke around some more I would conduct reads as CL one against 
>> nodes and see if they return the "deleted" data or not. This would help me 
>> understand if the tombstone is still out there. 
>> 
>> I would also poke around a lot in the logs to make sure repair was running 
>> as expected and completing. If you find anything suspicious post examples. 
>> 
>> Finally I would ensure CL QUROUM was been used. 
>> 
>> Hope that helps.
>> 
>> 
>> -----------------
>> Aaron Morton
>> Freelance Developer
>> @aaronmorton
>> http://www.thelastpickle.com
>> 
>> On 6/03/2012, at 10:13 PM, Stefan Reek wrote:
>> 
>>> Hi,
>>> 
>>> We were running a 3-node cluster of cassandra 0.6.13 with RF=3.
>>> After we added a fourth node, keeping RF=3, some old data appeared in the 
>>> database.
>>> As far as I understand this can only happen if nodetool repair wasn't run 
>>> for more than GCGraceSeconds.
>>> Our GCGraceSeconds is set to the default of 10 days (864000 seconds).
>>> We have  a scheduled cronjob to run repair once each week on every node, 
>>> each on another day.
>>> I'm sure that none of the nodes ever skipped running a repair.
>>> We don't run compact on the nodes explicitly as I understand that running 
>>> repair will trigger a
>>> major compaction. I'm not entirely sure if it does so, but in any case the 
>>> tombstones will be removed by a minor
>>> compaction. So I expected that the reappearing data, which is a couple of 
>>> months old in some cases, was long gone
>>> by the time we added the node.
>>> 
>>> Can anyone think of any reason why the old data reappeared?
>>> 
>>> Stefan
>> 
> 

Reply via email to