*Mike*, for now I can't upgrade my cluster. I'm going to check the servers time sync. Thanks;
*Bryan*, so u think it's not a distributed deleted problem. Thanks for bringing it up. Btw, hector should not be hiding any exception from me. Although there's a mutator reuse in my application. I'm gonna check if it may be the problem too. Guys, one more doubt: Is there any limitation on how many columns should I delete per delete operation? I'm currently sending 100 deletions each time. On Fri, Feb 15, 2013 at 4:46 PM, Bryan Talbot <btal...@aeriagames.com>wrote: > With a RF and CL of one, there is no replication so there can be no issue > with distributed deletes. Writes (and reads) can only go to the one host > that has the data and will be refused if that node is down. I'd guess that > your app isn't deleting records when you think that it is, or that the > delete is failing but not being detected as failed. > > -Bryan > > > > On Fri, Feb 15, 2013 at 10:21 AM, Mike <mthero...@yahoo.com> wrote: > >> If you increase the number of nodes to 3, with an RF of 3, then you >> should be able to read/delete utilizing a quorum consistency level, which I >> believe will help here. Also, make sure the time of your servers are in >> sync, utilizing NTP, as drifting time between you client and server could >> cause updates to be mistakenly dropped for being old. >> >> Also, make sure you are running with a gc_grace period that is high >> enough. The default is 10 days. >> >> Hope this helps, >> -Mike >> >> >> On 2/15/2013 1:13 PM, Víctor Hugo Oliveira Molinar wrote: >> >>> hello everyone! >>> >>> I have a column family filled with event objects which need to be >>> processed by query threads. >>> Once each thread query for those objects(spread among columns bellow a >>> row), it performs a delete operation for each object in cassandra. >>> It's done in order to ensure that these events wont be processed again. >>> Some tests has showed me that it works, but sometimes i'm not getting >>> those events deleted. I checked it through cassandra-cli,etc. >>> >>> So, reading it >>> (http://wiki.apache.org/**cassandra/DistributedDeletes<http://wiki.apache.org/cassandra/DistributedDeletes>) >>> I came to a conclusion that I may be reading old data. >>> My cluster is currently configured as: 2 nodes, RF1, CL 1. >>> In that case, what should I do? >>> >>> - Increase the consistency level for the write operations( in that case, >>> the deletions ). In order to ensure that those deletions are stored in all >>> nodes. >>> or >>> - Increase the consistency level for the read operations. In order to >>> ensure that I'm reading only those yet processed events(deleted). >>> >>> ? >>> >>> - >>> Thanks in advance >>> >>> >>> >> >