Update: I ran a test whereby I deleted ALL the rows in a column family, using a consistency level of ALL. To do this, I mapped the ColumnFamily and called remove on each row id. There were 1.5 million rows, so 1.5 million rows were deleted.
I ran a counter job immediately after. This job maps the same column family and tests if any data is returned. If not, it considers the row a "tombstone". If yes, it considers the row not deleted. Below are the hadoop counters for those jobs. Note the fluctuation in the number of rows with data over time, and the increase in time to map the column family after the destroy job. No other clients were accessing cassandra during this time. I'm thoroughly confused. Count: started 13:02:30 EDT, finished 13:11:33 EDT (9 minutes 2 seconds): ROWS: 1,542,479 TOMBSTONES: 69 Destroy: started 16:48:45 EDT, finished 17:07:36 EDT (18 minutes 50 seconds) DESTROYED: 1,542,548 Count: started 17:15:42 EDT, finished 17:31:03 EDT (15 minutes 21 seconds) ROWS 876,464 TOMBSTONES 666,084 Count: started 17:31:32, finished 17:47:16 (15mins, 44 seconds) ROWS 1,451,665 TOMBSTONES 90,883 Count: started 17:52:34, finished 18:10:28 (17mins, 53 seconds) ROWS 1,425,644 TOMBSTONES 116,904 On Tue, Apr 27, 2010 at 5:37 PM, Joost Ouwerkerk <jo...@openplaces.org> wrote: > Clocks are in sync: > > cluster04:~/cassandra$ dsh -g development "date" > Tue Apr 27 17:36:33 EDT 2010 > Tue Apr 27 17:36:33 EDT 2010 > Tue Apr 27 17:36:33 EDT 2010 > Tue Apr 27 17:36:33 EDT 2010 > Tue Apr 27 17:36:34 EDT 2010 > Tue Apr 27 17:36:34 EDT 2010 > Tue Apr 27 17:36:34 EDT 2010 > Tue Apr 27 17:36:34 EDT 2010 > Tue Apr 27 17:36:34 EDT 2010 > Tue Apr 27 17:36:35 EDT 2010 > Tue Apr 27 17:36:35 EDT 2010 > Tue Apr 27 17:36:35 EDT 2010 > > On Tue, Apr 27, 2010 at 5:35 PM, Nathan McCall <n...@vervewireless.com> wrote: >> Have you confirmed that your clocks are all synced in the cluster? >> This may be the result of an unintentional read-repair occurring if >> that were the case. >> >> -Nate >> >> On Tue, Apr 27, 2010 at 2:20 PM, Joost Ouwerkerk <jo...@openplaces.org> >> wrote: >>> Hmm... Even after deleting with cl.ALL, I'm getting data back for some >>> rows after having deleted them. Which rows return data is >>> inconsistent from one run of the job to the next. >>> >>> On Tue, Apr 27, 2010 at 1:44 PM, Joost Ouwerkerk <jo...@openplaces.org> >>> wrote: >>>> To check that rows are gone, I check that KeySlice.columns is empty. And >>>> as >>>> I mentioned, immediately after the delete job, this returns the expected >>>> number. >>>> Unfortunately I reproduced with QUORUM this morning. No node outages. I >>>> am >>>> going to try ALL to see if that changes anything, but I am starting to >>>> wonder if I'm doing something else wrong. >>>> On Mon, Apr 26, 2010 at 9:45 PM, Jonathan Ellis <jbel...@gmail.com> wrote: >>>>> >>>>> How are you checking that the rows are gone? >>>>> >>>>> Are you experiencing node outages during this? >>>>> >>>>> DC_QUORUM is unfinished code right now, you should avoid using it. >>>>> Can you reproduce with normal QUORUM? >>>>> >>>>> On Sat, Apr 24, 2010 at 12:23 PM, Joost Ouwerkerk <jo...@openplaces.org> >>>>> wrote: >>>>> > I'm having trouble deleting rows in Cassandra. After running a job that >>>>> > deletes hundreds of rows, I run another job that verifies that the rows >>>>> > are >>>>> > gone. Both jobs run correctly. However, when I run the verification >>>>> > job an >>>>> > hour later, the rows have re-appeared. This is not a case of "ghosting" >>>>> > because the verification job actually checks that there is data in the >>>>> > columns. >>>>> > >>>>> > I am running a cluster with 12 nodes and a replication factor of 3. I >>>>> > am >>>>> > using DC_QUORUM consistency when deleting. >>>>> > >>>>> > Any ideas? >>>>> > Joost. >>>>> > >>>>> >>>>> >>>>> >>>>> -- >>>>> Jonathan Ellis >>>>> Project Chair, Apache Cassandra >>>>> co-founder of Riptano, the source for professional Cassandra support >>>>> http://riptano.com >>>> >>>> >>> >> >