Good! :)

Can you reproduce w/o map/reduce, with raw get_range_slices?

On Wed, Apr 28, 2010 at 3:56 PM, Joost Ouwerkerk <jo...@openplaces.org> wrote:
> Yes! Reproduced on single-node cluster:
>
> 10/04/28 16:30:24 INFO mapred.JobClient:     ROWS=274884
> 10/04/28 16:30:24 INFO mapred.JobClient:     TOMBSTONES=951083
>
> 10/04/28 16:42:49 INFO mapred.JobClient:     ROWS=166580
> 10/04/28 16:42:49 INFO mapred.JobClient:     TOMBSTONES=1059387
>
> On Wed, Apr 28, 2010 at 10:43 AM, Jonathan Ellis <jbel...@gmail.com> wrote:
>> It sounds like either there is a fairly obvious bug, or you're doing
>> something wrong. :)
>>
>> Can you reproduce against a single node?
>>
>> On Tue, Apr 27, 2010 at 5:14 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>> wrote:
>>> Update: I ran a test whereby I deleted ALL the rows in a column
>>> family, using a consistency level of ALL.  To do this, I mapped the
>>> ColumnFamily and called remove on each row id.  There were 1.5 million
>>> rows, so 1.5 million rows were deleted.
>>>
>>> I ran a counter job immediately after.  This job maps the same column
>>> family and tests if any data is returned.  If not, it considers the
>>> row a "tombstone".  If yes, it considers the row not deleted.  Below
>>> are the hadoop counters for those jobs.  Note the fluctuation in the
>>> number of rows with data over time, and the increase in time to map
>>> the column family after the destroy job.  No other clients were
>>> accessing cassandra during this time.
>>>
>>> I'm thoroughly confused.
>>>
>>> Count: started 13:02:30 EDT, finished 13:11:33 EDT (9 minutes 2 seconds):
>>>   ROWS:        1,542,479
>>>   TOMBSTONES:  69
>>>
>>> Destroy: started 16:48:45 EDT, finished 17:07:36 EDT (18 minutes 50 seconds)
>>>   DESTROYED:  1,542,548
>>>
>>> Count: started 17:15:42 EDT, finished 17:31:03 EDT (15 minutes 21 seconds)
>>>   ROWS 876,464
>>>   TOMBSTONES   666,084
>>>
>>> Count: started 17:31:32, finished 17:47:16 (15mins, 44 seconds)
>>>   ROWS 1,451,665
>>>   TOMBSTONES   90,883
>>>
>>> Count: started 17:52:34, finished 18:10:28 (17mins, 53 seconds)
>>>   ROWS 1,425,644
>>>   TOMBSTONES   116,904
>>>
>>> On Tue, Apr 27, 2010 at 5:37 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>>> wrote:
>>>> Clocks are in sync:
>>>>
>>>> cluster04:~/cassandra$ dsh -g development "date"
>>>> Tue Apr 27 17:36:33 EDT 2010
>>>> Tue Apr 27 17:36:33 EDT 2010
>>>> Tue Apr 27 17:36:33 EDT 2010
>>>> Tue Apr 27 17:36:33 EDT 2010
>>>> Tue Apr 27 17:36:34 EDT 2010
>>>> Tue Apr 27 17:36:34 EDT 2010
>>>> Tue Apr 27 17:36:34 EDT 2010
>>>> Tue Apr 27 17:36:34 EDT 2010
>>>> Tue Apr 27 17:36:34 EDT 2010
>>>> Tue Apr 27 17:36:35 EDT 2010
>>>> Tue Apr 27 17:36:35 EDT 2010
>>>> Tue Apr 27 17:36:35 EDT 2010
>>>>
>>>> On Tue, Apr 27, 2010 at 5:35 PM, Nathan McCall <n...@vervewireless.com> 
>>>> wrote:
>>>>> Have you confirmed that your clocks are all synced in the cluster?
>>>>> This may be the result of an unintentional read-repair occurring if
>>>>> that were the case.
>>>>>
>>>>> -Nate
>>>>>
>>>>> On Tue, Apr 27, 2010 at 2:20 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>>>>> wrote:
>>>>>> Hmm... Even after deleting with cl.ALL, I'm getting data back for some
>>>>>> rows after having deleted them.  Which rows return data is
>>>>>> inconsistent from one run of the job to the next.
>>>>>>
>>>>>> On Tue, Apr 27, 2010 at 1:44 PM, Joost Ouwerkerk <jo...@openplaces.org> 
>>>>>> wrote:
>>>>>>> To check that rows are gone, I check that KeySlice.columns is empty.  
>>>>>>> And as
>>>>>>> I mentioned, immediately after the delete job, this returns the expected
>>>>>>> number.
>>>>>>> Unfortunately I reproduced with QUORUM this morning.  No node outages.  
>>>>>>> I am
>>>>>>> going to try ALL to see if that changes anything, but I am starting to
>>>>>>> wonder if I'm doing something else wrong.
>>>>>>> On Mon, Apr 26, 2010 at 9:45 PM, Jonathan Ellis <jbel...@gmail.com> 
>>>>>>> wrote:
>>>>>>>>
>>>>>>>> How are you checking that the rows are gone?
>>>>>>>>
>>>>>>>> Are you experiencing node outages during this?
>>>>>>>>
>>>>>>>> DC_QUORUM is unfinished code right now, you should avoid using it.
>>>>>>>> Can you reproduce with normal QUORUM?
>>>>>>>>
>>>>>>>> On Sat, Apr 24, 2010 at 12:23 PM, Joost Ouwerkerk 
>>>>>>>> <jo...@openplaces.org>
>>>>>>>> wrote:
>>>>>>>> > I'm having trouble deleting rows in Cassandra.  After running a job 
>>>>>>>> > that
>>>>>>>> > deletes hundreds of rows, I run another job that verifies that the 
>>>>>>>> > rows
>>>>>>>> > are
>>>>>>>> > gone.  Both jobs run correctly.  However, when I run the verification
>>>>>>>> > job an
>>>>>>>> > hour later, the rows have re-appeared.  This is not a case of 
>>>>>>>> > "ghosting"
>>>>>>>> > because the verification job actually checks that there is data in 
>>>>>>>> > the
>>>>>>>> > columns.
>>>>>>>> >
>>>>>>>> > I am running a cluster with 12 nodes and a replication factor of 3.  
>>>>>>>> > I
>>>>>>>> > am
>>>>>>>> > using DC_QUORUM consistency when deleting.
>>>>>>>> >
>>>>>>>> > Any ideas?
>>>>>>>> > Joost.
>>>>>>>> >
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> --
>>>>>>>> Jonathan Ellis
>>>>>>>> Project Chair, Apache Cassandra
>>>>>>>> co-founder of Riptano, the source for professional Cassandra support
>>>>>>>> http://riptano.com
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>>
>>
>> --
>> Jonathan Ellis
>> Project Chair, Apache Cassandra
>> co-founder of Riptano, the source for professional Cassandra support
>> http://riptano.com
>>
>



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of Riptano, the source for professional Cassandra support
http://riptano.com

Reply via email to