This talk from DataStax was talking about deletes as an Anti-pattern. It may be worth watching.
> Thanks for your interest in the following webinar: > > Avoiding anti-patterns: How to stay in love with Cassandra > > Here are the links to the video recording and presentation slides. > > Thanks, > > The DataStax Team On Mar 24, 2015, at 9:09 AM, joss Earl <j...@rareformnewmedia.com> wrote: > Hi Anuj > > Yes, thanks.. looking at my log file I see: > > ERROR [SharedPool-Worker-2] 2015-03-24 13:52:06,751 SliceQueryFilter.java:218 > - Scanned over 100000 tombstones in test1.msg; query aborted (see tombstone_\ > failure_threshold) > WARN [SharedPool-Worker-2] 2015-03-24 13:52:06,759 > AbstractTracingAwareExecutorService.java:169 - Uncaught exception on thread > Thread[SharedPool-Worker-2\ > ,5,main]: {} > java.lang.RuntimeException: > org.apache.cassandra.db.filter.TombstoneOverwhelmingException > > I'm reading up about how to deal with this now, thanks.. > > On 24 March 2015 at 13:16, Anuj Wadehra <anujw_2...@yahoo.co.in> wrote: > Hi Joss > > We faced similar issue recently. The problem seems to be related to huge > number of tombstones generated after deletion. I would suggest you to > increase tombstone warning and failure threshold in cassandra.yaml. > > Once you do that and run your program make sure that you monitor Cassandra > Heap usage using nodetool info command. If heap is near full Cassandra halts > are obvious. So you need to increase heap. > > Due to increased tombstones your query is unable to complete within short > time..I would suggest increasing read timeout in cassandra.yaml so that query > may complete. > > Please look at your logs to make sure that there are no exception. > > Thanks > Anuj Wadehra > > > From:"joss Earl" <j...@rareformnewmedia.com> > Date:Tue, 24 Mar, 2015 at 6:17 pm > Subject:Re: error deleting messages > > It inserts 100,000 messages, I then start deleting the messages by grabbing > chunks of 100 at a time and then individually deleting each message. > > So, the 100,000 messages get inserted without any trouble, I run into trouble > when I have deleted about half of them. I've run this on machines with 4,8, > and 16gig of ram and behaviour was consistent (I fail after 50000 or so > messages on that table, or maybe 30,000 messages on a table with more > columns). > > > > On 24 March 2015 at 12:35, Ali Akhtar <ali.rac...@gmail.com> wrote: > 50100 inserts or deletes? also how much ram / cpu do you have on the server > running this, and what's the ram / cpu usage at about the time it fails? > > On Tue, Mar 24, 2015 at 5:29 PM, joss Earl <j...@rareformnewmedia.com> wrote: > on a stock install, it gets to about 50100 before grinding to a halt > > > > On 24 March 2015 at 12:19, Ali Akhtar <ali.rac...@gmail.com> wrote: > What happens when you run it? How far does it get before stopping? > > On Tue, Mar 24, 2015 at 5:13 PM, joss Earl <j...@rareformnewmedia.com> wrote: > sure: https://gist.github.com/joss75321/7d85e4c75c06530e9d80 > > On 24 March 2015 at 12:04, Ali Akhtar <ali.rac...@gmail.com> wrote: > Can you put your code on gist.github.com or pastebin? > > On Tue, Mar 24, 2015 at 4:58 PM, joss Earl <j...@rareformnewmedia.com> wrote: > I run into trouble after a while if I delete rows, this happens in both 2.1.3 > and 2.0.13, and I encountered the same problem when using either the datastax > java driver or the stock python driver. > The problem is reproducible using the attached python program. > > Once the problem is encountered, the table becomes unusable.. > > cqlsh:test1> select id from msg limit 1; > Request did not complete within rpc_timeout. > > So, questions are: > am I doing something wrong ? > is this expected behaviour ? > is there some way to fix the table and make it usable again once this has > happened ? > if this is a bug, what is the best way of reporting it ? > > Many thanks > Joss > > > > > > >