> bloom_filter_fp_chance value that was changed from default to 0.1, looked at > the filters and they are about 2.5G on disk and I have around 8G of heap. > I will try increasing the value to 0.7 and report my results. You need to re-write the sstables on disk using nodetool upgradesstables. Otherwise only the new tables with have the 0.1 setting.
> I will try increasing the value to 0.7 and report my results. No need to, it will probably be something like "Oh no, really, what, how, please make it stop" :) 0.7 will mean reads will hit most / all of the SSTables for the CF. I covered a high row situation in on of my talks at the summit this month, the slide deck is here http://www.slideshare.net/aaronmorton/cassandra-sf-2013-in-case-of-emergency-break-glass and the videos will soon be up at Planet Cassandra. Rebuild the sstables, then reduce the index_interval if you still need to reduce mem pressure. Cheers ----------------- Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton http://www.thelastpickle.com On 22/06/2013, at 1:17 PM, sankalp kohli <kohlisank...@gmail.com> wrote: > I will take a heap dump and see whats in there rather than guessing. > > > On Fri, Jun 21, 2013 at 4:12 PM, Bryan Talbot <btal...@aeriagames.com> wrote: > bloom_filter_fp_chance = 0.7 is probably way too large to be effective and > you'll probably have issues compacting deleted rows and get poor read > performance with a value that high. I'd guess that anything larger than 0.1 > might as well be 1.0. > > -Bryan > > > > On Fri, Jun 21, 2013 at 5:58 AM, srmore <comom...@gmail.com> wrote: > > On Fri, Jun 21, 2013 at 2:53 AM, aaron morton <aa...@thelastpickle.com> wrote: >> > nodetool -h localhost flush didn't do much good. > Do you have 100's of millions of rows ? > If so see recent discussions about reducing the bloom_filter_fp_chance and > index_sampling. > Yes, I have 100's of millions of rows. > > > If this is an old schema you may be using the very old setting of 0.000744 > which creates a lot of bloom filters. > > bloom_filter_fp_chance value that was changed from default to 0.1, looked at > the filters and they are about 2.5G on disk and I have around 8G of heap. > I will try increasing the value to 0.7 and report my results. > > It also appears to be a case of hard GC failure (as Rob mentioned) as the > heap is never released, even after 24+ hours of idle time, the JVM needs to > be restarted to reclaim the heap. > > Cheers > > ----------------- > Aaron Morton > Freelance Cassandra Consultant > New Zealand > > @aaronmorton > http://www.thelastpickle.com > > On 20/06/2013, at 6:36 AM, Wei Zhu <wz1...@yahoo.com> wrote: > >> If you want, you can try to force the GC through Jconsole. Memory->Perform >> GC. >> >> It theoretically triggers a full GC and when it will happen depends on the >> JVM >> >> -Wei >> >> From: "Robert Coli" <rc...@eventbrite.com> >> To: user@cassandra.apache.org >> Sent: Tuesday, June 18, 2013 10:43:13 AM >> Subject: Re: Heap is not released and streaming hangs at 0% >> >> On Tue, Jun 18, 2013 at 10:33 AM, srmore <comom...@gmail.com> wrote: >> > But then shouldn't JVM C G it eventually ? I can still see Cassandra alive >> > and kicking but looks like the heap is locked up even after the traffic is >> > long stopped. >> >> No, when GC system fails this hard it is often a permanent failure >> which requires a restart of the JVM. >> >> > nodetool -h localhost flush didn't do much good. >> >> This adds support to the idea that your heap is too full, and not full >> of memtables. >> >> You could try nodetool -h localhost invalidatekeycache, but that >> probably will not free enough memory to help you. >> >> =Rob > > > >