We're working on better GC defaults for 0.6.2. Thanks! On Tue, May 11, 2010 at 12:00 PM, B. Todd Burruss <bburr...@real.com> wrote: > another note on this ... since all my nodes are very well balanced and were > started at the same time, i notice that they all do garbage collection at > about the same time. this of course causes a performance issue. > > i also have noticed that with the default JVM options and heavy load, > ConcMarkSweepGC can fall behind and require the JVM to unexpectedly pause > while it plays catchup. adding the following param can help this out, it > says to start processing when "CMS Old Gen" memory is 88% used. > > -XX:CMSInitiatingOccupancyFraction=88 > > my understanding of how the default is calculated, mine was about 92%, so i > only lowered this 4%, but now i can see GC starting earlier and haven't had > a pause like a saw before. > > > On 05/06/2010 02:42 PM, Todd Burruss wrote: > > i think you will see a slow down because of large values in your columns. > make sure you take a look at MemtableThroughputInMB in your config. if you > are writing 1MB of data per row, then you'll probably want to increase this > quite a bit so you are not constantly creating sstables. can't recall, did > you see compaction mgr reporting a lot of pending compactions? maybe try to > "chunk" your data into multiple columns or multiple rows. > > i too see slowness that exhibits in the same manner as you guys have > described. i'm still trying to track it down as well. > > On 05/06/2010 10:56 AM, Ran Tavory wrote: > > Jonathan, I think it's the case of large values in the columns. The > problematic CF is a key-value store, so it has only one column per row, > however the value of that column can be large. It's a java serialized object > (uncompressed) which, may be 100s of bytes, maybe even a few megs. This CF > also suffers from zero cache hits since each time a read is for a unique > key. > I ran stress.py and I see much better results (reads are < 1ms) so I assume > my cluster is healthy, so I need to fix the app. Would 1meg bytes object > explain a 30ms (sometimes even more) read latency? The boxes aren't fancy, > not sure exactly what hardware we have there but it's "commodity"... > Thanks! > > On Thu, May 6, 2010 at 5:22 PM, Jonathan Ellis <jbel...@gmail.com> wrote: >> >> columns, not CFs. >> >> put another way, how wide are the rows in the slow CF? >> >> On Wed, May 5, 2010 at 11:30 PM, Ran Tavory <ran...@gmail.com> wrote: >> > I have a few CFs but the one I'm seeing slowness in, which is the one >> > with >> > plenty of cache misses has only one column per key. >> > Latency varies b/w 10m and 60ms but I'd say average is 30ms. >> > >> > On Thu, May 6, 2010 at 4:25 AM, Jonathan Ellis <jbel...@gmail.com> >> > wrote: >> >> >> >> How many columns are in the rows you are reading from? >> >> >> >> 30ms is quite high, so I suspect you have relatively large rows, in >> >> which case decreasing the column index threshold may help. >> >> -- >> Jonathan Ellis >> Project Chair, Apache Cassandra >> co-founder of Riptano, the source for professional Cassandra support >> http://riptano.com > > >
-- Jonathan Ellis Project Chair, Apache Cassandra co-founder of Riptano, the source for professional Cassandra support http://riptano.com