I assume if I turn off swap it will just die earlier, no ? What is the mechanism of dying ?
>From the link you provided # Row cache is too large, or is caching large rows my row_cache is 0 # The memtable sizes are too large for the amount of heap allocated to the JVM Is my memtable size too large ? I have made it less to surely fit the "magical formula" Trying to analyze heap dumps gives me the following: In one case diagram has 3 Memtables about 64 Mb each + 72 Mb "Thread" + 700 Mb "Unreachable objects" suspected threats: 7 instances of "org.apache.cassandra.db.Memtable", loaded by "sun.misc.Launcher$AppClassLoader @ 0x7f29f4992d68" occupy 456,292,912 (48.36%) bytes. 25,211 instances of "org.apache.cassandra.io.sstable.SSTableReader", loaded by "sun.misc.Launcher$AppClassLoader @ 0x7f29f4992d68" occupy 294,908,984 (31.26%) byte 72 instances of "java.lang.Thread", loaded by "<system class loader>" occupy 143,632,624 (15.22%) bytes. In other cases memory analyzer hangs trying to parse 2Gb dump 22.04.2011, 17:26, "Jonathan Ellis" <jbel...@gmail.com>;: > (0) turn off swap > (1) > http://www.datastax.com/docs/0.7/troubleshooting/index#nodes-are-dying-with-oom-errors > > On Fri, Apr 22, 2011 at 8:00 AM, Nikolay Kоvshov <nkovs...@yandex.ru>;; > wrote: >> I am using Cassandra 0.7.0 with following settings >> >> binary_memtable_throughput_in_mb: 64 >> in_memory_compaction_limit_in_mb: 64 >> keys_cached 1 million >> rows_cached 0 >> >> RAM for Cassandra 2 GB >> >> I run very simple test >> >> 1 Node with 4 HDDs (1 HDD - commitlog and caches, 3 HDDs - data) >> 1 KS => 1 CF => 1 Column >> >> I insert data (random key 64 bytes + value 64 bytes) at a maximum possible >> speed, trying to hit disk i/o, calculate speed and make sure Cassandra stays >> alive. It doesn't, unfortunately. >> After several hundreds millions of inserts Cassandra always does down by >> OOM. Getting it up again doesn't help - after inserting some new data it >> goes down again. By this time Cassandra goes to swap and has a lot of tasks >> pending. I am not inserting anything now and tasks sloooowly disappear, but >> it will take her weeks to do all of them. >> >> compaction type: Minor >> column family: Standard1 >> bytes compacted: 3661003227 >> bytes total in progress: 4176296448 >> pending tasks: 630 >> >> So, what am I (or Cassandra) doing wrong ? I dont want to get Cassandra >> crashed without means of repair on heavy load circumstances. > -- > Jonathan Ellis > Project Chair, Apache Cassandra > co-founder of DataStax, the source for professional Cassandra support > http://www.datastax.com