Sounds like you need to increase sysctl vm.max_map_count On Fri, Feb 3, 2012 at 7:27 PM, Ajeet Grewal <asgre...@gmail.com> wrote: > Hey guys, > > I am getting an out of memory (mmap failed) error with Cassandra > 1.0.2. The relevant log lines are pasted at > http://pastebin.com/UM28ZC1g. > > Cassandra works fine until it reaches about 300-400GB of load (on one > instance, I have 12 nodes RF=2). Then nodes start failing with such > errors. The nodes are pretty beefy, 32GB of ram, 8 cores. Increasing > the JVM heap size does not help. > > I am running on a 64bit jvm. I am using jna. I have memlock unlimited > for the user. (I confirmed this by looking at /proc/<pid>/limits). > > I also tried restarting the process as root, but it crashes with the same > error. > > Also the number of files that I have in the data directory is about > ~300, so it should not be exceeding the open files limit. > > I don't know if this is relevant. I just have two column families, > counter_object and counter_time. I am using very wide columns, so row > sizes can be huge. You can see from the log link, that the *.db files > are sometimes pretty big. > > Please help! Thank you! > > -- > Regards, > Ajeet
-- Jonathan Ellis Project Chair, Apache Cassandra co-founder of DataStax, the source for professional Cassandra support http://www.datastax.com