Hi, I am not sure I understood your message correctly but I will try to answer it.
but, I think, in Cassandra case, it seems a matter of how much data we use > with how much memory we have. If you are saying you can use poor commodity servers (vertically scale poorly) and just add nodes (horizontal scaling) when the cluster is not powerful enough, you need to know that a minimum of vertical scaling is needed to have great performances or a good stability. Yet, tuning things, you can probably reach a stable state with t2.medium if there is enough t2.medium to handle the load. with default configuration except for leveledCompactionStrategy LeveledCompactionStrategy is heavier to maintain than STCS. On such an environment, read latency is probably not your main concern, and using STCS could give better results as it is way lighter in terms of compactions (Depends on your use case though). I also used 4GM RAM machine (t2.medium) > With 4GB of RAM you probably want to use 1 GB of heap. What version of cassandra are you using ? You might also need to tune bloomfilters, index_interval, memtables size and type, and a few other things to reduce the memory footprint. About compaction, use only half of the cores as concurrent compactors (one core) and see if this improves stability and compaction can still keep up. Or keep 2 and reduce it speed by lowering the compaction throughput. Use nodetool {tpstats, compactionstats, cfstats, cfhistograms} to monitor things and see what to tune. As told earlier, using this low spec machines is fine if you know how to tune Cassandra and can afford some research / tuning time... Alain ----------------------- Alain Rodriguez - al...@thelastpickle.com France The Last Pickle - Apache Cassandra Consulting http://www.thelastpickle.com 2016-03-12 6:58 GMT+01:00 Hiroyuki Yamada <mogwa...@gmail.com>: > Thank you all to respond and discuss my question. > > I agree with you all basically, > but, I think, in Cassandra case, it seems a matter of how much data we use > with how much memory we have. > > As Jack's (and datastax's) suggestion, > I also used 4GM RAM machine (t2.medium) with 1 billion records (about > 100GB in size) with default configuration except for > leveledCompactionStrategy, > but after completion of insertion from an application program, probably > compaction kept working, > and again, later Cassandra was killed by OOM killer. > > Insertion from application side is finished, so the issue is maybe from > compaction happening in background. > Is there any recommended configuration in compaction to make Cassandra > stable with large dataset (more than 100GB) with kind of low memory (4GB) > environment ? > > I think it would be the same thing if I try the experiment with 8GB memory > and larger data set (maybe more than 2 billion records). > (If it is not correct, please explain why.) > > > Best regards, > Hiro > > On Fri, Mar 11, 2016 at 4:19 AM, Robert Coli <rc...@eventbrite.com> wrote: > >> On Thu, Mar 10, 2016 at 3:27 AM, Alain RODRIGUEZ <arodr...@gmail.com> >> wrote: >> >>> So, like Jack, I globally really not recommend it unless you know what >>> you are doing and don't care about facing those issues. >>> >> >> Certainly a spectrum of views here, but everyone (including OP) seems to >> agree with the above. :D >> >> =Rob >> >> > >