The serializing cache is basically optimal. Your problem is really that row cache is not designed for wide rows at all. See https://issues.apache.org/jira/browse/CASSANDRA-1956
On Thu, Jan 12, 2012 at 10:46 PM, Todd Burruss <bburr...@expedia.com> wrote: > after looking through the code it seems fairly straight forward to create > some different cache providers and try some things. > > has anyone tried ehcache w/o persistence? I see this JIRA > https://issues.apache.org/jira/browse/CASSANDRA-1945 but the main > complaint was the disk serialization, which I don't think anyone wants. > > > On 1/12/12 6:18 PM, "Jonathan Ellis" <jbel...@gmail.com> wrote: > >>8x is pretty normal for JVM and bookkeeping overhead with the CLHCP. >> >>The SerializedCacheProvider is the default in 1.0 and is much >>lighter-weight. >> >>On Thu, Jan 12, 2012 at 6:07 PM, Todd Burruss <bburr...@expedia.com> >>wrote: >>> I'm using ConcurrentLinkedHashCacheProvider and my data on disk is >>>about 4gb, but the RAM used by the cache is around 25gb. I have 70k >>>columns per row, and only about 2500 rows so a lot more columns than >>>rows. has there been any discussion or JIRAs discussing reducing the >>>size of the cache? I can understand the overhead for column names, etc, >>>but the ratio seems a bit distorted. >>> >>> I'm tracing through the code, so any pointers to help me understand is >>>appreciated >>> >>> thx >> >> >> >>-- >>Jonathan Ellis >>Project Chair, Apache Cassandra >>co-founder of DataStax, the source for professional Cassandra support >>http://www.datastax.com > -- Jonathan Ellis Project Chair, Apache Cassandra co-founder of DataStax, the source for professional Cassandra support http://www.datastax.com