The serializing cache is basically optimal.  Your problem is really
that row cache is not designed for wide rows at all.  See
https://issues.apache.org/jira/browse/CASSANDRA-1956

On Thu, Jan 12, 2012 at 10:46 PM, Todd Burruss <bburr...@expedia.com> wrote:
> after looking through the code it seems fairly straight forward to create
> some different cache providers and try some things.
>
> has anyone tried ehcache w/o persistence?  I see this JIRA
> https://issues.apache.org/jira/browse/CASSANDRA-1945 but the main
> complaint was the disk serialization, which I don't think anyone wants.
>
>
> On 1/12/12 6:18 PM, "Jonathan Ellis" <jbel...@gmail.com> wrote:
>
>>8x is pretty normal for JVM and bookkeeping overhead with the CLHCP.
>>
>>The SerializedCacheProvider is the default in 1.0 and is much
>>lighter-weight.
>>
>>On Thu, Jan 12, 2012 at 6:07 PM, Todd Burruss <bburr...@expedia.com>
>>wrote:
>>> I'm using ConcurrentLinkedHashCacheProvider and my data on disk is
>>>about 4gb, but the RAM used by the cache is around 25gb.  I have 70k
>>>columns per row, and only about 2500 rows ­ so a lot more columns than
>>>rows.  has there been any discussion or JIRAs discussing reducing the
>>>size of the cache?  I can understand the overhead for column names, etc,
>>>but the ratio seems a bit distorted.
>>>
>>> I'm tracing through the code, so any pointers to help me understand is
>>>appreciated
>>>
>>> thx
>>
>>
>>
>>--
>>Jonathan Ellis
>>Project Chair, Apache Cassandra
>>co-founder of DataStax, the source for professional Cassandra support
>>http://www.datastax.com
>



-- 
Jonathan Ellis
Project Chair, Apache Cassandra
co-founder of DataStax, the source for professional Cassandra support
http://www.datastax.com

Reply via email to