I am new and trying to learn Cassandra.

Based on my understanding of the problem, almost 2Gb is taken up just for
the compression headers heap.

And 100MB per SSTable, and about 30,000 files gives about 3TB of data?

What is the hardware and memory configuration you are using to provide this
large data?

Should this be reduced to smaller data sets, and partitioned into multiple
nodes?

If my understanding is  totally wrong, please forgive and if possible
explain.







On Tue, Feb 18, 2014 at 2:58 PM, Plotnik, Alexey <aplot...@rhonda.ru> wrote:

> Compression buffers are located in Heap, I saw them in Heapdump. That is:
>
>
>
> ======================
>
> public class CompressedRandomAccessReader extends RandomAccessReader {
>
> …..
>
>    private ByteBuffer compressed; // ß THAT IS
>
> ======================
>
>
>
> *From:* Robert Coli [mailto:rc...@eventbrite.com]
> *Sent:* 19 февраля 2014 г. 6:24
> *To:* user@cassandra.apache.org
> *Subject:* Re: Turn off compression (1.2.11)
>
>
>
> On Mon, Feb 17, 2014 at 4:35 PM, Plotnik, Alexey <aplot...@rhonda.ru>
> wrote:
>
>
>
> As an aside, 1.2.0 beta moved a bunch of data related to compression off
> the heap. If you were to try to run the same cluster under 1.1, you'd
> probably OOM your heap immediately.
>
>
>
> https://issues.apache.org/jira/browse/CASSANDRA-4941
>
>
>
> =Rob
>
>
>

Reply via email to