I have try 1400M, and Cassandra OOM too.

Is there another solution ? My data isn't very big.

It seems that is the merge of the db


On Wed, Apr 21, 2010 at 2:14 PM, Mark Greene <green...@gmail.com> wrote:

> Trying increasing Xmx. 1G is probably not enough for the amount of inserts
> you are doing.
>
>
> On Wed, Apr 21, 2010 at 8:10 AM, Nicolas Labrot <nith...@gmail.com> wrote:
>
>> Hello,
>>
>> For my first message I will first thanks Cassandra contributors for their
>> great works.
>>
>> I have a parameter issue with Cassandra (I hope it's just a parameter
>> issue). I'm using Cassandra 6.0.1 with Hector client on my desktop. It's a
>> simple dual core with 4GB of RAM on WinXP. I have keep the default JVM
>> option inside cassandra.bat (Xmx1G)
>>
>> I'm trying to insert 3 millions of SC with 6 Columns each inside 1 CF
>> (named Super1). The insertion go to 1 millions of SC (without slowdown) and
>> Cassandra crash because of an OOM. (I store an average of 100 bytes per SC
>> with a max of 10kB).
>> I have aggressively decreased all the memories parameters without any
>> respect to the consistency (My config is here [1]), the cache is turn off
>> but Cassandra still go to OOM. I have joined the last line of the Cassandra
>> life [2].
>>
>> What can I do to fix my issue ?  Is there another solution than increasing
>> the Xmx ?
>>
>> Thanks for your help,
>>
>> Nicolas
>>
>>
>>
>>
>>
>> [1]
>>   <Keyspaces>
>>     <Keyspace Name="Keyspace1">
>>       <ColumnFamily Name="Super1"
>>                     ColumnType="Super"
>>                     CompareWith="BytesType"
>>                     CompareSubcolumnsWith="BytesType" />
>>
>> <ReplicaPlacementStrategy>org.apache.cassandra.locator.RackUnawareStrategy</ReplicaPlacementStrategy>
>>       <ReplicationFactor>1</ReplicationFactor>
>>
>> <EndPointSnitch>org.apache.cassandra.locator.EndPointSnitch</EndPointSnitch>
>>     </Keyspace>
>>   </Keyspaces>
>>   <CommitLogRotationThresholdInMB>32</CommitLogRotationThresholdInMB>
>>
>>   <DiskAccessMode>auto</DiskAccessMode>
>>   <RowWarningThresholdInMB>64</RowWarningThresholdInMB>
>>   <SlicedBufferSizeInKB>64</SlicedBufferSizeInKB>
>>   <FlushDataBufferSizeInMB>16</FlushDataBufferSizeInMB>
>>   <FlushIndexBufferSizeInMB>4</FlushIndexBufferSizeInMB>
>>   <ColumnIndexSizeInKB>64</ColumnIndexSizeInKB>
>>
>>   <MemtableThroughputInMB>16</MemtableThroughputInMB>
>>   <BinaryMemtableThroughputInMB>32</BinaryMemtableThroughputInMB>
>>   <MemtableOperationsInMillions>0.01</MemtableOperationsInMillions>
>>   <MemtableObjectCountInMillions>0.01</MemtableObjectCountInMillions>
>>   <MemtableFlushAfterMinutes>60</MemtableFlushAfterMinutes>
>>   <ConcurrentReads>4</ConcurrentReads>
>>   <ConcurrentWrites>8</ConcurrentWrites>
>> </Storage>
>>
>>
>> [2]
>>  INFO 13:36:41,062 Super1 has reached its threshold; switching in a fresh
>> Memtable at
>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>> position=5417524)
>>  INFO 13:36:41,062 Enqueuing flush of Memtable(Super1)@15385755
>>  INFO 13:36:41,062 Writing Memtable(Super1)@15385755
>>  INFO 13:36:42,062 Completed flushing
>> d:\cassandra\data\Keyspace1\Super1-711-Data.db
>>  INFO 13:36:45,781 Super1 has reached its threshold; switching in a fresh
>> Memtable at
>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>> position=6065637)
>>  INFO 13:36:45,781 Enqueuing flush of Memtable(Super1)@15578910
>>  INFO 13:36:45,796 Writing Memtable(Super1)@15578910
>>  INFO 13:36:46,109 Completed flushing
>> d:\cassandra\data\Keyspace1\Super1-712-Data.db
>>  INFO 13:36:54,296 GC for ConcurrentMarkSweep: 7149 ms, 58337240 reclaimed
>> leaving 922392600 used; max is 1174208512
>>  INFO 13:36:54,593 Super1 has reached its threshold; switching in a fresh
>> Memtable at
>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>> position=6722241)
>>  INFO 13:36:54,593 Enqueuing flush of Memtable(Super1)@24468872
>>  INFO 13:36:54,593 Writing Memtable(Super1)@24468872
>>  INFO 13:36:55,421 Completed flushing
>> d:\cassandra\data\Keyspace1\Super1-713-Data.dbjava.lang.OutOfMemoryError:
>> Java heap space
>>  INFO 13:37:08,281 GC for ConcurrentMarkSweep: 5561 ms, 9432 reclaimed
>> leaving 971904520 used; max is 1174208512
>>
>>
>

Reply via email to