Maybe, maybe not. Presumably if you are running a RDMS with any reasonable
amount of traffic now a days, it's sitting on a machine with 4-8G of memory
at least.

On Wed, Apr 21, 2010 at 10:48 AM, Nicolas Labrot <nith...@gmail.com> wrote:

> Thanks Mark.
>
> Cassandra is maybe too much for my need ;)
>
>
>
> On Wed, Apr 21, 2010 at 4:45 PM, Mark Greene <green...@gmail.com> wrote:
>
>> Hit send to early....
>>
>> That being said a lot of people running Cassandra in production are using
>> 4-6GB max heaps on 8GB machines, don't know if that helps but hopefully
>> gives you some perspective.
>>
>>
>> On Wed, Apr 21, 2010 at 10:39 AM, Mark Greene <green...@gmail.com> wrote:
>>
>>> RAM doesn't necessarily need to be proportional but I would say the
>>> number of nodes does. You can't just throw a bazillion inserts at one node.
>>> This is the main benefit of Cassandra is that if you start hitting your
>>> capacity, you add more machines and distribute the keys across more
>>> machines.
>>>
>>>
>>> On Wed, Apr 21, 2010 at 9:07 AM, Nicolas Labrot <nith...@gmail.com>wrote:
>>>
>>>> So does it means the RAM needed is proportionnal with the data handled ?
>>>>
>>>> Or Cassandra need a minimum amount or RAM when dataset is big?
>>>>
>>>> I must confess this OOM behaviour is strange.
>>>>
>>>>
>>>> On Wed, Apr 21, 2010 at 2:54 PM, Mark Jones <mjo...@imagehawk.com>wrote:
>>>>
>>>>>  On my 4GB machine I’m giving it 3GB and having no trouble with 60+
>>>>> million 500 byte columns
>>>>>
>>>>>
>>>>>
>>>>> *From:* Nicolas Labrot [mailto:nith...@gmail.com]
>>>>> *Sent:* Wednesday, April 21, 2010 7:47 AM
>>>>> *To:* user@cassandra.apache.org
>>>>> *Subject:* Re: Cassandra tuning for running test on a desktop
>>>>>
>>>>>
>>>>>
>>>>> I have try 1400M, and Cassandra OOM too.
>>>>>
>>>>> Is there another solution ? My data isn't very big.
>>>>>
>>>>> It seems that is the merge of the db
>>>>>
>>>>>  On Wed, Apr 21, 2010 at 2:14 PM, Mark Greene <green...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> Trying increasing Xmx. 1G is probably not enough for the amount of
>>>>> inserts you are doing.
>>>>>
>>>>>
>>>>>
>>>>> On Wed, Apr 21, 2010 at 8:10 AM, Nicolas Labrot <nith...@gmail.com>
>>>>> wrote:
>>>>>
>>>>> Hello,
>>>>>
>>>>> For my first message I will first thanks Cassandra contributors for
>>>>> their great works.
>>>>>
>>>>> I have a parameter issue with Cassandra (I hope it's just a parameter
>>>>> issue). I'm using Cassandra 6.0.1 with Hector client on my desktop. It's a
>>>>> simple dual core with 4GB of RAM on WinXP. I have keep the default JVM
>>>>> option inside cassandra.bat (Xmx1G)
>>>>>
>>>>> I'm trying to insert 3 millions of SC with 6 Columns each inside 1 CF
>>>>> (named Super1). The insertion go to 1 millions of SC (without slowdown) 
>>>>> and
>>>>> Cassandra crash because of an OOM. (I store an average of 100 bytes per SC
>>>>> with a max of 10kB).
>>>>> I have aggressively decreased all the memories parameters without any
>>>>> respect to the consistency (My config is here [1]), the cache is turn off
>>>>> but Cassandra still go to OOM. I have joined the last line of the 
>>>>> Cassandra
>>>>> life [2].
>>>>>
>>>>> What can I do to fix my issue ?  Is there another solution than
>>>>> increasing the Xmx ?
>>>>>
>>>>> Thanks for your help,
>>>>>
>>>>> Nicolas
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> [1]
>>>>>   <Keyspaces>
>>>>>     <Keyspace Name="Keyspace1">
>>>>>       <ColumnFamily Name="Super1"
>>>>>                     ColumnType="Super"
>>>>>                     CompareWith="BytesType"
>>>>>                     CompareSubcolumnsWith="BytesType" />
>>>>>
>>>>> <ReplicaPlacementStrategy>org.apache.cassandra.locator.RackUnawareStrategy</ReplicaPlacementStrategy>
>>>>>       <ReplicationFactor>1</ReplicationFactor>
>>>>>
>>>>> <EndPointSnitch>org.apache.cassandra.locator.EndPointSnitch</EndPointSnitch>
>>>>>     </Keyspace>
>>>>>   </Keyspaces>
>>>>>   <CommitLogRotationThresholdInMB>32</CommitLogRotationThresholdInMB>
>>>>>
>>>>>   <DiskAccessMode>auto</DiskAccessMode>
>>>>>   <RowWarningThresholdInMB>64</RowWarningThresholdInMB>
>>>>>   <SlicedBufferSizeInKB>64</SlicedBufferSizeInKB>
>>>>>   <FlushDataBufferSizeInMB>16</FlushDataBufferSizeInMB>
>>>>>   <FlushIndexBufferSizeInMB>4</FlushIndexBufferSizeInMB>
>>>>>   <ColumnIndexSizeInKB>64</ColumnIndexSizeInKB>
>>>>>
>>>>>   <MemtableThroughputInMB>16</MemtableThroughputInMB>
>>>>>   <BinaryMemtableThroughputInMB>32</BinaryMemtableThroughputInMB>
>>>>>   <MemtableOperationsInMillions>0.01</MemtableOperationsInMillions>
>>>>>   <MemtableObjectCountInMillions>0.01</MemtableObjectCountInMillions>
>>>>>   <MemtableFlushAfterMinutes>60</MemtableFlushAfterMinutes>
>>>>>   <ConcurrentReads>4</ConcurrentReads>
>>>>>   <ConcurrentWrites>8</ConcurrentWrites>
>>>>> </Storage>
>>>>>
>>>>>
>>>>> [2]
>>>>>  INFO 13:36:41,062 Super1 has reached its threshold; switching in a
>>>>> fresh Memtable at
>>>>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>>>>> position=5417524)
>>>>>  INFO 13:36:41,062 Enqueuing flush of Memtable(Super1)@15385755
>>>>>  INFO 13:36:41,062 Writing Memtable(Super1)@15385755
>>>>>  INFO 13:36:42,062 Completed flushing
>>>>> d:\cassandra\data\Keyspace1\Super1-711-Data.db
>>>>>  INFO 13:36:45,781 Super1 has reached its threshold; switching in a
>>>>> fresh Memtable at
>>>>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>>>>> position=6065637)
>>>>>  INFO 13:36:45,781 Enqueuing flush of Memtable(Super1)@15578910
>>>>>  INFO 13:36:45,796 Writing Memtable(Super1)@15578910
>>>>>  INFO 13:36:46,109 Completed flushing
>>>>> d:\cassandra\data\Keyspace1\Super1-712-Data.db
>>>>>  INFO 13:36:54,296 GC for ConcurrentMarkSweep: 7149 ms, 58337240
>>>>> reclaimed leaving 922392600 used; max is 1174208512
>>>>>  INFO 13:36:54,593 Super1 has reached its threshold; switching in a
>>>>> fresh Memtable at
>>>>> CommitLogContext(file='d:/cassandra/commitlog\CommitLog-1271849783703.log',
>>>>> position=6722241)
>>>>>  INFO 13:36:54,593 Enqueuing flush of Memtable(Super1)@24468872
>>>>>  INFO 13:36:54,593 Writing Memtable(Super1)@24468872
>>>>>  INFO 13:36:55,421 Completed flushing
>>>>> d:\cassandra\data\Keyspace1\Super1-713-Data.dbjava.lang.OutOfMemoryError:
>>>>> Java heap space
>>>>>  INFO 13:37:08,281 GC for ConcurrentMarkSweep: 5561 ms, 9432 reclaimed
>>>>> leaving 971904520 used; max is 1174208512
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to