Sorry, I meant 15GB heap on the one machine that has less nice CPU% now.
The others are 6GB

On Tue, Dec 16, 2014 at 12:50 PM, Arne Claassen <a...@emotient.com> wrote:
>
> AWS r3.xlarge, 30GB, but only using a Heap of 10GB, new 2GB because we
> might go c3.2xlarge instead if CPU is more important than RAM
> Storage is optimized EBS SSD (but iostat shows no real IO going on)
> Each node only has about 10GB with ownership of 67%, 64.7% & 68.3%.
>
> The node on which I set the Heap to 10GB from 6GB the utlilization has
> dropped to 46%nice now, but the ParNew log messages still continue at the
> same pace. I'm gonna up the HEAP to 20GB for a bit, see if that brings that
> nice CPU further down.
>
> No TombstoneOverflowingExceptions.
>
> On Tue, Dec 16, 2014 at 11:50 AM, Ryan Svihla <rsvi...@datastax.com>
> wrote:
>>
>> What's CPU, RAM, Storage layer, and data density per node? Exact heap
>> settings would be nice. In the logs look for TombstoneOverflowingException
>>
>>
>> On Tue, Dec 16, 2014 at 1:36 PM, Arne Claassen <a...@emotient.com> wrote:
>>>
>>> I'm running 2.0.10.
>>>
>>> The data is all time series data and as we change our pipeline, we've
>>> been periodically been reprocessing the data sources, which causes each
>>> time series to be overwritten, i.e. every row per partition key is deleted
>>> and re-written, so I assume i've been collecting a bunch of tombstones.
>>>
>>> Also, the presence of the ever present and never completing compaction
>>> types, i assumed were an artifact of tombstoning, but i fully admit to
>>> conjecture based on about ~20 blog posts and stackoverflow questions i've
>>> surveyed.
>>>
>>> I doubled the Heap on one node and it changed nothing regarding the load
>>> or the ParNew log statements. New Generation Usage is 50%, Eden itself is
>>> 56%.
>>>
>>> Anything else i should look at and report, let me know.
>>>
>>> On Tue, Dec 16, 2014 at 11:14 AM, Jonathan Lacefield <
>>> jlacefi...@datastax.com> wrote:
>>>>
>>>> Hello,
>>>>
>>>>   What version of Cassandra are you running?
>>>>
>>>>   If it's 2.0, we recently experienced something similar with 8447 [1],
>>>> which 8485 [2] should hopefully resolve.
>>>>
>>>>   Please note that 8447 is not related to tombstones.  Tombstone
>>>> processing can put a lot of pressure on the heap as well. Why do you think
>>>> you have a lot of tombstones in that one particular table?
>>>>
>>>>   [1] https://issues.apache.org/jira/browse/CASSANDRA-8447
>>>>   [2] https://issues.apache.org/jira/browse/CASSANDRA-8485
>>>>
>>>> Jonathan
>>>>
>>>> [image: datastax_logo.png]
>>>>
>>>> Jonathan Lacefield
>>>>
>>>> Solution Architect | (404) 822 3487 | jlacefi...@datastax.com
>>>>
>>>> [image: linkedin.png] <http://www.linkedin.com/in/jlacefield/> [image:
>>>> facebook.png] <https://www.facebook.com/datastax> [image: twitter.png]
>>>> <https://twitter.com/datastax> [image: g+.png]
>>>> <https://plus.google.com/+Datastax/about>
>>>> <http://feeds.feedburner.com/datastax> <https://github.com/datastax/>
>>>>
>>>> On Tue, Dec 16, 2014 at 2:04 PM, Arne Claassen <a...@emotient.com>
>>>> wrote:
>>>>>
>>>>> I have a three node cluster that has been sitting at a load of 4 (for
>>>>> each node), 100% CPI utilization (although 92% nice) for that last 12
>>>>> hours, ever since some significant writes finished. I'm trying to 
>>>>> determine
>>>>> what tuning I should be doing to get it out of this state. The debug log 
>>>>> is
>>>>> just an endless series of:
>>>>>
>>>>> DEBUG [ScheduledTasks:1] 2014-12-16 19:03:35,042 GCInspector.java
>>>>> (line 118) GC for ParNew: 166 ms for 10 collections, 4400928736 used; max
>>>>> is 8000634880
>>>>> DEBUG [ScheduledTasks:1] 2014-12-16 19:03:36,043 GCInspector.java
>>>>> (line 118) GC for ParNew: 165 ms for 10 collections, 4440011176 used; max
>>>>> is 8000634880
>>>>> DEBUG [ScheduledTasks:1] 2014-12-16 19:03:37,043 GCInspector.java
>>>>> (line 118) GC for ParNew: 135 ms for 8 collections, 4402220568 used;
>>>>> max is 8000634880
>>>>>
>>>>> iostat shows virtually no I/O.
>>>>>
>>>>> Compaction may enter into this, but i don't really know what to make
>>>>> of compaction stats since they never change:
>>>>>
>>>>> [root@cassandra-37919c3a ~]# nodetool compactionstats
>>>>> pending tasks: 10
>>>>>           compaction type        keyspace           table
>>>>> completed           total      unit  progress
>>>>>                Compaction           mediamedia_tracks_raw
>>>>> 271651482       563615497     bytes    48.20%
>>>>>                Compaction           mediamedia_tracks_raw
>>>>>  30308910     21676695677     bytes     0.14%
>>>>>                Compaction           mediamedia_tracks_raw
>>>>>  1198384080      1815603161     bytes    66.00%
>>>>> Active compaction remaining time :   0h22m24s
>>>>>
>>>>> 5 minutes later:
>>>>>
>>>>> [root@cassandra-37919c3a ~]# nodetool compactionstats
>>>>> pending tasks: 9
>>>>>           compaction type        keyspace           table
>>>>> completed           total      unit  progress
>>>>>                Compaction           mediamedia_tracks_raw
>>>>> 271651482       563615497     bytes    48.20%
>>>>>                Compaction           mediamedia_tracks_raw
>>>>>  30308910     21676695677     bytes     0.14%
>>>>>                Compaction           mediamedia_tracks_raw
>>>>>  1198384080      1815603161     bytes    66.00%
>>>>> Active compaction remaining time :   0h22m24s
>>>>>
>>>>> Sure the pending tasks went down by one, but the rest is identical.
>>>>> media_tracks_raw likely has a bunch of tombstones (can't figure out how to
>>>>> get stats on that).
>>>>>
>>>>> Is this behavior something that indicates that i need more Heap,
>>>>> larger new generation? Should I be manually running compaction on tables
>>>>> with lots of tombstones?
>>>>>
>>>>> Any suggestions or places to educate myself better on performance
>>>>> tuning would be appreciated.
>>>>>
>>>>> arne
>>>>>
>>>>
>>
>> --
>>
>> [image: datastax_logo.png] <http://www.datastax.com/>
>>
>> Ryan Svihla
>>
>> Solution Architect
>>
>> [image: twitter.png] <https://twitter.com/foundev> [image: linkedin.png]
>> <http://www.linkedin.com/pub/ryan-svihla/12/621/727/>
>>
>> DataStax is the fastest, most scalable distributed database technology,
>> delivering Apache Cassandra to the world’s most innovative enterprises.
>> Datastax is built to be agile, always-on, and predictably scalable to any
>> size. With more than 500 customers in 45 countries, DataStax is the
>> database technology and transactional backbone of choice for the worlds
>> most innovative companies such as Netflix, Adobe, Intuit, and eBay.
>>
>>

Reply via email to