> On Feb 12, 2016, at 9:24 AM, Skvazh Roman <r...@skvazh.com> wrote:
> 
> I have disabled autocompaction and stop it on highload node.

Does the load decrease and the node answers requests “normally” when you do 
disable auto-compaction? You actually see pending compactions on nodes having 
high load correct?

> Heap is 8Gb. gc_grace is 86400
> All sstables is about 200-300 Mb.

All seems legit here. Using G1 GC?

> $ nodetool compactionstats
> pending tasks: 14

Try to increase the compactors from 4 to 6-8 on a node, disable gossip and let 
it finish compacting and put it back in the ring by enabling gossip. See what 
happens.

The tombstones count growing is because the auto-aucompactions are disabled on 
these nodes. Probably not your issue.

   J.


> 
> $ dstat -lvnr 10
> ---load-avg--- ---procs--- ------memory-usage----- ---paging-- -dsk/total- 
> ---system-- ----total-cpu-usage---- -net/total- --io/total-
> 1m   5m  15m |run blk new| used  buff  cach  free|  in   out | read  writ| 
> int   csw |usr sys idl wai hiq siq| recv  send| read  writ
> 29.4 28.6 23.5|0.0   0 1.2|11.3G  190M 17.6G  407M|   0     0 |7507k 7330k|  
> 13k   40k| 11   1  88   0   0   0|   0     0 |96.5  64.6
> 29.3 28.6 23.5| 29   0 0.9|11.3G  190M 17.6G  408M|   0     0 |   0   
> 189k|9822  2319 | 99   0   0   0   0   0| 138k  120k|   0  4.30
> 29.4 28.6 23.6| 30   0 2.0|11.3G  190M 17.6G  408M|   0     0 |   0    
> 26k|8689  2189 |100   0   0   0   0   0| 139k  120k|   0  2.70
> 29.4 28.7 23.6| 29   0 3.0|11.3G  190M 17.6G  408M|   0     0 |   0    
> 20k|8722  1846 | 99   0   0   0   0   0| 136k  120k|   0  1.50 ^C
> 
> 
> JvmTop 0.8.0 alpha - 15:20:37,  amd64, 16 cpus, Linux 3.14.44-3, load avg 
> 28.09
> http://code.google.com/p/jvmtop
> 
> PID 32505: org.apache.cassandra.service.CassandraDaemon
> ARGS:
> VMARGS: -ea -javaagent:/usr/share/cassandra/lib/jamm-0.3.0.jar -XX:+CMSCl[...]
> VM: Oracle Corporation Java HotSpot(TM) 64-Bit Server VM 1.8.0_65
> UP:  8:31m  #THR: 334  #THRPEAK: 437  #THRCREATED: 4694 USER: cassandra
> GC-Time:  0: 8m   #GC-Runs: 6378      #TotalLoadedClasses: 5926
> CPU: 97.96% GC:  0.00% HEAP:6049m /7540m NONHEAP:  82m /  n/a
> 
>  TID   NAME                                    STATE    CPU  TOTALCPU 
> BLOCKEDBY
>    447 SharedPool-Worker-45                 RUNNABLE 60.47%     1.03%
>    343 SharedPool-Worker-2                  RUNNABLE 56.46%     3.07%
>    349 SharedPool-Worker-8                  RUNNABLE 56.43%     1.61%
>    456 SharedPool-Worker-25                 RUNNABLE 55.25%     1.06%
>    483 SharedPool-Worker-40                 RUNNABLE 53.06%     1.04%
>    475 SharedPool-Worker-53                 RUNNABLE 52.31%     1.03%
>    464 SharedPool-Worker-20                 RUNNABLE 52.00%     1.11%
>    577 SharedPool-Worker-71                 RUNNABLE 51.73%     1.02%
>    404 SharedPool-Worker-10                 RUNNABLE 51.10%     1.29%
>    486 SharedPool-Worker-34                 RUNNABLE 51.06%     1.03%
> Note: Only top 10 threads (according cpu load) are shown!
> 
> 
>> On 12 Feb 2016, at 18:14, Julien Anguenot <jul...@anguenot.org> wrote:
>> 
>> At the time when the load is high and you have to restart, do you see any 
>> pending compactions when using `nodetool compactionstats`?
>> 
>> Possible to see a `nodetool compactionstats` taken *when* the load is too 
>> high?  Have you checked the size of your SSTables for that big table? Any 
>> large ones in there?  What about the Java HEAP configuration on these nodes?
>> 
>> If you have too many tombstones I would try to decrease gc_grace_seconds so 
>> they get cleared out earlier during compactions.
>> 
>>  J.
>> 
>>> On Feb 12, 2016, at 8:45 AM, Skvazh Roman <r...@skvazh.com> wrote:
>>> 
>>> There is 1-4 compactions at that moment.
>>> We have many tombstones, which does not removed.
>>> DroppableTombstoneRatio is 5-6 (greater than 1)
>>> 
>>>> On 12 Feb 2016, at 15:53, Julien Anguenot <jul...@anguenot.org> wrote:
>>>> 
>>>> Hey, 
>>>> 
>>>> What about compactions count when that is happening?
>>>> 
>>>> J.
>>>> 
>>>> 
>>>>> On Feb 12, 2016, at 3:06 AM, Skvazh Roman <r...@skvazh.com> wrote:
>>>>> 
>>>>> Hello!
>>>>> We have a cluster of 25 c3.4xlarge nodes (16 cores, 32 GiB) with attached 
>>>>> 1.5 TB 4000 PIOPS EBS drive.
>>>>> Sometimes one or two nodes user cpu spikes to 100%, load average to 20-30 
>>>>> - read requests drops of.
>>>>> Only restart of this cassandra services helps.
>>>>> Please advice.
>>>>> 
>>>>> One big table with wide rows. 600 Gb per node.
>>>>> LZ4Compressor
>>>>> LeveledCompaction
>>>>> 
>>>>> concurrent compactors: 4
>>>>> compactor throughput: tried from 16 to 128
>>>>> Concurrent_readers: from 16 to 32
>>>>> Concurrent_writers: 128
>>>>> 
>>>>> 
>>>>> https://gist.github.com/rskvazh/de916327779b98a437a6
>>>>> 
>>>>> 
>>>>> JvmTop 0.8.0 alpha - 06:51:10,  amd64, 16 cpus, Linux 3.14.44-3, load avg 
>>>>> 19.35
>>>>> http://code.google.com/p/jvmtop
>>>>> 
>>>>> Profiling PID 9256: org.apache.cassandra.service.CassandraDa
>>>>> 
>>>>> 95.73% (     4.31s) 
>>>>> ....google.common.collect.AbstractIterator.tryToComputeN()
>>>>> 1.39% (     0.06s) com.google.common.base.Objects.hashCode()
>>>>> 1.26% (     0.06s) io.netty.channel.epoll.Native.epollWait()
>>>>> 0.85% (     0.04s) net.jpountz.lz4.LZ4JNI.LZ4_compress_limitedOutput()
>>>>> 0.46% (     0.02s) net.jpountz.lz4.LZ4JNI.LZ4_decompress_fast()
>>>>> 0.26% (     0.01s) com.google.common.collect.Iterators$7.computeNext()
>>>>> 0.06% (     0.00s) io.netty.channel.epoll.Native.eventFdWrite()
>>>>> 
>>>>> 
>>>>> ttop:
>>>>> 
>>>>> 2016-02-12T08:20:25.605+0000 Process summary
>>>>> process cpu=1565.15%
>>>>> application cpu=1314.48% (user=1354.48% sys=-40.00%)
>>>>> other: cpu=250.67%
>>>>> heap allocation rate 146mb/s
>>>>> [000405] user=76.25% sys=-0.54% alloc=     0b/s - SharedPool-Worker-9
>>>>> [000457] user=75.54% sys=-1.26% alloc=     0b/s - SharedPool-Worker-14
>>>>> [000451] user=73.52% sys= 0.29% alloc=     0b/s - SharedPool-Worker-16
>>>>> [000311] user=76.45% sys=-2.99% alloc=     0b/s - SharedPool-Worker-4
>>>>> [000389] user=70.69% sys= 2.62% alloc=     0b/s - SharedPool-Worker-6
>>>>> [000388] user=86.95% sys=-14.28% alloc=     0b/s - SharedPool-Worker-5
>>>>> [000404] user=70.69% sys= 0.10% alloc=     0b/s - SharedPool-Worker-8
>>>>> [000390] user=72.61% sys=-1.82% alloc=     0b/s - SharedPool-Worker-7
>>>>> [000255] user=87.86% sys=-17.87% alloc=     0b/s - SharedPool-Worker-1
>>>>> [000444] user=72.21% sys=-2.30% alloc=     0b/s - SharedPool-Worker-12
>>>>> [000310] user=71.50% sys=-2.31% alloc=     0b/s - SharedPool-Worker-3
>>>>> [000445] user=69.68% sys=-0.83% alloc=     0b/s - SharedPool-Worker-13
>>>>> [000406] user=72.61% sys=-4.40% alloc=     0b/s - SharedPool-Worker-10
>>>>> [000446] user=69.78% sys=-1.65% alloc=     0b/s - SharedPool-Worker-11
>>>>> [000452] user=66.86% sys= 0.22% alloc=     0b/s - SharedPool-Worker-15
>>>>> [000256] user=69.08% sys=-2.42% alloc=     0b/s - SharedPool-Worker-2
>>>>> [004496] user=29.99% sys= 0.59% alloc=   30mb/s - CompactionExecutor:15
>>>>> [004906] user=29.49% sys= 0.74% alloc=   39mb/s - CompactionExecutor:16
>>>>> [010143] user=28.58% sys= 0.25% alloc=   26mb/s - CompactionExecutor:17
>>>>> [000785] user=27.87% sys= 0.70% alloc=   38mb/s - CompactionExecutor:12
>>>>> [012723] user= 9.09% sys= 2.46% alloc= 2977kb/s - RMI TCP 
>>>>> Connection(2673)-127.0.0.1
>>>>> [000555] user= 5.35% sys=-0.08% alloc=  474kb/s - SharedPool-Worker-24
>>>>> [000560] user= 3.94% sys= 0.07% alloc=  434kb/s - SharedPool-Worker-22
>>>>> [000557] user= 3.94% sys=-0.17% alloc=  339kb/s - SharedPool-Worker-25
>>>>> [000447] user= 2.73% sys= 0.60% alloc=  436kb/s - SharedPool-Worker-19
>>>>> [000563] user= 3.33% sys=-0.04% alloc=  460kb/s - SharedPool-Worker-20
>>>>> [000448] user= 2.73% sys= 0.27% alloc=  414kb/s - SharedPool-Worker-21
>>>>> [000554] user= 1.72% sys= 0.70% alloc=  232kb/s - SharedPool-Worker-26
>>>>> [000558] user= 1.41% sys= 0.39% alloc=  213kb/s - SharedPool-Worker-23
>>>>> [000450] user= 1.41% sys=-0.03% alloc=  158kb/s - SharedPool-Worker-17
>>>> 
>> 
>> 
>> 
> 

--
Julien Anguenot (@anguenot)
USA +1.832.408.0344 <tel:+1.832.408.0344>  
FR +33.7.86.85.70.44

Reply via email to