There is 1-4 compactions at that moment.
We have many tombstones, which does not removed.
DroppableTombstoneRatio is 5-6 (greater than 1)

> On 12 Feb 2016, at 15:53, Julien Anguenot <jul...@anguenot.org> wrote:
> 
> Hey, 
> 
> What about compactions count when that is happening?
> 
>   J.
> 
> 
>> On Feb 12, 2016, at 3:06 AM, Skvazh Roman <r...@skvazh.com> wrote:
>> 
>> Hello!
>> We have a cluster of 25 c3.4xlarge nodes (16 cores, 32 GiB) with attached 
>> 1.5 TB 4000 PIOPS EBS drive.
>> Sometimes one or two nodes user cpu spikes to 100%, load average to 20-30 - 
>> read requests drops of.
>> Only restart of this cassandra services helps.
>> Please advice.
>> 
>> One big table with wide rows. 600 Gb per node.
>> LZ4Compressor
>> LeveledCompaction
>> 
>> concurrent compactors: 4
>> compactor throughput: tried from 16 to 128
>> Concurrent_readers: from 16 to 32
>> Concurrent_writers: 128
>> 
>> 
>> https://gist.github.com/rskvazh/de916327779b98a437a6
>> 
>> 
>> JvmTop 0.8.0 alpha - 06:51:10,  amd64, 16 cpus, Linux 3.14.44-3, load avg 
>> 19.35
>> http://code.google.com/p/jvmtop
>> 
>> Profiling PID 9256: org.apache.cassandra.service.CassandraDa
>> 
>> 95.73% (     4.31s) 
>> ....google.common.collect.AbstractIterator.tryToComputeN()
>>  1.39% (     0.06s) com.google.common.base.Objects.hashCode()
>>  1.26% (     0.06s) io.netty.channel.epoll.Native.epollWait()
>>  0.85% (     0.04s) net.jpountz.lz4.LZ4JNI.LZ4_compress_limitedOutput()
>>  0.46% (     0.02s) net.jpountz.lz4.LZ4JNI.LZ4_decompress_fast()
>>  0.26% (     0.01s) com.google.common.collect.Iterators$7.computeNext()
>>  0.06% (     0.00s) io.netty.channel.epoll.Native.eventFdWrite()
>> 
>> 
>> ttop:
>> 
>> 2016-02-12T08:20:25.605+0000 Process summary
>> process cpu=1565.15%
>> application cpu=1314.48% (user=1354.48% sys=-40.00%)
>> other: cpu=250.67%
>> heap allocation rate 146mb/s
>> [000405] user=76.25% sys=-0.54% alloc=     0b/s - SharedPool-Worker-9
>> [000457] user=75.54% sys=-1.26% alloc=     0b/s - SharedPool-Worker-14
>> [000451] user=73.52% sys= 0.29% alloc=     0b/s - SharedPool-Worker-16
>> [000311] user=76.45% sys=-2.99% alloc=     0b/s - SharedPool-Worker-4
>> [000389] user=70.69% sys= 2.62% alloc=     0b/s - SharedPool-Worker-6
>> [000388] user=86.95% sys=-14.28% alloc=     0b/s - SharedPool-Worker-5
>> [000404] user=70.69% sys= 0.10% alloc=     0b/s - SharedPool-Worker-8
>> [000390] user=72.61% sys=-1.82% alloc=     0b/s - SharedPool-Worker-7
>> [000255] user=87.86% sys=-17.87% alloc=     0b/s - SharedPool-Worker-1
>> [000444] user=72.21% sys=-2.30% alloc=     0b/s - SharedPool-Worker-12
>> [000310] user=71.50% sys=-2.31% alloc=     0b/s - SharedPool-Worker-3
>> [000445] user=69.68% sys=-0.83% alloc=     0b/s - SharedPool-Worker-13
>> [000406] user=72.61% sys=-4.40% alloc=     0b/s - SharedPool-Worker-10
>> [000446] user=69.78% sys=-1.65% alloc=     0b/s - SharedPool-Worker-11
>> [000452] user=66.86% sys= 0.22% alloc=     0b/s - SharedPool-Worker-15
>> [000256] user=69.08% sys=-2.42% alloc=     0b/s - SharedPool-Worker-2
>> [004496] user=29.99% sys= 0.59% alloc=   30mb/s - CompactionExecutor:15
>> [004906] user=29.49% sys= 0.74% alloc=   39mb/s - CompactionExecutor:16
>> [010143] user=28.58% sys= 0.25% alloc=   26mb/s - CompactionExecutor:17
>> [000785] user=27.87% sys= 0.70% alloc=   38mb/s - CompactionExecutor:12
>> [012723] user= 9.09% sys= 2.46% alloc= 2977kb/s - RMI TCP 
>> Connection(2673)-127.0.0.1
>> [000555] user= 5.35% sys=-0.08% alloc=  474kb/s - SharedPool-Worker-24
>> [000560] user= 3.94% sys= 0.07% alloc=  434kb/s - SharedPool-Worker-22
>> [000557] user= 3.94% sys=-0.17% alloc=  339kb/s - SharedPool-Worker-25
>> [000447] user= 2.73% sys= 0.60% alloc=  436kb/s - SharedPool-Worker-19
>> [000563] user= 3.33% sys=-0.04% alloc=  460kb/s - SharedPool-Worker-20
>> [000448] user= 2.73% sys= 0.27% alloc=  414kb/s - SharedPool-Worker-21
>> [000554] user= 1.72% sys= 0.70% alloc=  232kb/s - SharedPool-Worker-26
>> [000558] user= 1.41% sys= 0.39% alloc=  213kb/s - SharedPool-Worker-23
>> [000450] user= 1.41% sys=-0.03% alloc=  158kb/s - SharedPool-Worker-17
> 

Reply via email to