> Changed memtable_total_space_in_mb to 1024 still no luck.
Reducing memtable_total_space_in_mb will increase the frequency of flushing to 
disk, which will create more for compaction to do and result in increased IO. 

You should return it to the default.

> when I send traffic to one node its performance is 2x more than when I send 
> traffic to all the nodes.
>  
What are you measuring, request latency or local read/write latency ? 

If it’s write latency it’s probably GC, if it’s read is probably IO or data 
model. 

Hope that helps. 

-----------------
Aaron Morton
New Zealand
@aaronmorton

Co-Founder & Principal Consultant
Apache Cassandra Consulting
http://www.thelastpickle.com

On 7/12/2013, at 8:05 am, srmore <comom...@gmail.com> wrote:

> Changed memtable_total_space_in_mb to 1024 still no luck.
> 
> 
> On Fri, Dec 6, 2013 at 11:05 AM, Vicky Kak <vicky....@gmail.com> wrote:
> Can you set the memtable_total_space_in_mb value, it is defaulting to 1/3 
> which is 8/3 ~ 2.6 gb in capacity
> http://www.datastax.com/dev/blog/whats-new-in-cassandra-1-0-improved-memory-and-disk-space-management
> 
> The flushing of 2.6 gb to the disk might slow the performance if frequently 
> called, may be you have lots of write operations going on.
> 
> 
> 
> On Fri, Dec 6, 2013 at 10:06 PM, srmore <comom...@gmail.com> wrote:
> 
> 
> 
> On Fri, Dec 6, 2013 at 9:59 AM, Vicky Kak <vicky....@gmail.com> wrote:
> You have passed the JVM configurations and not the cassandra configurations 
> which is in cassandra.yaml.
> 
> Apologies, was tuning JVM and that's what was in my mind. 
> Here are the cassandra settings http://pastebin.com/uN42GgYT
> 
>  
> The spikes are not that significant in our case and we are running the 
> cluster with 1.7 gb heap.
> 
> Are these spikes causing any issue at your end?
> 
> There are no big spikes, the overall performance seems to be about 40% low.
>  
> 
> 
> 
> 
> On Fri, Dec 6, 2013 at 9:10 PM, srmore <comom...@gmail.com> wrote:
> 
> 
> 
> On Fri, Dec 6, 2013 at 9:32 AM, Vicky Kak <vicky....@gmail.com> wrote:
> Hard to say much without knowing about the cassandra configurations.
>  
> The cassandra configuration is 
> -Xms8G
> -Xmx8G
> -Xmn800m
> -XX:+UseParNewGC
> -XX:+UseConcMarkSweepGC
> -XX:+CMSParallelRemarkEnabled
> -XX:SurvivorRatio=4
> -XX:MaxTenuringThreshold=2
> -XX:CMSInitiatingOccupancyFraction=75
> -XX:+UseCMSInitiatingOccupancyOnly
> 
>  
> Yes compactions/GC's could skipe the CPU, I had similar behavior with my 
> setup.
> 
> Were you able to get around it ?
>  
> 
> -VK
> 
> 
> On Fri, Dec 6, 2013 at 7:40 PM, srmore <comom...@gmail.com> wrote:
> We have a 3 node cluster running cassandra 1.2.12, they are pretty big 
> machines 64G ram with 16 cores, cassandra heap is 8G. 
> 
> The interesting observation is that, when I send traffic to one node its 
> performance is 2x more than when I send traffic to all the nodes. We ran 
> 1.0.11 on the same box and we observed a slight dip but not half as seen with 
> 1.2.12. In both the cases we were writing with LOCAL_QUORUM. Changing CL to 
> ONE make a slight improvement but not much.
> 
> The read_Repair_chance is 0.1. We see some compactions running.
> 
> following is my iostat -x output, sda is the ssd (for commit log) and sdb is 
> the spinner.
> 
> avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>           66.46    0.00    8.95    0.01    0.00   24.58
> 
> Device:         rrqm/s   wrqm/s   r/s   w/s   rsec/s   wsec/s avgrq-sz 
> avgqu-sz   await  svctm  %util
> sda               0.00    27.60  0.00  4.40     0.00   256.00    58.18     
> 0.01    2.55   1.32   0.58
> sda1              0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> sda2              0.00    27.60  0.00  4.40     0.00   256.00    58.18     
> 0.01    2.55   1.32   0.58
> sdb               0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> sdb1              0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> dm-0              0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> dm-1              0.00     0.00  0.00  0.60     0.00     4.80     8.00     
> 0.00    5.33   2.67   0.16
> dm-2              0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> dm-3              0.00     0.00  0.00 24.80     0.00   198.40     8.00     
> 0.24    9.80   0.13   0.32
> dm-4              0.00     0.00  0.00  6.60     0.00    52.80     8.00     
> 0.01    1.36   0.55   0.36
> dm-5              0.00     0.00  0.00  0.00     0.00     0.00     0.00     
> 0.00    0.00   0.00   0.00
> dm-6              0.00     0.00  0.00 24.80     0.00   198.40     8.00     
> 0.29   11.60   0.13   0.32
> 
> 
> 
> I can see I am cpu bound here but couldn't figure out exactly what is causing 
> it, is this caused by GC or Compaction ? I am thinking it is compaction, I 
> see a lot of context switches and interrupts in my vmstat output.
> 
> I don't see GC activity in the logs but see some compaction activity. Has 
> anyone seen this ? or know what can be done to free up the CPU.
> 
> Thanks,
> Sandeep
> 
> 
> 
> 
> 
> 
> 
> 

Reply via email to