Is there a sustained difference or did it settle back ? 
Could this have been compaction or repair or upgrade tables working ? 

Do the read / write counts available in nodetool cfstats show anything 
different ? 

Cheers
-----------------
Aaron Morton
Freelance Cassandra Developer
New Zealand

@aaronmorton
http://www.thelastpickle.com

On 19/12/2012, at 6:26 AM, Alain RODRIGUEZ <arodr...@gmail.com> wrote:

> Hi,
> 
> I am experimenting a strange issue in my C* 1.1.6, 3 node, RF=3 cluster.
> 
> root@ip-10-64-177-38:~# nodetool ring
> Note: Ownership information does not include topology, please specify a 
> keyspace.
> Address         DC          Rack        Status State   Load            Owns   
>              Token
>                                                                               
>              141784319550391032739561396922763706368
> 10.64.167.32    eu-west     1b          Up     Normal  178.37 GB       33.33% 
>              28356863910078203714492389662765613056
> 10.250.202.154  eu-west     1b          Up     Normal  174.93 GB       33.33% 
>              85070591730234615865843651857942052863
> 10.64.177.38    eu-west     1b          Up     Normal  167.13 GB       33.33% 
>              141784319550391032739561396922763706368
> 
> root@ip-10-64-177-38:~# nodetool ring cassa_teads
> Address         DC          Rack        Status State   Load            
> Effective-Ownership Token
>                                                                               
>              141784319550391032739561396922763706368
> 10.64.167.32    eu-west     1b          Up     Normal  178.37 GB       
> 100.00%             28356863910078203714492389662765613056
> 10.250.202.154  eu-west     1b          Up     Normal  174.93 GB       
> 100.00%             85070591730234615865843651857942052863
> 10.64.177.38    eu-west     1b          Up     Normal  167.13 GB       
> 100.00%             141784319550391032739561396922763706368
> 
> My cluster is well balanced, all the nodes have an identical configuration, 
> but yet I have a lot of disk reads on one of them as you can see in these 
> screenshots:
> 
> Datastax OpsCenter : 
> http://img4.imageshack.us/img4/2528/datastaxopscenterheighr.png
> or
> AWS console : http://img59.imageshack.us/img59/5223/ec2managementconsole.png
> 
> I have tried to see what is read from any nodes with "inotifywatch -r -t300 
> /raid0 > inotifywatch5min" and get the following result:
> 
> root@ip-10-64-177-38:~# cat inotifywatch5min
> total       access  close_nowrite  open      filename
> 234580   113280  60691             60609    
> /raid0/cassandra/data/cassa_teads/data_viewer/
> 56013      27108   14454            14451     
> /raid0/cassandra/data/cassa_teads/data_ip_viewer/
> 30748      14998   7884              7866      
> /raid0/cassandra/data/cassa_teads/algo_ad_newcapping/
> 301          147      76                  78         
> /raid0/cassandra/data/cassa_teads/data_transac/
> 191           95       48                  48         
> /raid0/cassandra/data/cassa_teads/data_cust_website_viewer/
> 6               0        3                    3           /raid0/cassandra/
> 2               0        1                    1           
> /raid0/cassandra/data/
> 2               0        1                    1           
> /raid0/cassandra/commitlog/
> 2               0        1                    1           
> /raid0/cassandra/saved_caches/
> 
> 
> root@ip-10-250-202-154:~# cat inotifywatch5min
> total       access  modify  close_write  close_nowrite  open  moved_from  
> moved_to  create  delete  filename
> 307378  115456  77706   12                57119             57035        
> /raid0/cassandra/data/cassa_teads/data_viewer/
> 55395    26878   0           0                 14259             14258       
> /raid0/cassandra/data/cassa_teads/data_ip_viewer/
> 36155    17653   0           0                 9256               9246        
> /raid0/cassandra/data/cassa_teads/algo_ad_newcapping/
> 7377      188      7153      6                 4                    11        
>    /raid0/cassandra/data/cassa_teads/data_action/
> 401        0         364        6                 4                    12     
>       /raid0/cassandra/data/cassa_teads/stats_ad_uv/
> 244        120      0           0                 62                   62     
>      /raid0/cassandra/data/cassa_teads/data_transac/
> 160        76        0           0                 42                   42    
>      /raid0/cassandra/data/cassa_teads/data_cust_website_viewer/
> 26          0         0           0                 13                   13   
>        /raid0/cassandra/data/cassa_teads/
> 12          0         2           2                 1                     3   
>         /raid0/cassandra/commitlog/
> 6            0         0           0                 3                     3  
>         /raid0/cassandra/
> 2            0         0           0                 1                     1  
>         /raid0/cassandra/data/
> 2            0         0           0                 1                     1  
>         /raid0/cassandra/saved_caches/
> 
> 
> root@ip-10-64-167-32:~# cat inotifywatch5min
> total       access  modify  close_write  close_nowrite  open    filename
> 218489   105379  0            0                  56615         56495  
> /raid0/cassandra/data/cassa_teads/data_viewer/
> 61553     29733   0             0                  15911        15909  
> /raid0/cassandra/data/cassa_teads/data_ip_viewer/
> 33262     16214   0             0                  8528          8520    
> /raid0/cassandra/data/cassa_teads/algo_ad_newcapping/
> 407         0         370          6                  4               12      
>   /raid0/cassandra/data/cassa_teads/stats_ad_uv/
> 280         134      0             0                  73              73      
>  /raid0/cassandra/data/cassa_teads/data_transac/
> 163          79       0             0                  42              42     
>   /raid0/cassandra/data/cassa_teads/data_cust_website_viewer/
> 17            0        3             3                  1                 4   
>      /raid0/cassandra/commitlog/
> 8             0        0              0                  4                 4  
>       /raid0/cassandra/data/cassa_teads/
> 6             0        0              0                  3                 3  
>       /raid0/cassandra/
> 2             0        0              0                  1                 1  
>       /raid0/cassandra/data/
> 2             0        0              0                  1                 1  
>       /raid0/cassandra/saved_caches/
> 
> I see no reason to have a md0 disk read throughput up to 8 time bigger on the 
> .38 node...
> 
> root@ip-10-64-167-32:~# iostat -mx 5 100
> 
> Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s avgrq-sz 
> avgqu-sz   await  svctm  %util
> xvdap1            0.00     0.40    0.20    0.40     0.00     0.00    18.67    
>  0.00    6.67   6.67   0.40
> xvdb              3.60     0.00  471.60    0.00    13.48     0.00    58.53    
> 33.31   69.58   1.79  84.20
> xvdc              3.40     0.00  474.40    0.00    13.76     0.00    59.42    
> 30.06   62.63   1.72  81.60
> xvdd              2.60     0.00  478.20    0.00    13.69     0.00    58.61    
> 28.74   61.74   1.76  84.40
> xvde              3.40     0.00  472.40    0.00    13.34     0.00    57.82    
> 23.08   47.96   1.74  82.40
> md0               0.00     0.00 1954.20    0.00    55.57     0.00    58.24    
>  0.00    0.00   0.00   0.00
> 
> avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>           11.95   11.88   11.54   48.33    0.20   16.10
> 
> Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s avgrq-sz 
> avgqu-sz   await  svctm  %util
> xvdap1            0.00     2.40    0.40    1.00     0.00     0.01    25.14    
>  0.15  110.00  55.71   7.80
> xvdb              3.40     0.00  481.60   63.20    13.94     2.59    62.13    
> 33.79   62.86   1.56  84.80
> xvdc              3.60     0.00  469.20   64.40    13.43     2.68    61.81    
> 40.35   76.23   1.54  82.40
> xvdd              6.60     0.20  480.80   64.40    13.93     2.65    62.28    
> 45.99   84.72   1.63  88.60
> xvde              6.60     0.00  482.00   65.60    14.20     2.68    63.14    
> 48.76   89.77   1.61  88.00
> md0               0.00     0.00 1910.80  257.80    54.92    10.60    61.88    
>  0.00    0.00   0.00   0.00
> 
> avg-cpu:  %user   %nice %system %iowait  %steal   %idle
>           12.41   12.82   10.93   36.15    0.21   27.48
> 
> Device:         rrqm/s   wrqm/s     r/s     w/s    rMB/s    wMB/s avgrq-sz 
> avgqu-sz   await  svctm  %util
> xvdap1            0.00     0.60    2.40    0.60     0.03     0.00    23.47    
>  0.18   58.67  24.67   7.40
> xvdb              3.80     0.00  429.20    4.80    12.41     0.20    59.48    
> 16.44   37.94   1.70  73.80
> xvdc              3.00     0.00  432.60    4.80    12.47     0.20    59.33    
> 20.74   47.44   1.71  75.00
> xvdd              2.20     0.00  439.60    4.80    12.56     0.20    58.80    
> 17.57   39.44   1.78  79.20
> xvde              2.00     0.00  449.20    3.60    12.84     0.15    58.73    
> 15.96   35.29   1.68  76.00
> md0               0.00     0.00 1744.40   18.00    49.63     0.75    58.55    
>  0.00    0.00   0.00   0.00
> 
> Did anyone experiments something similar ?
> 
> How can I dig further to see what is wrong ?
> 
> Any idea on what on C* can be responsible of this king of targeted reads ?
> 
> Alain

Reply via email to