I'm not sure if it's significant, but on first notice the IP addresses all have the same octets in the ProperyFileSnitch, yet the EC2Snitch, all the octets are different.
Ergo: PropertyFileSnitch states that are all in the same data centre [168] and the same rac [2]. EC2Snitch states that all nodes in 3 different data centres [20, 73, 236]. I'm still new at this too and may not have the full answer as we are prepping our prod env with the PropertyFileSnitch 2DC's and 3 nodes per DC. Though our QA environment is configured much the same way only it's 3 nodes in a single DC: consistency: LOCAL_QUORUM strategy: NetworkTopologyStrategy strategy_options: datacenter1:3 Our distribution is 33% equally. Just reading the docs on the datastax website I'm starting to wonder how the PropertyFileSnitch distributes the data across the DC's: For NetworkTopologyStrategy, it specifies the number of replicas per data center in a comma separated list of datacenter_name:number_of_replicas. I'm wondering if you need to increase your replication factor to 3 to see the data replicate across the DC's Anthony On 17/09/2011, at 8:36 PM, Chris Marino wrote: > Anthony, We used the Ec2Snitch for one sets of runs, but for another set > we're using PropertyFileSnitch. > > With the PropertyFileSnitch we see: > > Address DC Rack Status State Load Owns > Token > > > 85070591730234615865843651857942052865 > > 192.168.2.1 us-east 1b Up Normal 60.59 MB 50.00% > 0 > > 192.168.2.6 us-west 1c Up Normal 26.5 MB 0.00% > 1 > > 192.168.2.2 us-east 1b Up Normal 29.86 MB 50.00% > 85070591730234615865843651857942052864 > > 192.168.2.7 us-west 1c Up Normal 60.63 MB 0.00% > 85070591730234615865843651857942052865 > > > > While with the EC2Snitch wwe see: > Address DC Rack Status State Load Owns > Token > > > 85070591730234615865843651857942052865 > > 107.20.68.176 us-east 1b Up Normal 59.95 MB 50.00% > 0 > > 204.236.179.193 us-west 1c Up Normal 53.67 MB 0.00% > 1 > > 184.73.133.171 us-east 1b Up Normal 60.65 MB 50.00% > 85070591730234615865843651857942052864 > > 204.236.166.4 us-west 1c Up Normal 26.33 MB 0.00% > 85070591730234615865843651857942052865 > > > > What also strange is that the Load on the nodes changes as well. For example, > node 204.236.166.4 sometimes is very low (~26KB), other times its closer to > 30MB. We see the same kind of variability in both clusters. > > > For both clusters, we're running stress tests with the following options: > > > --consistency-level=LOCAL_QUORUM --threads=4 > --replication-strategy=NetworkTopologyStrategy > --strategy-properties=us-east:2,us-west:2 --column-size=128 --keep-going > --num-keys=100000 -r > > Any clues to what is going on here are greatly appreciated. > > Thanks > CM > > On Sat, Sep 17, 2011 at 12:15 PM, Ikeda Anthony <anthony.ikeda....@gmail.com> > wrote: > What snitch do you have configured? We typically see a proper spread of data > across all our nodes equally. > > Anthony > > > On 17/09/2011, at 10:06 AM, Chris Marino wrote: > >> Hi, I have a question about what to expect when running a cluster across >> datacenters with Local Quorum consistency. >> >> My simplistic assumption is that the performance of an 8 node cluster split >> across 2 data centers and running with local quorum would perform roughly >> the same as a 4 node cluster in one data center. >> >> I'm 95% certain we've set up the keyspace so that the entire range is in one >> datacenter and the client is local. I see the keyspace split across all the >> local nodes, with remote nodes owning 0%. Yet when I run the stress tests >> against this configuration with local quorum, I see dramatically different >> results from when I ran the same tests against a 4 node cluster. I'm still >> 5% unsure of this because the documentation on how to configure this is >> pretty thin. >> >> My understanding of Local Quorum was that once the data was written to a >> local quorum, the commit would complete. I also believed that this would >> eliminate any WAN latency required for replication to the other DC. >> >> It not just that the split cluster runs slower, its also that there is >> enormous variability in identical tests. Sometimes by a factor of 2 or more. >> It seems as though the WAN latency is not only impacting performance, but >> that it's also introducing a wide variation on overally performance. >> >> Should WAN latency be completely hidden with local quorum? Or are there >> second order issues involved that will impact performance?? >> >> I'm running in EC2 across us-east/west regions. I already know how >> unpredictable EC2 performance can be, but what I'm seeing with here is far >> beyond normal.performance variability for EC2 >> >> Is there something obvious that I'm missing that would explain why the >> results are so different?? >> >> Here's the config when we run a 2x2 cluster: >> >> Address DC Rack Status State Load Owns >> Token >> >> 85070591730234615865843651857942052865 >> 192.168.2.1 us-east 1b Up Normal 25.26 MB >> 50.00% 0 >> 192.168.2.6 us-west 1c Up Normal 12.68 MB 0.00% >> 1 >> 192.168.2.2 us-east 1b Up Normal 12.56 MB >> 50.00% 85070591730234615865843651857942052864 >> 192.168.2.7 us-west 1c Up Normal 25.48 MB 0.00% >> 85070591730234615865843651857942052865 >> >> Thanks in advance. >> CM > >