Can you provide the full create keyspace statement ? > Yes – using NetworkTopologyStrategy mmm, maybe it thinks the other nodes are down.
Cheers ----------------- Aaron Morton Freelance Cassandra Consultant New Zealand @aaronmorton http://www.thelastpickle.com On 22/03/2013, at 6:42 AM, Dwight Smith <dwight.sm...@genesyslab.com> wrote: > Yes – using NetworkTopologyStrategy > > From: aaron morton [mailto:aa...@thelastpickle.com] > Sent: Thursday, March 21, 2013 10:22 AM > To: user@cassandra.apache.org > Subject: Re: Question regarding multi datacenter and LOCAL_QUORUM > > DEBUG [Thrift:1] 2013-03-19 00:00:53,313 ReadCallback.java (line 79) Blockfor > is 2; setting up requests to /xx.yy.zz.146,/xx.yy.zz.143,/xx.yy.zz.145 > DEBUG [Thrift:1] 2013-03-19 00:00:53,334 CassandraServer.java (line 306) > get_slice > DEBUG [Thrift:1] 2013-03-19 00:00:53,334 ReadCallback.java (line 79) Blockfor > is 2; setting up requests to /xx.yy.zz.146,/xx.yy.zz.143 > DEBUG [Thrift:1] 2013-03-19 00:00:53,366 CassandraServer.java (line 306) > get_slice > DEBUG [Thrift:1] 2013-03-19 00:00:53,367 ReadCallback.java (line 79) Blockfor > is 2; setting up requests to /xx.yy.zz.146,/xx.yy.zz.143,/xx.yy.zz.145 > This is Read Repair, as controlled by the read_repaur_chance and > dclocal_read_repair_chance CF settings, in action. > > "Blockfor" is how many nodes the read operation is going to wait for. When > the number of nodes in the request is more than blockfor it means Read Repair > is active, we are reading from all UP nodes and will repair any detected > differences in the background. Your read is waiting for 2 nodes to respond > only (including the one we ask for the data.) > > The odd thing here is that there are only 3 replicas nodes. Are you using the > Network Topology Strategy ? If so I would expect there to be 6 nodes in the > the request with RR, 3 in each DC. > > Cheers > > > ----------------- > Aaron Morton > Freelance Cassandra Consultant > New Zealand > > @aaronmorton > http://www.thelastpickle.com > > ----------------- > Aaron Morton > Freelance Cassandra Consultant > New Zealand > > @aaronmorton > http://www.thelastpickle.com > > On 21/03/2013, at 12:38 PM, Tyler Hobbs <ty...@datastax.com> wrote: > > > > On Wed, Mar 20, 2013 at 3:18 PM, Tycen Stafford <tstaff...@medio.com> wrote: > I don’t think that’s correct for a mult-dc ring, but you’ll want to hear a > final answer from someone more authoritative. I could easily be wrong. Try > using the built in token generating tool (token-generator) – I don’t seem to > have it on my hosts (1.1.6 also) so I can’t confirm. I used the > tokentoolv2.py tool (from > herehttp://www.datastax.com/docs/1.0/initialize/token_generation) and got the > following (which looks to me evenly spaced and not using offsets): > > tstafford@tycen-linux:Cassandra$ ./tokentoolv2.py 3 3 > { > "0": { > "0": 0, > "1": 56713727820156410577229101238628035242, > "2": 113427455640312821154458202477256070485 > }, > "1": { > "0": 28356863910078205288614550619314017621, > "1": 85070591730234615865843651857942052863, > "2": 141784319550391026443072753096570088106 > } > } > > For multi-DC clusters, the only requirement for a balanced cluster is that > all tokens within a DC must be balanced; you can basically treat each DC as a > separate ring (as long as your tokens don't line up exactly). So, either > using an offset for the second DC or evenly spacing all nodes is acceptable. > > -- > Tyler Hobbs > DataStax