>>> When reading, sometimes the data is there, >>> sometimes it is not, which we think is a replication issue, even >>> though we have left it plenty of time after the writes. Can you provide some more information on this ? Are you talking about writes to one DC and reads from another ?
Cheers ----------------- Aaron Morton Freelance Cassandra Developer New Zealand @aaronmorton http://www.thelastpickle.com On 4/12/2012, at 6:45 AM, Шамим <sre...@yandex.ru> wrote: > Yes, it's should be. dc1:3 means u have 3 copy of every piece of row, with > local quorum you always get a good consistency from 3 nodes. > First you have to calculate token for data center dc1 and add offset 100 to > token for the second data center which will resolve your problem. After > creating the keyspace you can run the nodetool command with ring KEYSPACE > NAME which should show u the load for every node as 33% > Hope it will help u > Shamim >> Hi Shamim >> >> I have read a bit about the Tokens. I understand how that could effect >> the data distribution at first, but I don't understand if we have >> specified Options: [dc1:3, dc2:3], surely after a while all the data >> will be on every server? >> >> Thanks, >> >> Owen >> >> On 3 December 2012 14:06, Шамим <sre...@yandex.ru> wrote: >> >>> Hello Owen, >>> Seems you did not configure token for all nodes correctly. See the section >>> Calculating Tokens for multiple data centers here >>> http://www.datastax.com/docs/0.8/install/cluster_init >>> >>> Best regards >>> Shamim >>> --- >>> On Mon, Dec 3, 2012 at 4:42 PM, Owen Davies <cassan...@obduk.com> wrote: >>> >>> We have a 2 data center test cassandra setup running, and are writing >>> to it using LOCAL_QUORUM. When reading, sometimes the data is there, >>> sometimes it is not, which we think is a replication issue, even >>> though we have left it plenty of time after the writes. >>> >>> We have the following setup: >>> >>> cassandra -v: 1.1.6 >>> >>> cassandra.yaml >>> ----------------------- >>> >>> cluster_name: something >>> >>> endpoint_snitch: PropertyFileSnitch >>> >>> cassandra-topology.properties >>> -------------------------------------------- >>> 192.168.1.1=dc1:rack1 >>> 192.168.1.2=dc1:rack1 >>> 192.168.1.3=dc1:rack1 >>> >>> 192.168.2.1=dc2:rack1 >>> 192.168.2.2=dc2:rack1 >>> 192.168.2.3=dc3:rack1 >>> >>> default=nodc:norack >>> >>> cassandra-cli >>> -------------------- >>> Keyspace: example: >>> Replication Strategy: org.apache.cassandra.locator.NetworkTopologyStrategy >>> Durable Writes: true >>> Options: [dc1:3, dc2:3] >>> >>> nodetool ring >>> ------------------- >>> >>> Address DC Rack Status State Load >>> Effective-Ownership Token >>> >>> 159447687142037741049740936276011715300 >>> 192.168.1.2 dc1 rack1 Up Normal 111.17 GB >>> 100.00% 67165620003619490909052924699950283577 >>> 192.168.1.1 dc1 rack1 Up Normal 204.57 GB >>> 100.00% 71045951808949151217931264995073558408 >>> 192.168.2.1 dc2 rack1 Up Normal 209.92 GB >>> 100.00% 107165019770579893816561717940612111506 >>> 192.168.1.3 dc1 rack1 Up Normal 209.92 GB >>> 100.00% 114165363957966360026729000065495595953 >>> 192.168.2.3 dc2 rack1 Up Normal 198.22 GB >>> 100.00% 147717787092318068320268200174271353451 >>> 192.168.2.2 dc2 rack1 Up Normal 179.31 GB >>> 100.00% 159447687142037741049740936276011715300 >>> >>> Does anyone have any ideas why every server does not have the same >>> amount of data on? >>> >>> Thanks, >>> >>> Owen > > > --