Yes, if you can reproduce easily, please see if 0.8.3 fixes it for by
any chance.
Otherwise, please open a JIRA ticket with as much info on how to reproduce.

--
Sylvain

On Tue, Aug 9, 2011 at 11:04 AM, Andrii Denysenko <andr...@gmail.com> wrote:
> Try 0.8.3
> They fixed https://issues.apache.org/jira/browse/CASSANDRA-2968 - and this
> produced erroneous records for counters.
> Not sure this is exactly yours, but similar.
>
> On Tue, Aug 9, 2011 at 5:28 AM, Boris Yen <yulin...@gmail.com> wrote:
>>
>> Hi,
>> I am not sure if this is a bug or we use the counter the wrong way, but I
>> keep getting a enormous counter number in our deployment. After a few tries,
>> I am finally able to reproduce it. The following are the settings of my
>> development:
>> -----------------------------------------------------
>> I have two-node cluster with the following keyspace and column family
>> settings.
>> Cluster Information:
>>    Snitch: org.apache.cassandra.locator.SimpleSnitch
>>    Partitioner: org.apache.cassandra.dht.RandomPartitioner
>>    Schema versions:
>> 63fda700-c243-11e0-0000-2d03dcafebdf: [172.17.19.151, 172.17.19.152]
>> Keyspace: test:
>>   Replication Strategy:
>> org.apache.cassandra.locator.NetworkTopologyStrategy
>>   Durable Writes: true
>>     Options: [datacenter1:2]
>>   Column Families:
>>     ColumnFamily: testCounter (Super)
>>     "APP status information."
>>       Key Validation Class: org.apache.cassandra.db.marshal.BytesType
>>       Default column value validator:
>> org.apache.cassandra.db.marshal.CounterColumnType
>>       Columns sorted by:
>> org.apache.cassandra.db.marshal.BytesType/org.apache.cassandra.db.marshal.BytesType
>>       Row cache size / save period in seconds: 0.0/0
>>       Key cache size / save period in seconds: 200000.0/14400
>>       Memtable thresholds: 1.1578125/1440/247 (millions of ops/MB/minutes)
>>       GC grace seconds: 864000
>>       Compaction min/max thresholds: 4/32
>>       Read repair chance: 1.0
>>       Replicate on write: true
>>       Built indexes: []
>> Then, I use a test program based on hector to add a counter column
>> (testCounter[sc][column]) 1000 times. In the middle the adding process, I
>> intentional shut down the node 172.17.19.152. In addition to that, the test
>> program is smart enough to switch the consistency level from Quorum to One,
>> so that the following adding actions would not fail.
>> After all the adding actions are done, I start the cassandra
>> on 172.17.19.152, and I use cassandra-cli to check if the counter is correct
>> on both nodes, and I got a result 1001 which should be reasonable because
>> hector will retry once. However, when I shut down 172.17.19.151 and
>> after 172.17.19.152 is aware of 172.17.19.151 is down, I try to start the
>> cassandra on 172.17.19.151 again. Then, I check the counter again, this time
>> I got a result 481387 which is so wrong.
>> I was wondering if anyone could explain why this happens, is this a bug or
>> do I use the counter the wrong way?.
>> Regards
>> Boris
>
>
> --
> Regards,
> Andriy Denysenko
>
>

Reply via email to