Crystal clear, we use a lot of counters and I am always happy to learn this
kind of things.

Thanks a lot.

Alain


2013/6/12 Sylvain Lebresne <sylv...@datastax.com>

>
> Is there something special of this kind regarding counters over multiDC ?
>>
>
> No. Counters behave exactly as other writes as far the consistency level
> is concerned.
> Technically, the counter write path is different from the normal write
> path in the sense that a counter write
> will be written to one replica first and then written to the rest of the
> replicas in a second time (with a local
> read on the first replica in between, which is why counter writes are
> slower than normal ones). But,
> outside of the obvious performance impact, this has no impact on the
> behavior observed from a
> client point of view. The consistency level has the exact same meaning in
> particular (though one
> small difference is that counters don't support CL.ANY).
>
> --
> Sylvain
>
>
>>
>> Thank you anyway Sylvain
>>
>>
>> 2013/6/12 Sylvain Lebresne <sylv...@datastax.com>
>>
>>> It is the normal behavior, but that's true of any update, not only of
>>> counters.
>>>
>>> The consistency level does *not* influence which replica are written to.
>>> Cassandra always write to all replicas. The consistency level only decides
>>> how replica acknowledgement are waited for.
>>>
>>> --
>>> Sylvain
>>>
>>>
>>> On Wed, Jun 12, 2013 at 4:56 AM, Alain RODRIGUEZ <arodr...@gmail.com>wrote:
>>>
>>>> "counter will replicate to all replicas during write regardless the
>>>> consistency level"
>>>>
>>>> I that the normal behavior or a bug ?
>>>>
>>>>
>>>> 2013/6/11 Daning Wang <dan...@netseer.com>
>>>>
>>>>> It is counter caused the problem. counter will replicate to all
>>>>> replicas during write regardless the consistency level.
>>>>>
>>>>> In our case. we don't need to sync the counter across the center. so
>>>>> moving counter to new keyspace and all the replica in one
>>>>> center solved problem.
>>>>>
>>>>> There is option replicate_on_write on table. If you turn that off for
>>>>> counter might have better performance. but you are on high risk to lose
>>>>> data and create inconsistency. I did not try this option.
>>>>>
>>>>> Daning
>>>>>
>>>>>
>>>>> On Sat, Jun 8, 2013 at 6:53 AM, srmore <comom...@gmail.com> wrote:
>>>>>
>>>>>> I am seeing the similar behavior, in my case I have 2 nodes in each
>>>>>> datacenter and one node always has high latency (equal to the latency
>>>>>> between the two datacenters). When one of the datacenters is shutdown the
>>>>>> latency drops.
>>>>>>
>>>>>> I am curious to know whether anyone else has these issues and if yes
>>>>>> how did to get around it.
>>>>>>
>>>>>> Thanks !
>>>>>>
>>>>>>
>>>>>> On Fri, Jun 7, 2013 at 11:49 PM, Daning Wang <dan...@netseer.com>wrote:
>>>>>>
>>>>>>> We have deployed multi-center but got performance issue. When the
>>>>>>> nodes on other center are up, the read response time from clients is 4 
>>>>>>> or 5
>>>>>>> times higher. when we take those nodes down, the response time becomes
>>>>>>> normal(compare to the time before we changed to multi-center).
>>>>>>>
>>>>>>> We have high volume on the cluster, the consistency level is one for
>>>>>>> read. so my understanding is most of traffic between data center should 
>>>>>>> be
>>>>>>> read repair. but seems that could not create much delay.
>>>>>>>
>>>>>>> What could cause the problem? how to debug this?
>>>>>>>
>>>>>>> Here is the keyspace,
>>>>>>>
>>>>>>> [default@dsat] describe dsat;
>>>>>>> Keyspace: dsat:
>>>>>>>   Replication Strategy:
>>>>>>> org.apache.cassandra.locator.NetworkTopologyStrategy
>>>>>>>   Durable Writes: true
>>>>>>>     Options: [dc2:1, dc1:3]
>>>>>>>   Column Families:
>>>>>>>     ColumnFamily: categorization_cache
>>>>>>>
>>>>>>>
>>>>>>> Ring
>>>>>>>
>>>>>>> Datacenter: dc1
>>>>>>> ===============
>>>>>>> Status=Up/Down
>>>>>>> |/ State=Normal/Leaving/Joining/Moving
>>>>>>> --  Address           Load       Tokens  Owns (effective)  Host ID
>>>>>>>                             Rack
>>>>>>> UN  xx.xx.xx..111       59.2 GB    256     37.5%
>>>>>>> 4d6ed8d6-870d-4963-8844-08268607757e  rac1
>>>>>>> DN  xx.xx.xx..121       99.63 GB   256     37.5%
>>>>>>> 9d0d56ce-baf6-4440-a233-ad6f1d564602  rac1
>>>>>>> UN  xx.xx.xx..120       66.32 GB   256     37.5%
>>>>>>> 0fd912fb-3187-462b-8c8a-7d223751b649  rac1
>>>>>>> UN  xx.xx.xx..118       63.61 GB   256     37.5%
>>>>>>> 3c6e6862-ab14-4a8c-9593-49631645349d  rac1
>>>>>>> UN  xx.xx.xx..117       68.16 GB   256     37.5%
>>>>>>> ee6cdf23-d5e4-4998-a2db-f6c0ce41035a  rac1
>>>>>>> UN  xx.xx.xx..116       32.41 GB   256     37.5%
>>>>>>> f783eeef-1c51-4f91-ab7c-a60669816770  rac1
>>>>>>> UN  xx.xx.xx..115       64.24 GB   256     37.5%
>>>>>>> e75105fb-b330-4f40-aa4f-8e6e11838e37  rac1
>>>>>>> UN  xx.xx.xx..112       61.32 GB   256     37.5%
>>>>>>> 2547ee54-88dd-4994-a1ad-d9ba367ed11f  rac1
>>>>>>> Datacenter: dc2
>>>>>>> ===============
>>>>>>> Status=Up/Down
>>>>>>> |/ State=Normal/Leaving/Joining/Moving
>>>>>>> --  Address           Load       Tokens  Owns (effective)  Host ID
>>>>>>>                             Rack
>>>>>>> DN  xx.xx.xx.199    58.39 GB   256     50.0%
>>>>>>> 6954754a-e9df-4b3c-aca7-146b938515d8  rac1
>>>>>>> DN  xx.xx.xx..61      33.79 GB   256     50.0%
>>>>>>> 91b8d510-966a-4f2d-a666-d7edbe986a1c  rac1
>>>>>>>
>>>>>>>
>>>>>>> Thank you in advance,
>>>>>>>
>>>>>>> Daning
>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to