Those latencies look like the difference between a couple of disk seeks and
reading something that's already in the os cache.

The dynamic snitch will favor nodes with lower latencies.  Once a node has
served enough reads, it might not have to hit disk very often, which
produces lower latencies.  So, if you have a hot dataset that fits into
memory, the dynamic snitch starts a positive feedback loop where most reads
will be served from one replica.

I'm guessing the node with the low latencies is serving most of your reads.
You can look at how quickly the total read count is increasing for each of
the replicas to confirm this.  It's not easy to do with only nodetool
cfstats, but something like OpsCenter would help.

On Fri, Jul 6, 2012 at 9:11 PM, Deno Vichas <d...@syncopated.net> wrote:

> all,
>
> what would explain a huge different (12ms vs 0.1ms) in read latency from
> node to node.  i've got a 4 node cluster w/ replication factor of 3 using
> hector.  i'm seeing these numbers with nodetool cfstats.
>
>
> thx,
> deno
>



-- 
Tyler Hobbs
DataStax <http://datastax.com/>

Reply via email to