Hi all,

Test scenario:

    4 nodes (.1, .2, .3, .4)
    RF=3
    CL=QUORUM
    1.1.2

I noticed that in ReadCallback's constructor, it determines the 'blockfor' number of 2 for RF=3, CL=QUORUM.

According to the API page on the wiki[1] for reads at CL=QUORUM:

   Will query *all* replicas and return the record with the most recent
   timestamp once it has at least a majority of replicas (N / 2 + 1)
   reported.


However, in ReadCallback's constructor, it determines blockfor to be 2, then calls filterEndpoints. filterEndpoints is given a list of the three replicas, but at the very end of the method, the endpoint list to only two replicas. Those two replicas are then used in StorageProxy to execute the read/digest calls. So it ends up as 2 nodes, not all three as stated on the wiki.

In my test case, I kill a node and then immediately issue a query for a key that has a replica on the downed node. For the live nodes in the system, it doesn't immediately know that the other node is down yet. Rather than contacting *all* nodes as the wiki states, the coordinator contacts only two -- one of which is the downed node. Since it blocks for two, one of which is down, the query times out. Attempting the read again produces the same effect, even when trying different nodes as coordinators. I end up retrying a few times until the failure detectors on the live nodes realize that the node is down.

So, the end result is that if a client attempts to read a row that has a replica on a newly downed node, it will timeout repeatedly until the ~30 seconds failure detector window has passed -- even though there are enough live replicas to satisfy the request. We basically have a scenario wherein a value is not retrievable for upwards of 30 seconds. The percentage of keys that exhibit this possibility shrinks as the ring grows, but it's still non-zero.

This doesn't seem right and I'm sure I'm missing something.

Thanks,
Kirk

[1] http://wiki.apache.org/cassandra/API

Reply via email to