Re: Performance impact with ALLOW FILTERING clause.

2019-08-17 Thread Alex Ott
Spark connector doesn't do the "select * from table;" - it does reads by token ranges, reading the data (see https://github.com/datastax/spark-cassandra-connector/blob/master/spark-cassandra-connector/src/main/scala/com/datastax/spark/connector/rdd/partitioner/CassandraPartition.scala#L14) Jac

Consistent hashing vnodes and ring in cassandra

2019-08-17 Thread Inquistive allen
I am a newbie in cassandra. I have asked this question on various platforms but never got a satisfying answer. Hence thought of bringing up the topic here. Sorry for this might be a simple question. 1. I studied the paper on consistent hashing (which is being implemented in Cassandra) 2. Cassan

Re: Consistent hashing vnodes and ring in cassandra

2019-08-17 Thread Jeff Jirsa
> On Aug 17, 2019, at 10:53 AM, Inquistive allen wrote: > > I am a newbie in cassandra. I have asked this question on various platforms > but never got a satisfying answer. > Hence thought of bringing up the topic here. Sorry for this might be a simple > question. > > 1. I studied the paper

Re: Consistent hashing vnodes and ring in cassandra

2019-08-17 Thread Inquistive allen
Hello Jeff, Thanks for the response. I just got it right. One Last thing, when a read request comes in to the coordinator node, the partition key is hashed and a node is located where corresponding data is previously stored. How does the coordinator node locate the replica nodes for this row. The

Re: Consistent hashing vnodes and ring in cassandra

2019-08-17 Thread Jeff Jirsa
The has results in a token The replicas are typically the RF instances with tokens numerically larger than the hash So if a row hashes to token 1, and the instances are -100,0,100,200,300,400,500 the replica instances are 100,200,300, and cassandra considers them to be identical for nearly all

Re: Performance impact with ALLOW FILTERING clause.

2019-08-17 Thread Devopam Mittra
Hi Asad, Seems to me that your development team will need to remodel the tables sooner than later. This problem can't be left unattended for long once it starts hitting severely. The way Cassandra is, you may want to have them replicate the same table with different PK / structure to suitably embed