Re: Cassandra/Spark failing to process large table

2018-03-03 Thread Ben Slater
Both CQLSH and the Spark Cassandra query at consistent level ONE (LOCAL_ONE for Spark connector) by default so if there is any inconsistency in your replicas this can resulting in inconsistent query results. See http://cassandra.apache.org/doc/latest/tools/cqlsh.html and https://github.com/datasta

Re: Cassandra/Spark failing to process large table

2018-03-03 Thread Kant Kodali
The fact that cqlsh itself gives different results tells me that this has nothing to do with spark. Moreover, spark results are monotonically increasing which seem to be more consistent than cqlsh. so I believe spark can be taken out of the equation. Now, while you are running these queries is th