Hello All, We are using DSE 4.6.6 & Cassandra 2.0.14.425.
I am facing this exception right now. We got this exception couple of times & repair jobs helped us temporarily. As the data is growing significantly we are experiencing this exception more than couple of times. Does any one have any thoughts on this ? Exception in thread "main" org.apache.spark.SparkException: Job aborted due to stage failure: Task 114 in stage 17.0 failed 4 times, most recent failure: Lost task 114.3 in stage 17.0 (TID 196, ): com.datastax.driver.core.exceptions.NoHostAvailableException: All host(s) tried for query failed (tried: [All IP addresses] - use getErrors() for details) com.datastax.driver.core.exceptions.NoHostAvailableException.copy( NoHostAvailableException.java:65) com.datastax.driver.core.DefaultResultSetFuture. extractCauseFromExecutionException(DefaultResultSetFuture.java:259) com.datastax.driver.core.ArrayBackedResultSet$ MultiPage.prepareNextRow(ArrayBackedResultSet.java:279) com.datastax.driver.core.ArrayBackedResultSet$MultiPage.isExhausted( ArrayBackedResultSet.java:239) com.datastax.driver.core.ArrayBackedResultSet$1. hasNext(ArrayBackedResultSet.java:122) com.datastax.spark.connector.rdd.reader. PrefetchingResultSetIterator.hasNext(PrefetchingResultSetIterator.scala:16) scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371) com.datastax.spark.connector.util.CountingIterator.hasNext( CountingIterator.scala:10) scala.collection.Iterator$$anon$14.hasNext(Iterator.scala:388) scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:327) org.apache.spark.storage.MemoryStore.unrollSafely( MemoryStore.scala:235) org.apache.spark.CacheManager.putInBlockManager( CacheManager.scala:163) org.apache.spark.CacheManager.getOrCompute(CacheManager.scala:70) org.apache.spark.rdd.RDD.iterator(RDD.scala:227) org.apache.spark.rdd.FilteredRDD.compute(FilteredRDD.scala:34) org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262) org.apache.spark.rdd.RDD.iterator(RDD.scala:229) org.apache.spark.rdd.FilteredRDD.compute(FilteredRDD.scala:34) org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:262) org.apache.spark.rdd.RDD.iterator(RDD.scala:229) org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:62) org.apache.spark.scheduler.Task.run(Task.scala:54) org.apache.spark.executor.Executor$TaskRunner.run( Executor.scala:177) java.util.concurrent.ThreadPoolExecutor.runWorker( ThreadPoolExecutor.java:1142) java.util.concurrent.ThreadPoolExecutor$Worker.run( ThreadPoolExecutor.java:617) java.lang.Thread.run(Thread.java:745) Thanks, Venkat.