What about ephemeral storage on ssd ? If performance is required it's
generally for production so the cluster would never be stopped. Then a
spark job to backup/restore on S3 allows to shut down completely the cluster

Le 3 déc. 2016 1:28 PM, "David Mitchell" <jdavidmitch...@gmail.com> a
écrit :

> To get a node local read from Spark to Cassandra, one has to use a read
> consistency level of LOCAL_ONE.  For some use cases, this is not an
> option.  For example, if you need to use a read consistency level
> of LOCAL_QUORUM, as many use cases demand, then one is not going to get a
> node local read.
>
> Also, to insure a node local read, one has to set spark.locality.wait to
> zero.  Whether or not a partition will be streamed to another node or
> computed locally is dependent on the spark.locality.wait parameters. This
> parameter can be set to 0 to force all partitions to only be computed on
> local nodes.
>
> If you do some testing, please post your performance numbers.
>
>
>

Reply via email to