Hi Oleg. Spark can be configured to have high availability without the need for Mesos ( https://spark.apache.org/docs/latest/spark-standalone.html#high-availability), for instance using Zookeeper and standby masters. If I'm not wrong Storm doesn't need Mesos to work, so I imagine you use it to make nimbus fault tolerant, am I correct? In any case Mesos also deals with high availability (http://mesos.apache.org/documentation/latest/high-availability/) so I don't see the SPOF. What am I missing?
And I agree with DuyHai, have you tried Spark Streaming or similar? Perhaps it fits your needs. Paco 2014-09-10 20:20 GMT+02:00 Oleg Ruchovets <oruchov...@gmail.com>: > Interesting things actually: > We have hadoop in our eco system. It has single point of failure and I > am not sure about inter data center replication. > Plan is to use cassandra - no single point of failure , there is data > center replication. > For aggregation/transformation using SPARK. BUT storm requires mesos which > has SINGLE POINT of failure ( and it will require the same maintenance like > with secondary name node with hadoop) :-) :-). > > Question : is there a way to have storage and processing without single > point of failure and inter data center replication ? > > Thanks > Oleg. > > On Thu, Sep 11, 2014 at 2:09 AM, DuyHai Doan <doanduy...@gmail.com> wrote: > >> "As far as I know, the Datastax connector uses thrift to connect Spark >> with Cassandra although thrift is already deprecated, could someone confirm >> this point?" >> >> --> the Scala connector is using the latest Java driver, so no there is >> no Thrift there. >> >> For the Java version, I'm not sure, have not looked into it but I think >> it also uses the new Java driver >> >> >> On Wed, Sep 10, 2014 at 7:27 PM, Francisco Madrid-Salvador < >> pmad...@stratio.com> wrote: >> >>> Hi Oleg, >>> >>> Stratio Deep is just a library you must include in your Spark deployment >>> so it doesn't guarantee any high availability at all. To achieve HA you >>> must use Mesos or any other 3rd party resource manager. >>> >>> Stratio doesn't currently support PySpark, just Scala and Java. Perhaps >>> in the future... >>> >>> It should be ready for production use, but like always please test >>> before on a testing environment ;-) >>> >>> As far as I know, the Datastax connector uses thrift to connect Spark >>> with Cassandra although thrift is already deprecated, could someone confirm >>> this point? >>> >>> Paco >>> >> >> >