Interesting Tobias, when you said "Instead we transferred the data to Apache Kudu", did you transfer all Cassandra data into Kudu from with a single migration and then tap into Kudo for aggregation or did you run data import every day/week/month from Cassandra into Kudu ?
>From my point of view, the difficulty is not to have a static set of data and run aggregation on it, there are a lot of alternatives out there. The difficulty is to be able to run analytics on a live/production/changing dataset with all the data movement & update that it implies. Regards On Thu, Jun 8, 2017 at 3:37 PM, Tobias Eriksson <tobias.eriks...@qvantel.com > wrote: > Hi > > Something to consider before moving to Apache Spark and Cassandra > > I have a background where we have tons of data in Cassandra, and we wanted > to use Apache Spark to run various jobs > > We loved what we could do with Spark, BUT…. > > We realized soon that we wanted to run multiple jobs in parallel > > Some jobs would take 30 minutes and some 45 seconds > > Spark is by default arranged so that it will take up all the resources > there is, this can be tweaked by using Mesos or Yarn > > But even with Mesos and Yarn we found it complicated to run multiple jobs > in parallel. > > So eventually we ended up throwing out Spark, > > Instead we transferred the data to Apache Kudu, and then we ran our > analysis on Kudu, and what a difference ! > > “my two cents!” > > -Tobias > > > > > > > > *From: *한 승호 <shha...@outlook.com> > *Date: *Thursday, 8 June 2017 at 10:25 > *To: *"user@cassandra.apache.org" <user@cassandra.apache.org> > *Subject: *Cassandra & Spark > > > > Hello, > > > > I am Seung-ho and I work as a Data Engineer in Korea. I need some advice. > > > > My company recently consider replacing RDMBS-based system with Cassandra > and Hadoop. > > The purpose of this system is to analyze Cadssandra and HDFS data with > Spark. > > > > It seems many user cases put emphasis on data locality, for instance, both > Cassandra and Spark executor should be on the same node. > > > > The thing is, my company's data analyst team wants to analyze > heterogeneous data source, Cassandra and HDFS, using Spark. > > So, I wonder what would be the best practices of using Cassandra and > Hadoop in such case. > > > > Plan A: Both HDFS and Cassandra with NodeManager(Spark Executor) on the > same node > > > > Plan B: Cassandra + Node Manager / HDFS + NodeManager in each node > separately but the same cluster > > > > > > Which would be better or correct, or would be a better way? > > > > I appreciate your advice in advance :) > > > > Best Regards, > > Seung-Ho Han > > > > > > Windows 10용 메일 <https://go.microsoft.com/fwlink/?LinkId=550986>에서 보냄 > > >