Hi, Victor, 

the issue for you to have different version in driver and cluster is that you 
the master will shutdown your application due to the inconsistent 
SerialVersionID in ExecutorState

Best, 

-- 
Nan Zhu


On Tuesday, August 26, 2014 at 10:10 PM, Matei Zaharia wrote:

> Things will definitely compile, and apps compiled on 1.0.0 should even be 
> able to link against 1.0.2 without recompiling. The only problem is if you 
> run your driver with 1.0.0 on its classpath, but the cluster has 1.0.2 in 
> executors.
> 
> For Mesos and YARN vs standalone, the difference is that they just have more 
> features, at the expense of more complicated setup. For example, they have 
> richer support for cross-application sharing (see 
> https://spark.apache.org/docs/latest/job-scheduling.html), and the ability to 
> run non-Spark applications on the same cluster.
> 
> Matei 
> 
> On August 26, 2014 at 6:53:33 PM, Victor Tso-Guillen (v...@paxata.com 
> (mailto:v...@paxata.com)) wrote:
> 
> > Yes, we are standalone right now. Do you have literature why one would want 
> > to consider Mesos or YARN for Spark deployments? 
> > 
> > Sounds like I should try upgrading my project and seeing if everything 
> > compiles without modification. Then I can connect to an existing 1.0.0 
> > cluster and see what what happens... 
> > 
> > Thanks, Matei :) 
> > 
> > 
> > On Tue, Aug 26, 2014 at 6:37 PM, Matei Zaharia <matei.zaha...@gmail.com 
> > (mailto:matei.zaha...@gmail.com)> wrote:
> > > Is this a standalone mode cluster? We don't currently make this 
> > > guarantee, though it will likely work in 1.0.0 to 1.0.2. The problem 
> > > though is that the standalone mode grabs the executors' version of Spark 
> > > code from what's installed on the cluster, while your driver might be 
> > > built against another version. On YARN and Mesos, you can more easily mix 
> > > different versions of Spark, since each application ships its own Spark 
> > > JAR (or references one from a URL), and this is used for both the driver 
> > > and executors. 
> > > 
> > > Matei 
> > > 
> > > 
> > > On August 26, 2014 at 6:10:57 PM, Victor Tso-Guillen (v...@paxata.com 
> > > (mailto:v...@paxata.com)) wrote:
> > > 
> > > > I wanted to make sure that there's full compatibility between minor 
> > > > releases. I have a project that has a dependency on spark-core so that 
> > > > it can be a driver program and that I can test locally. However, when 
> > > > connecting to a cluster you don't necessarily know what version you're 
> > > > connecting to. Is a 1.0.0 cluster binary compatible with a 1.0.2 driver 
> > > > program? Is a 1.0.0 driver program binary compatible with a 1.0.2 
> > > > cluster?
> > > > 
> > > > 
> > > > 
> > > 
> > > 
> > > 
> > > 
> > 
> > 
> > 

Reply via email to