Just build with -Pmesos
http://spark.apache.org/docs/latest/building-spark.html#building-with-mesos-support

On Tue, Jan 10, 2017 at 8:56 AM, Olivier Girardot <
o.girar...@lateral-thoughts.com> wrote:

> I had the same problem, added spark-mesos as dependency and now I get :
> [2017-01-10 17:45:16,575] {bash_operator.py:77} INFO - Exception in thread
> "main" java.lang.NoClassDefFoundError: Could not initialize class
> org.apache.mesos.MesosSchedulerDriver
> [2017-01-10 17:45:16,576] {bash_operator.py:77} INFO - at
> org.apache.spark.scheduler.cluster.mesos.MesosSchedulerUtils$class.
> createSchedulerDriver(MesosSchedulerUtils.scala:105)
> [2017-01-10 17:45:16,576] {bash_operator.py:77} INFO - at
> org.apache.spark.scheduler.cluster.mesos.MesosCoarseGrainedSchedulerBac
> kend.createSchedulerDriver(MesosCoarseGrainedSchedulerBackend.scala:48)
> [2017-01-10 17:45:16,576] {bash_operator.py:77} INFO - at
> org.apache.spark.scheduler.cluster.mesos.MesosCoarseGrainedSchedulerBac
> kend.start(MesosCoarseGrainedSchedulerBackend.scala:155)
> [2017-01-10 17:45:16,577] {bash_operator.py:77} INFO - at
> org.apache.spark.scheduler.TaskSchedulerImpl.start(
> TaskSchedulerImpl.scala:156)
> [2017-01-10 17:45:16,577] {bash_operator.py:77} INFO - at
> org.apache.spark.SparkContext.<init>(SparkContext.scala:509)
> [2017-01-10 17:45:16,577] {bash_operator.py:77} INFO - at
> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2313)
> [2017-01-10 17:45:16,577] {bash_operator.py:77} INFO - at
> org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:868)
> [2017-01-10 17:45:16,577] {bash_operator.py:77} INFO - at
> org.apache.spark.sql.SparkSession$Builder$$anonfun$
> 6.apply(SparkSession.scala:860)
> [2017-01-10 17:45:16,578] {bash_operator.py:77} INFO - at
> scala.Option.getOrElse(Option.scala:121)
> [2017-01-10 17:45:16,578] {bash_operator.py:77} INFO - at
> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.
> scala:860)
>
> Is there any other dependency to add for spark 2.1.0 ?
>
>
>
> On Tue, Jan 10, 2017 1:26 AM, Abhishek Bhandari abhi10...@gmail.com wrote:
>
>> Glad that you found it.
>> ᐧ
>>
>> On Mon, Jan 9, 2017 at 3:29 PM, Richard Siebeling <rsiebel...@gmail.com>
>> wrote:
>>
>> Probably found it, it turns out that Mesos should be explicitly added
>> while building Spark, I assumed I could use the old build command that I
>> used for building Spark 2.0.0... Didn't see the two lines added in the
>> documentation...
>>
>> Maybe these kind of changes could be added in the changelog under changes
>> of behaviour or changes in the build process or something like that,
>>
>> kind regards,
>> Richard
>>
>>
>> On 9 January 2017 at 22:55, Richard Siebeling <rsiebel...@gmail.com>
>> wrote:
>>
>> Hi,
>>
>> I'm setting up Apache Spark 2.1.0 on Mesos and I am getting a "Could not
>> parse Master URL: 'mesos://xx.xx.xxx.xxx:5050'" error.
>> Mesos is running fine (both the master as the slave, it's a single
>> machine configuration).
>>
>> I really don't understand why this is happening since the same
>> configuration but using a Spark 2.0.0 is running fine within Vagrant.
>> Could someone please help?
>>
>> thanks in advance,
>> Richard
>>
>>
>>
>>
>>
>>
>>
>> --
>> *Abhishek J Bhandari*
>> Mobile No. +1 510 493 6205 <(510)%20493-6205> (USA)
>> Mobile No. +91 96387 93021 <+91%2096387%2093021> (IND)
>> *R & D Department*
>> *Valent Software Inc. CA*
>> Email: *abhis...@valent-software.com <abhis...@valent-software.com>*
>>
>
>
> *Olivier Girardot* | Associé
> o.girar...@lateral-thoughts.com
> +33 6 24 09 17 94
>



-- 
Michael Gummelt
Software Engineer
Mesosphere

Reply via email to