My team is successfully running on Spark on MapR.

However, we add the mapr jars to the SPARK_CLASSAPTH on the workers, as
well as making sure they are on the classpath of the driver.

I'm not sure if we need every jar that we currently add but below is what
we currently use. The important file in the conf directory
is mapr-clusters.conf.

I do not think we add the jars to the classpath via addJars and prefer to
reference them from the mapr installation. This is partly because we
colocate our workers on our mapr worker nodes.

export
SPARK_CLASSPATH=/opt/mapr/hadoop/hadoop-0.20.2/conf:/opt/mapr/hadoop/hadoop-0.20.2/lib/hadoop-0.20.2-dev-core.jar:/opt/mapr/lib/maprfs-1.0.3-mapr-3.0.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/lib/commons-logging-1.0.4.jar:/opt/mapr/hadoop/hadoop-0.20.2/lib/maprfs-1.0.3-mapr-3.0.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/lib/zookeeper-3.3.6.jar




On Mon, May 26, 2014 at 10:50 AM, Mayur Rustagi <mayur.rust...@gmail.com>wrote:

> Did you try in standalone mode. You may not see serialization issues in
> local threaded mode.
> Serialization errors are unlikely to be cause of Mapr hadoop version.
> Regards
> Mayur
>
> Mayur Rustagi
> Ph: +1 (760) 203 3257
> http://www.sigmoidanalytics.com
> @mayur_rustagi <https://twitter.com/mayur_rustagi>
>
>
>
> On Mon, May 26, 2014 at 3:18 PM, nelson <nelson.verd...@ysance.com> wrote:
>
>> Hi all,
>>
>> I am trying to run spark over a MapR cluster. I successfully ran several
>> custom applications on a previous non-mapr hadoop cluster but i can't get
>> them working on the mapr one. To be more specific, i am not able to read
>> or
>> write on mfs without running into a serialization error from Java. Note
>> that
>> everything works fine when i am running the app in local mode which make
>> me
>> think of a dependancy error.
>>
>> The test application is built using sbt with the following dependency:
>>  - org.apache.spark spark-core 0.9.1
>>
>> In my test_app/lib directory i have:
>>  - hadoop-core-1.0.3-mapr-3.0.2.jar
>>  - json-20140107.jar
>>  - maprfs-1.0.3-mapr-3.0.2.jar
>>
>> Finally, i add those jars with conf.setJars so that they are distributed
>> on
>> the cluster.
>>
>> Am I compiling with the wrong dependencies? Should i get a "mapr version"
>> of
>> spark-core?
>>
>> Regards, Nelson
>>
>>
>>
>> --
>> View this message in context:
>> http://apache-spark-user-list.1001560.n3.nabble.com/maprfs-and-spark-libraries-tp6392.html
>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>
>
>


-- 

SUREN HIRAMAN, VP TECHNOLOGY
Velos
Accelerating Machine Learning

440 NINTH AVENUE, 11TH FLOOR
NEW YORK, NY 10001
O: (917) 525-2466 ext. 105
F: 646.349.4063
E: suren.hiraman@v <suren.hira...@sociocast.com>elos.io
W: www.velos.io

Reply via email to