I suspect this is due to not setting SPARK_EXECUTOR_URI. I’ve run Zeppelin with Spark on Mesos. I ran into a similar exception where Zeppelin was not able to parse the MASTER URL, which is “mesos://leader.mesos <mesos://leader.mesos>:5050” in my case. Then I found out that I had the following setting: SPARK_EXECUTOR_URI=https://www.apache.org/dist/spark/spark-2.1.0/spark-2.1.0-bin-hadoop2.6.tgz <https://www.apache.org/dist/spark/spark-2.1.0/spark-2.1.0-bin-hadoop2.6.tgz> which is not built for mesos.
After changing it to the following SPARK_EXECUTOR_URI=https://downloads.mesosphere.com/spark/assets/spark-2.1.0-bin-2.6.tgz <https://downloads.mesosphere.com/spark/assets/spark-2.1.0-bin-2.6.tgz> the exception was gone. In your case, you might want to look at this page: http://archive-primary.cloudera.com/cdh5/cdh/5/ <http://archive-primary.cloudera.com/cdh5/cdh/5/> So I guess something like http://archive-primary.cloudera.com/cdh5/cdh/5/spark-1.6.0-cdh5.7.6.tar.gz <http://archive-primary.cloudera.com/cdh5/cdh/5/spark-1.6.0-cdh5.7.6.tar.gz> should work as a value for SPARK_EXECUTOR_URI. -- Chaoran Yu > On Apr 12, 2017, at 4:16 PM, Ben Vogan <b...@shopkick.com> wrote: > > Hello all, > > I am trying to install Zeppelin 0.7.1 on my CDH 5.7 Cluster. I have been > following the instructions here: > > https://zeppelin.apache.org/docs/0.7.1/install/install.html > <https://zeppelin.apache.org/docs/0.7.1/install/install.html> > https://zeppelin.apache.org/docs/0.7.1/install/configuration.html > <https://zeppelin.apache.org/docs/0.7.1/install/configuration.html> > https://zeppelin.apache.org/docs/0.7.1/interpreter/spark.html > <https://zeppelin.apache.org/docs/0.7.1/interpreter/spark.html> > > I copied the zeppelin-env.sh.template into zeppelin-env.sh and made the > following changes: > export JAVA_HOME=/usr/java/latest > export MASTER=yarn-client > > export ZEPPELIN_LOG_DIR=/var/log/services/zeppelin > export ZEPPELIN_PID_DIR=/services/zeppelin/data > export ZEPPELIN_WAR_TEMPDIR=/services/zeppelin/data/jetty_tmp > export ZEPPELIN_NOTEBOOK_DIR=/services/zeppelin/data/notebooks > export ZEPPELIN_NOTEBOOK_PUBLIC=true > > export SPARK_HOME=/opt/cloudera/parcels/CDH/lib/spark > export HADOOP_CONF_DIR=/etc/spark/conf/yarn-conf > export PYSPARK_PYTHON=/usr/lib/python > > I then start Zeppelin and hit the UI in my browser and create a spark note: > > %spark > sqlContext.sql("select 1+1").collect().foreach(println) > > And I get this error: > > org.apache.spark.SparkException: Could not parse Master URL: 'yarn' > at > org.apache.spark.SparkContext$.org$apache$spark$SparkContext$$createTaskScheduler(SparkContext.scala:2746) > at org.apache.spark.SparkContext.<init>(SparkContext.scala:533) > at > org.apache.zeppelin.spark.SparkInterpreter.createSparkContext_1(SparkInterpreter.java:484) > at > org.apache.zeppelin.spark.SparkInterpreter.createSparkContext(SparkInterpreter.java:382) > at > org.apache.zeppelin.spark.SparkInterpreter.getSparkContext(SparkInterpreter.java:146) > at > org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:828) > at > org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70) > at > org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:483) > at org.apache.zeppelin.scheduler.Job.run(Job.java:175) > at > org.apache.zeppelin.scheduler.FIFOScheduler$1.run(FIFOScheduler.java:139) > at > java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) > at java.util.concurrent.FutureTask.run(FutureTask.java:266) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) > at > java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) > at > java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) > at > java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) > at java.lang.Thread.run(Thread.java:745) > > I specified "yarn-client" as indicated by the instructions so I'm not sure > where it is getting "yarn" from. In my spark-defaults.conf it > spark.master=yarn-client as well. > > Help would be greatly appreciated. > > Thanks, > -- > BENJAMIN VOGAN | Data Platform Team Lead > > <http://www.shopkick.com/> > <https://www.facebook.com/shopkick> <https://www.instagram.com/shopkick/> > <https://www.pinterest.com/shopkick/> <https://twitter.com/shopkickbiz> > <https://www.linkedin.com/company-beta/831240/?pathWildcard=831240>