Hi Keiji, In the paragraph you would write: %sh spark-submit myapp.jar ...
The %sh interpreter is a shell, and runs as the zeppelin service user with whatever permissions it has. You can run any shell commands in it. Although, this is a fairly strange way to run zeppelin so I’m not really sure that is what you want. You can just use the %spark.pyspark interpreter and write your python spark code in there. The spark interpreters in Zeppelin already create the Spark Context for you, as well as sqlContext and spark session. These are available as sc, sqlContext and spark. If you have a program that is ready for spark submit, I would use some other tool to schedule and run it, like cron, oozie, NiFi, Luigi, Airflow etc. Or if you want to run manually just use spark submit from the shell directly or ssh. Dave From: 小野圭二<mailto:onoke...@gmail.com> Sent: Tuesday, 3 October 2017 8:43 PM To: users@zeppelin.apache.org<mailto:users@zeppelin.apache.org> Subject: Re: How to execute spark-submit on Note Thank you for your quick reply again, Jeff. Yes i know the difference of the execution environment between "%sh" and ">spark-submit". And my question was "how to execute spark-submit as shell interpreter". That mean, i am searching how to execute a binary program from a note of zeppelin. This time it has been limited on Spark. Seems like Zeppelin have several procedure to execute Spark shell, like spark.pyspark, spark.sql.... So how to do "spark-submit" was my wondering. I am sorry for bothering Your time, but at the same time, i am appreciated if You get my wondering clearly, and show me some tips. -Keiji 2017-10-03 18:30 GMT+09:00 Jeff Zhang <zjf...@gmail.com<mailto:zjf...@gmail.com>>: %sh is shell interpreter, you can run spark-submit just as you run it in shell terminal. 小野圭二 <onoke...@gmail.com<mailto:onoke...@gmail.com>>于2017年10月3日周二 下午4:58写道: Thank you for your reply, Jeff "%sh" ? "sh" seems like request something execution code. I tried "%sh", then %sh <program name with full path> %sh bash: <program name>: no permission I made binary file from .py to .pyc, but the answer was as same. I am sorry seems like doubting you, but Is "%sh" the resolution? -Keiji 2017-10-03 17:35 GMT+09:00 Jianfeng (Jeff) Zhang <jzh...@hortonworks.com<mailto:jzh...@hortonworks.com>>: I am surprised why would you use %spark-submit, there’s no document about %spark-submit. If you want to use spark-submit in zeppelin, then you could use %sh Best Regard, Jeff Zhang From: 小野圭二 <onoke...@gmail.com<mailto:onoke...@gmail.com>> Reply-To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" <users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>> Date: Tuesday, October 3, 2017 at 12:49 PM To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" <users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>> Subject: How to execute spark-submit on Note Hi all, I searched this topic on the archive of ml, but still could not find out the solution clearly. So i have tried to post this again(maybe). I am using ver 0.8.0, and have installed spark 2.2 on the other path, just for checking my test program. Then i wrote a quite simple sample python code to check the how to. 1. the code works fine on a note in Zeppelin 2. the same code but added the initialize code for SparkContext in it works fine on the Spark by using 'spark-submit'. 3. tried to execute "2" from a note in Zeppelin with the following script. yes, "spark" interpreter has been implemented in the note. then on the note, %spark-submit <program name with full path> -> interpreter not found error 4.I have arranged 'SPARK_SUBMIT_OPTIONS' in zeppelin-env.sh order by the doc ex. export SPARK_SUBMIT_OPTIONS='--packages com.databricks:spark-csv_2.10:1.2.0' 5. then running.... %spark-submit <program name with full path> -> interpreter not found error (as same as "3") How can i use spark-submit from a note? Any advice thanks. -Keiji