Hi Keiji,

In the paragraph you would write:
%sh
spark-submit myapp.jar ...

The %sh interpreter is a shell, and runs as the zeppelin service user with 
whatever permissions it has. You can run any shell commands in it.

Although, this is a fairly strange way to run zeppelin so I’m not really sure 
that is what you want.

You can just use the %spark.pyspark interpreter and write your python spark 
code in there. The spark interpreters in Zeppelin already create the Spark 
Context for you, as well as sqlContext and spark session. These are available 
as sc, sqlContext and spark. If you have a program that is ready for spark 
submit, I would use some other tool to schedule and run it, like cron, oozie, 
NiFi, Luigi, Airflow etc. Or if you want to run manually just use spark submit 
from the shell directly or ssh.


Dave

From: 小野圭二<mailto:onoke...@gmail.com>
Sent: Tuesday, 3 October 2017 8:43 PM
To: users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>
Subject: Re: How to execute spark-submit on Note

Thank you for your quick reply again, Jeff.

Yes i know the difference of the execution environment between "%sh" and 
">spark-submit".
And my question was "how to execute spark-submit as shell interpreter".
That mean, i am searching how to execute a binary program from a note of 
zeppelin.
This time it has been limited on Spark.

Seems like Zeppelin have several procedure to execute Spark shell, like 
spark.pyspark, spark.sql....
So how to do "spark-submit" was my wondering.

I am sorry for bothering Your time, but at the same time, i am appreciated if 
You get my wondering clearly, and show me some tips.

-Keiji


2017-10-03 18:30 GMT+09:00 Jeff Zhang 
<zjf...@gmail.com<mailto:zjf...@gmail.com>>:
%sh is shell interpreter, you can run spark-submit just as you run it in shell 
terminal.

小野圭二 <onoke...@gmail.com<mailto:onoke...@gmail.com>>于2017年10月3日周二 下午4:58写道:
Thank you for your reply, Jeff

"%sh" ?
"sh" seems like request something execution code.
I tried "%sh", then

%sh <program name with full path>
      %sh bash: <program name>: no permission

I made binary file from .py to .pyc, but the answer was as same.
I am sorry seems like doubting you, but Is "%sh" the resolution?

-Keiji

2017-10-03 17:35 GMT+09:00 Jianfeng (Jeff) Zhang 
<jzh...@hortonworks.com<mailto:jzh...@hortonworks.com>>:

I am surprised why would you use %spark-submit, there’s no document about 
%spark-submit.   If you want to use spark-submit in zeppelin, then you could 
use %sh


Best Regard,
Jeff Zhang


From: 小野圭二 <onoke...@gmail.com<mailto:onoke...@gmail.com>>
Reply-To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Date: Tuesday, October 3, 2017 at 12:49 PM
To: "users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>" 
<users@zeppelin.apache.org<mailto:users@zeppelin.apache.org>>
Subject: How to execute spark-submit on Note

Hi all,

I searched this topic on the archive of ml, but still could not find out the 
solution clearly.
So i have tried to post this again(maybe).

I am using ver 0.8.0, and have installed spark 2.2 on the other path, just for 
checking my test program.
Then i wrote a quite simple sample python code to check the how to.

1. the code works fine on a note in Zeppelin
2. the same code but added the initialize code for SparkContext in it works 
fine on the Spark by using 'spark-submit'.
3. tried to execute "2" from a note in Zeppelin with the following script.
    yes, "spark" interpreter has been implemented in the note.
    then on the note,
        %spark-submit <program name with full path>
          -> interpreter not found error
4.I have arranged 'SPARK_SUBMIT_OPTIONS' in zeppelin-env.sh order by the doc
    ex. export SPARK_SUBMIT_OPTIONS='--packages 
com.databricks:spark-csv_2.10:1.2.0'
5. then running....
     %spark-submit <program name with full path>
      -> interpreter not found error  (as same as "3")

How can i use spark-submit from a note?
Any advice thanks.

-Keiji


Reply via email to