Hi Stephane,

I mean running spark sql job concurrently via %spark.sql just by
setting zeppelin.spark.concurrentSQL
to be true.

See the details here
http://zeppelin.apache.org/docs/0.9.0-preview1/interpreter/spark.html#sparksql

<stephane.d...@orange.com> 于2020年7月22日周三 上午12:21写道:

> Hi Jeff,
>
>
>
> * You can also run multiple spark sql jobs concurrently in one spark app*
>
>
>
> Can you please elaborate on this? What I see (with Zeppelin 0.8) is that
> with shared interpreter, each job is ran one after one. When going to one
> interpreter per user, many users can run a job at the same time, but each
> user can run only one job at one time. How is it possible to run multiple
> sql jobs concurrently in one spark app?
>
>
>
> Thanks,
>
>
>
> Stéphane
>
>
>
>
>
> *From:* Jeff Zhang [mailto:zjf...@gmail.com]
> *Sent:* Tuesday, July 21, 2020 17:54
> *To:* users
> *Subject:* Re: Monitoring a Notebook in Spark UI
>
>
>
> Regarding how many spark apps, it depends on the interpreter binding mode,
> you can refer to this document.
> http://zeppelin.apache.org/docs/0.9.0-preview1/usage/interpreter/interpreter_binding_mode.html
>
> Internally, each spark app run a scala shell to execute scala code and
> python shell to execute pyspark code.
>
>
>
> Regarding the interpreter concurrency,  it depends on how you define
> interpreter concurrency, you can run each spark app for each user or each
> note, that depends on the interpreter binding mode I refer above. You can
> also run multiple spark sql jobs concurrently in one spark app
>
>
>
> Joshua Conlin <conlin.jos...@gmail.com> 于2020年7月21日周二 下午11:00写道:
>
> Hello,
>
>
>
> I'm looking for documentation to better understand pyspark/scala notebook
> execution in Spark.  I typically see application runtimes that can be very
> long, is there always a spark "application" running for a notebook or
> zeppelin session?  Those that are not actually being run in zeppelin
> typically have very low resource utilization.  Are these applications in
> spark tied to the zeppelin user's session?
>
>
>
> Also, how can I find out more about hive, pyspark and scala interpreter
> concurrency?  How many users/notebooks/paragraphs can execute these
> interpreters concurrently and how is this tunable?
>
>
>
> Any insight you can provide would be appreciated.
>
>
>
> Thanks,
>
>
>
> Josh
>
>
>
>
> --
>
> Best Regards
>
> Jeff Zhang
>


-- 
Best Regards

Jeff Zhang

Reply via email to