Regarding how many spark apps, it depends on the interpreter binding mode, you can refer to this document. http://zeppelin.apache.org/docs/0.9.0-preview1/usage/interpreter/interpreter_binding_mode.html Internally, each spark app run a scala shell to execute scala code and python shell to execute pyspark code.
Regarding the interpreter concurrency, it depends on how you define interpreter concurrency, you can run each spark app for each user or each note, that depends on the interpreter binding mode I refer above. You can also run multiple spark sql jobs concurrently in one spark app Joshua Conlin <conlin.jos...@gmail.com> 于2020年7月21日周二 下午11:00写道: > Hello, > > I'm looking for documentation to better understand pyspark/scala notebook > execution in Spark. I typically see application runtimes that can be very > long, is there always a spark "application" running for a notebook or > zeppelin session? Those that are not actually being run in zeppelin > typically have very low resource utilization. Are these applications in > spark tied to the zeppelin user's session? > > Also, how can I find out more about hive, pyspark and scala interpreter > concurrency? How many users/notebooks/paragraphs can execute these > interpreters concurrently and how is this tunable? > > Any insight you can provide would be appreciated. > > Thanks, > > Josh > -- Best Regards Jeff Zhang