It looks like one of our big problems is that zeppelin doesn’t always kill all
completed processes. Is there an accepted way to kill a spark instance? Most of
the time executing sys.exit in a paragraph will kill the spark instance in yarn
and I believe also kill the corresponding zeppelin proces
Running in a VM we’ve noticed Zeppelin consume lot of memory and have
encountered out-of-memory and GC issues - with a couple of users.
In our case, I attributed it to the use-case: Spark interpreter connecting to a
Postgres DB to load few tables into data frames which had large number of rows