dungnguyen created ZEPPELIN-3205: ------------------------------------ Summary: restarting interpreter setting in a notebook abort running jobs of other notebooks Key: ZEPPELIN-3205 URL: https://issues.apache.org/jira/browse/ZEPPELIN-3205 Project: Zeppelin Issue Type: Bug Reporter: dungnguyen
I'm aware that there is resolved issues https://issues.apache.org/jira/browse/ZEPPELIN-1770 But it's pretty simple to reproduce, I can configure spark or python interpreters in per-note isolated mode, and start a long running job in 2 notebooks #1 and #2. If I restart spark or python (depends on the type of running job) interpreter in notebook #1, the job in notebook #2 is aborted. It is worse for pyspark since not just the job is aborted, the pyspark python process of notebook #2 is also killed, but notebook #2 will be hanging afterward, the only way to fix is to restart notebook #2 I also found a related issue for python interpreter https://issues.apache.org/jira/browse/ZEPPELIN-3171 -- This message was sent by Atlassian JIRA (v7.6.3#76005)