Github user matyix commented on the issue: https://github.com/apache/zeppelin/pull/2637 @echarles Currently there are two ways to add external dependencies: add a new paragraph to the notebook using `spark.dep` interpreter and `z.load()`. This works because it downloads the dependencies inside the driver. There is an issue with this on Spark 2.2 and Scala 2.11.8 similar to https://issues.apache.org/jira/browse/ZEPPELIN-2475. Adding artifacts to Interpreter setting doesnât work in case of Spark `cluster` mode, since dependencies are downloaded locally. Maybe we can think about something like: in case of `deployMode=cluster` add artifact to --packages automatically and donât download it locally.
---