Github user echarles commented on the issue: https://github.com/apache/zeppelin/pull/2637 @matyix There is a long history in Zeppelin on `spark.dep` vs `external dependency in interpreter settings`. I am fan of the later (interpreter settings), so if the --packages flag can make it work, this would be wonderful. I don't see in Spark doc that --package add the jars on the executor classpath. The `spark.jars` property (Comma-separated list of local jars to include on the driver and executor classpaths) may be an alternative.
---