Hi Jeff, Apologies for raising a somewhat unrelated issue:
Should the JDBC interpreter also be able to find a driver if the JAR file containing the driver is loaded using the dep interpreter (instead of defining it as an "artifact" at the bottom of the interpreter configuration page)? Thanks, Sanjay On Mon, May 28, 2018 at 7:54 AM, Jeff Zhang <zjf...@gmail.com> wrote: > And this seems a bug in the DepInterpreter, I have created ticket > https://issues.apache.org/jira/browse/ZEPPELIN-3506 > > > > Jeff Zhang <zjf...@gmail.com>于2018年5月28日周一 上午8:24写道: > > > > > You can use %spark.conf to add custom jars to spark interpreter. > > %spark.conf is more powerful that it could not customize jars and also > > other spark configurations. > > > > e.g. > > > > %spark.conf > > > > spark.jars /tmp/product.jar > > > > > > See the Generic ConfInterpeter Section of this article > > https://medium.com/@zjffdu/zeppelin-0-8-0-new-features-ea53e8810235 > > > > Yohana Khoury <yohana.kho...@gigaspaces.com>于2018年5月27日周日 下午10:22写道: > > > >> Hi, > >> > >> I am trying to run the following on Zeppelin 0.8.0-RC2, Spark 2.3: > >> > >> *%dep* > >> *z.load("/tmp/product.jar")* > >> > >> and then > >> > >> *%spark* > >> *import model.v1._* > >> > >> The result is: > >> > >> *<console>:25: error: not found: value model* > >> * import model.v1._* > >> * ^* > >> > >> > >> It appears that Zeppelin does not load the jar file into the spark > >> interpreter. > >> Are you dropping this option from Zeppelin 0.8.0 ? Is it going to be > >> fixed? > >> > >> It worked with Zeppelin 0.7.2. > >> > >> > >> Thanks! > >> > > >