Hi Jeff, I'm building my custom zeppelin:
mvn clean package -DskipTests -Pweb-angular -Pspark-scala-2.12 -Pinclude-hadoop -Pspark-3.0 -Phadoop2 -Pbuild-distr -pl zeppelin-interpreter,zeppelin-zengine,spark/interpreter,spark/spark-dependencies,zeppelin-web-angular,zeppelin-server,zeppelin-distribution,jdbc,zeppelin-plugins/notebookrepo/filesystem -am Running the following code: %spark.pyspark import pandas as pd from pyspark.sql.types import StringType import numpy as np def np_sqrt(v): return np.__path__ spark.udf.register("np_sqrt", np_sqrt, StringType()) df = spark.range(10).createOrReplaceTempView("d") spark.sql("select np_sqrt(id) as arr from d").show(truncate=False) Thanks On Sun, Sep 12, 2021 at 6:09 PM Jeff Zhang <zjf...@gmail.com> wrote: > We have integration test for spark 3.1.2, it is supposed to work. Could > you let us know how to reproduce it ? > > Lior Chaga <lio...@taboola.com> 于2021年9月12日周日 下午8:17写道: > >> Hi, >> >> I'm running zeppelin 0.10 with spark 3.1.2 >> When trying to run %spark.pyspark paragraph, I >> get ClassNotFoundException: py4j.GatewayServer$GatewayServerBuilder >> >> I see that py4j is a provided scope dependency of spark interpreter, >> should I explicitly add py4j to the interpreter dependencies? >> Setting zeppelin.py4j.useAuth to false doesn't help, as the code >> uses sparkInterpreter.getSparkVersion().isSecretSocketSupported(), and >> doesn't allow manual control over it. >> >> Thanks, >> Lior >> > > > -- > Best Regards > > Jeff Zhang >