Do you have set of  up to date mvn compile prameters because I want to
build from source.

Also I need the build to be compatible with spark 3.



On Sun, 21 Jun 2020, 23:49 Jeff Zhang, <zjf...@gmail.com> wrote:

> JDK issue ?
>
> 17 more Caused by: java.lang.NullPointerException at
> org.apache.commons.lang3.SystemUtils.isJavaVersionAtLeast(SystemUtils.java:1654)
> at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:207)
> at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
> ... 28 more
>
> Anwar AliKhan <anwaralikhan...@gmail.com> 于2020年6月22日周一 上午6:21写道:
>
>>
>> The only change I am making is to set SPARK_HOME I have made the setting
>> in config files bashed file . In the Zeppelin interpreter settings. I am
>> trying to run scala files which comes Zeppelin so I can develop spark scala
>> app. I keep the same same message. Any ideas  ?
>>
>>
>> org.apache.zeppelin.interpreter.InterpreterException:
>> org.apache.zeppelin.interpreter.InterpreterException: Fail to open
>> SparkInterpreter at
>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:76)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
>> at
>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:577)
>> at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at
>> org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
>> at
>> org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
>> at
>> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>> at
>> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>> at java.base/java.lang.Thread.run(Thread.java:834) Caused by:
>> org.apache.zeppelin.interpreter.InterpreterException: Fail to open
>> SparkInterpreter at
>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:114)
>> at
>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
>> ... 8 more Caused by: java.lang.reflect.InvocationTargetException at
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>> Method) at
>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>> at
>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>> at java.base/java.lang.reflect.Method.invoke(Method.java:566) at
>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.spark2CreateContext(BaseSparkScalaInterpreter.scala:292)
>> at
>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.createSparkContext(BaseSparkScalaInterpreter.scala:223)
>> at
>> org.apache.zeppelin.spark.SparkScala212Interpreter.open(SparkScala212Interpreter.scala:90)
>> at
>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:98)
>> ... 9 more Caused by: java.lang.ExceptionInInitializerError at
>> org.apache.spark.storage.BlockManagerMasterEndpoint.<init>(BlockManagerMasterEndpoint.scala:93)
>> at org.apache.spark.SparkEnv$.$anonfun$create$9(SparkEnv.scala:370) at
>> org.apache.spark.SparkEnv$.registerOrLookupEndpoint$1(SparkEnv.scala:311)
>> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:359) at
>> org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:189) at
>> org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:267) at
>> org.apache.spark.SparkContext.<init>(SparkContext.scala:442) at
>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2555) at
>> org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$1(SparkSession.scala:930)
>> at scala.Option.getOrElse(Option.scala:189) at
>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921)
>> ... 17 more Caused by: java.lang.NullPointerException at
>> org.apache.commons.lang3.SystemUtils.isJavaVersionAtLeast(SystemUtils.java:1654)
>> at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:207) at
>> org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala) ... 28
>> more
>> <http://www.backbutton.co.uk/>
>>
>>
>
> --
> Best Regards
>
> Jeff Zhang
>

Reply via email to