awesome

Anwar AliKhan <anwaralikhan...@gmail.com> 于2020年6月23日周二 上午1:18写道:

> I found the cause of issue , kind of.
> I built spark from source so now all is good.
>
>
> On Sun, 21 Jun 2020, 23:49 Jeff Zhang, <zjf...@gmail.com> wrote:
>
>> JDK issue ?
>>
>> 17 more Caused by: java.lang.NullPointerException at
>> org.apache.commons.lang3.SystemUtils.isJavaVersionAtLeast(SystemUtils.java:1654)
>> at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:207)
>> at org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala)
>> ... 28 more
>>
>> Anwar AliKhan <anwaralikhan...@gmail.com> 于2020年6月22日周一 上午6:21写道:
>>
>>>
>>> The only change I am making is to set SPARK_HOME I have made the setting
>>> in config files bashed file . In the Zeppelin interpreter settings. I am
>>> trying to run scala files which comes Zeppelin so I can develop spark scala
>>> app. I keep the same same message. Any ideas  ?
>>>
>>>
>>> org.apache.zeppelin.interpreter.InterpreterException:
>>> org.apache.zeppelin.interpreter.InterpreterException: Fail to open
>>> SparkInterpreter at
>>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:76)
>>> at
>>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:668)
>>> at
>>> org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:577)
>>> at org.apache.zeppelin.scheduler.Job.run(Job.java:172) at
>>> org.apache.zeppelin.scheduler.AbstractScheduler.runJob(AbstractScheduler.java:130)
>>> at
>>> org.apache.zeppelin.scheduler.FIFOScheduler.lambda$runJobInScheduler$0(FIFOScheduler.java:39)
>>> at
>>> java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
>>> at
>>> java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
>>> at java.base/java.lang.Thread.run(Thread.java:834) Caused by:
>>> org.apache.zeppelin.interpreter.InterpreterException: Fail to open
>>> SparkInterpreter at
>>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:114)
>>> at
>>> org.apache.zeppelin.interpreter.LazyOpenInterpreter.open(LazyOpenInterpreter.java:70)
>>> ... 8 more Caused by: java.lang.reflect.InvocationTargetException at
>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native
>>> Method) at
>>> java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>>> at
>>> java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>>> at java.base/java.lang.reflect.Method.invoke(Method.java:566) at
>>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.spark2CreateContext(BaseSparkScalaInterpreter.scala:292)
>>> at
>>> org.apache.zeppelin.spark.BaseSparkScalaInterpreter.createSparkContext(BaseSparkScalaInterpreter.scala:223)
>>> at
>>> org.apache.zeppelin.spark.SparkScala212Interpreter.open(SparkScala212Interpreter.scala:90)
>>> at
>>> org.apache.zeppelin.spark.SparkInterpreter.open(SparkInterpreter.java:98)
>>> ... 9 more Caused by: java.lang.ExceptionInInitializerError at
>>> org.apache.spark.storage.BlockManagerMasterEndpoint.<init>(BlockManagerMasterEndpoint.scala:93)
>>> at org.apache.spark.SparkEnv$.$anonfun$create$9(SparkEnv.scala:370) at
>>> org.apache.spark.SparkEnv$.registerOrLookupEndpoint$1(SparkEnv.scala:311)
>>> at org.apache.spark.SparkEnv$.create(SparkEnv.scala:359) at
>>> org.apache.spark.SparkEnv$.createDriverEnv(SparkEnv.scala:189) at
>>> org.apache.spark.SparkContext.createSparkEnv(SparkContext.scala:267) at
>>> org.apache.spark.SparkContext.<init>(SparkContext.scala:442) at
>>> org.apache.spark.SparkContext$.getOrCreate(SparkContext.scala:2555) at
>>> org.apache.spark.sql.SparkSession$Builder.$anonfun$getOrCreate$1(SparkSession.scala:930)
>>> at scala.Option.getOrElse(Option.scala:189) at
>>> org.apache.spark.sql.SparkSession$Builder.getOrCreate(SparkSession.scala:921)
>>> ... 17 more Caused by: java.lang.NullPointerException at
>>> org.apache.commons.lang3.SystemUtils.isJavaVersionAtLeast(SystemUtils.java:1654)
>>> at org.apache.spark.storage.StorageUtils$.<init>(StorageUtils.scala:207) at
>>> org.apache.spark.storage.StorageUtils$.<clinit>(StorageUtils.scala) ... 28
>>> more
>>> <http://www.backbutton.co.uk/>
>>>
>>>
>>
>> --
>> Best Regards
>>
>> Jeff Zhang
>>
>

-- 
Best Regards

Jeff Zhang

Reply via email to