Could you try to run your code in spark shell first ?

Kang Minwoo <minwoo.k...@outlook.com>于2017年9月25日周一 下午7:04写道:

> Helle, Users
>
> I think it is not a zeppelin issue.
> But I ask for your help.
>
> I am using zeppelin version 0.7.1, Spark 2.0.2, Hadoop 2.7.1
>
> When I start zeppelin notebook it includes spark code, I got an error.
> The error log is below.
>
> ---
>
> Caused by: java.lang.IllegalStateException: unread block data
>   at
> java.io.ObjectInputStream$BlockDataInputStream.setBlockDataMode(ObjectInputStream.java:2431)
>   at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1383)
>   at
> java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000)
>   at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924)
>   at
> java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801)
>   at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351)
>   at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371)
>   at
> org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:75)
>   at
> org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:114)
>   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:298)
>   ... 3 more
>
> ---
>
> Does anyone of you have the same problem with me?
>
> I checked master and slave JDK version, spark version...
>
> Best regards,
> Minwoo Kang
>

Reply via email to