Could you try to run your code in spark shell first ?
Kang Minwoo <minwoo.k...@outlook.com>于2017年9月25日周一 下午7:04写道: > Helle, Users > > I think it is not a zeppelin issue. > But I ask for your help. > > I am using zeppelin version 0.7.1, Spark 2.0.2, Hadoop 2.7.1 > > When I start zeppelin notebook it includes spark code, I got an error. > The error log is below. > > --- > > Caused by: java.lang.IllegalStateException: unread block data > at > java.io.ObjectInputStream$BlockDataInputStream.setBlockDataMode(ObjectInputStream.java:2431) > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1383) > at > java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000) > at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924) > at > java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) > at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) > at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371) > at > org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:75) > at > org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:114) > at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:298) > ... 3 more > > --- > > Does anyone of you have the same problem with me? > > I checked master and slave JDK version, spark version... > > Best regards, > Minwoo Kang >