Thanks! I'll try. Best regards, Minwoo Kang
________________________________________ 보낸 사람: Jeff Zhang <zjf...@gmail.com> 보낸 날짜: 2017년 9월 25일 월요일 오후 8:17:29 받는 사람: users@zeppelin.apache.org 제목: Re: unread block data Could you try to run your code in spark shell first ? Kang Minwoo <minwoo.k...@outlook.com<mailto:minwoo.k...@outlook.com>>于2017年9月25日周一 下午7:04写道: Helle, Users I think it is not a zeppelin issue. But I ask for your help. I am using zeppelin version 0.7.1, Spark 2.0.2, Hadoop 2.7.1 When I start zeppelin notebook it includes spark code, I got an error. The error log is below. --- Caused by: java.lang.IllegalStateException: unread block data at java.io.ObjectInputStream$BlockDataInputStream.setBlockDataMode(ObjectInputStream.java:2431) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1383) at java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:2000) at java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1924) at java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1801) at java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1351) at java.io.ObjectInputStream.readObject(ObjectInputStream.java:371) at org.apache.spark.serializer.JavaDeserializationStream.readObject(JavaSerializer.scala:75) at org.apache.spark.serializer.JavaSerializerInstance.deserialize(JavaSerializer.scala:114) at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:298) ... 3 more --- Does anyone of you have the same problem with me? I checked master and slave JDK version, spark version... Best regards, Minwoo Kang