(Sorry if this mail is duplicate, but it seems that my previous mail could not reach the mailing list.)
Hi, When my spark program calls JavaSparkContext.stop(), the following errors occur. 14/12/11 16:24:19 INFO Main: sc.stop { 14/12/11 16:24:20 ERROR ConnectionManager: Corresponding SendingConnection to ConnectionManagerId(cluster02,38918) not found 14/12/11 16:24:20 ERROR SendingConnection: Exception while reading SendingConnection to ConnectionManagerId(cluster04,59659) java.nio.channels.ClosedChannelException at sun.nio.ch.SocketChannelImpl.ensureReadOpen(SocketChannelImpl.java:252) at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:295) at org.apache.spark.network.SendingConnection.read(Connection.scala:390) at org.apache.spark.network.ConnectionManager$$anon$6.run(ConnectionManager.sca la:205) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:11 45) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:6 15) at java.lang.Thread.run(Thread.java:745) 14/12/11 16:24:20 ERROR ConnectionManager: Corresponding SendingConnection to ConnectionManagerId(cluster03,59821) not found 14/12/11 16:24:20 ERROR ConnectionManager: Corresponding SendingConnection to ConnectionManagerId(cluster02,38918) not found 14/12/11 16:24:20 WARN ConnectionManager: All connections not cleaned up 14/12/11 16:24:20 INFO Main: sc.stop } How can I fix this? The configuration is as follows: - Spark version is 1.1.1 - Client runs on Windows 7 - The cluster is Linux(CentOS 6.5). - spark.master=yarn-client - Since Spark has a problem submitting job from Windows to Linux, I applied my patch to the Spark source code. (Please see https://github.com/apache/spark/pull/899 ) Spark 1.0.0 did not have this problem. Thanks.