Hello, I tried to execute a simple spark application using sparkSQL. At first 
try, it worked as I exepcted but after then, it doesn't run and shows an stderr 
like below:  Spark Executor Command: "java" "-cp" 
"::/opt/spark-1.0.2-bin-hadoop2/conf:/opt/spark-1.0.2-bin-hadoop2/lib/spark-assembly-1.0.2-hadoop2.4.0.jar:/opt/hadoop2/etc/hadoop:/opt/hadoop2/etc/hadoop"
 "-XX:MaxPermSize=128m" "-Xms14336M" "-Xmx14336M" 
"org.apache.spark.executor.CoarseGrainedExecutorBackend" 
"akka.tcp://spark@saturn00:35894/user/CoarseGrainedScheduler" "9" "saturn09" 
"4" "akka.tcp://sparkWorker@saturn09:45636/user/Worker" 
"app-20140908223656-0000"========================================
14/09/08 22:36:57 INFO spark.SecurityManager: Changing view acls to: 
root14/09/08 22:36:57 INFO spark.SecurityManager: SecurityManager: 
authentication disabled; ui acls disabled; users with view permissions: 
Set(root)14/09/08 22:36:57 INFO slf4j.Slf4jLogger: Slf4jLogger started14/09/08 
22:36:57 INFO Remoting: Starting remoting14/09/08 22:36:57 INFO Remoting: 
Remoting started; listening on addresses 
:[akka.tcp://sparkExecutor@saturn09:44260]14/09/08 22:36:57 INFO Remoting: 
Remoting now listens on addresses: 
[akka.tcp://sparkExecutor@saturn09:44260]14/09/08 22:36:57 INFO 
executor.CoarseGrainedExecutorBackend: Connecting to driver: 
akka.tcp://spark@saturn00:35894/user/CoarseGrainedScheduler14/09/08 22:36:57 
INFO worker.WorkerWatcher: Connecting to worker 
akka.tcp://sparkWorker@saturn09:45636/user/Worker14/09/08 22:36:57 INFO 
worker.WorkerWatcher: Successfully connected to 
akka.tcp://sparkWorker@saturn09:45636/user/Worker14/09/08 22:36:57 INFO 
executor.CoarseGrainedExecutorBackend: Succe
 ssfully registered with driver14/09/08 22:36:57 INFO spark.SecurityManager: 
Changing view acls to: root14/09/08 22:36:57 INFO spark.SecurityManager: 
SecurityManager: authentication disabled; ui acls disabled; users with view 
permissions: Set(root)14/09/08 22:36:58 INFO slf4j.Slf4jLogger: Slf4jLogger 
started14/09/08 22:36:58 INFO Remoting: Starting remoting14/09/08 22:36:58 INFO 
Remoting: Remoting started; listening on addresses 
:[akka.tcp://spark@saturn09:39880]14/09/08 22:36:58 INFO Remoting: Remoting now 
listens on addresses: [akka.tcp://spark@saturn09:39880]14/09/08 22:36:58 INFO 
spark.SparkEnv: Connecting to MapOutputTracker: 
akka.tcp://spark@saturn00:35894/user/MapOutputTracker14/09/08 22:36:58 INFO 
spark.SparkEnv: Connecting to BlockManagerMaster: 
akka.tcp://spark@saturn00:35894/user/BlockManagerMaster14/09/08 22:36:58 INFO 
storage.DiskBlockManager: Created local directory at 
/hadoop/spark/spark-local-20140908223658-569914/09/08 22:36:58 INFO 
storage.MemoryStore: MemoryStore s
 tarted with capacity 4.0 GB.14/09/08 22:36:58 INFO network.ConnectionManager: 
Bound socket to port 49090 with id = 
ConnectionManagerId(saturn09,49090)14/09/08 22:36:58 INFO 
storage.BlockManagerMaster: Trying to register BlockManager14/09/08 22:36:58 
INFO storage.BlockManagerMaster: Registered BlockManager14/09/08 22:36:58 INFO 
spark.HttpFileServer: HTTP File server directory is 
/tmp/spark-379704ff-05f2-4c93-8814-ffbe1cc8cd5314/09/08 22:36:58 INFO 
spark.HttpServer: Starting HTTP Server14/09/08 22:36:58 INFO server.Server: 
jetty-8.y.z-SNAPSHOT14/09/08 22:36:58 INFO server.AbstractConnector: Started 
SocketConnector@0.0.0.0:40257 [akka.tcp://spark@saturn00:35894] disassociated! 
Shutting down.
 Here, saturn00 is a master and there are 10 nodes in my cluster 
(saturn01~saturn10) At the last message of the error, what is the meaning of 
"Driver Disassociated?" How can I resolve this issue? Thanks 


// Yoonmin Nam


---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
For additional commands, e-mail: user-h...@spark.apache.org

Reply via email to