Hi, I am trying to use spark to perform some basic text processing on news articles Recently I am facing issues on codes which ran perfectly well on the same data before
I am pasting the last few lines, including the exception message. I am using python. Can anybody suggest a remedy Thanks, Sunandan 14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 10 on beaker-3.news.cs.nyu.edu: remote Akka client disassociated 14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 10 from TaskSet 1.0 14/10/21 01:52:23 WARN TaskSetManager: Lost task 8.2 in stage 1.0 (TID 29, beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 11.2 in stage 1.0 (TID 31, beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 9.2 in stage 1.0 (TID 28, beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 10.2 in stage 1.0 (TID 30, beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 INFO DAGScheduler: Executor lost: 10 (epoch 12) 14/10/21 01:52:23 INFO BlockManagerMasterActor: Trying to remove executor 10 from BlockManagerMaster. 14/10/21 01:52:23 INFO BlockManagerMaster: Removed 10 successfully in removeExecutor 14/10/21 01:52:23 INFO AppClient$ClientActor: Executor updated: app-20141021015207-0000/10 is now EXITED (Command exited with code 53) 14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor app-20141021015207-0000/10 removed: Command exited with code 53 14/10/21 01:52:23 ERROR SparkDeploySchedulerBackend: Application has been killed. Reason: Master removed our application: FAILED 14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor 11 disconnected, so removing it 14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 11 on beaker-12.news.cs.nyu.edu: remote Akka client disassociated 14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 11 from TaskSet 1.0 14/10/21 01:52:23 WARN TaskSetManager: Lost task 7.2 in stage 1.0 (TID 32, beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 5.2 in stage 1.0 (TID 35, beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 6.2 in stage 1.0 (TID 34, beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 4.2 in stage 1.0 (TID 33, beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor 12 disconnected, so removing it 14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 12 on beaker-13.news.cs.nyu.edu: remote Akka client disassociated 14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 12 from TaskSet 1.0 14/10/21 01:52:23 WARN TaskSetManager: Lost task 15.2 in stage 1.0 (TID 38, beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 14.2 in stage 1.0 (TID 37, beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 13.2 in stage 1.0 (TID 36, beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 WARN TaskSetManager: Lost task 12.2 in stage 1.0 (TID 39, beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost) 14/10/21 01:52:23 INFO SparkUI: Stopped Spark web UI at http://beaker-2.news.cs.nyu.edu:4040 14/10/21 01:52:23 INFO DAGScheduler: Stopping DAGScheduler 14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Shutting down all executors 14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Asking each executor to shut down 14/10/21 01:52:24 INFO DAGScheduler: Failed to run saveAsTextFile at NativeMethodAccessorImpl.java: -2 Traceback (most recent call last): File "/home/sunandan/www/test.py", line 74, in <module> weekly.saveAsTextFile('part11') File "/home/sunandan/spark/spark-1.1.0/python/pyspark/rdd.py", line 1324, in saveAsTextFile keyed._jrdd.map(self.ctx._jvm.BytesToString()).saveAsTextFile(path) File "/home/sunandan/spark/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py", line 538, in __call__ File "/home/sunandan/spark/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/protocol.py", line 300, in get_return_value py4j.protocol.Py4JJavaError: An error occurred while calling o70.saveAsTextFile. : org.apache.spark.SparkException: Job cancelled because SparkContext was shut down at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:694) at org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:693) at scala.collection.mutable.HashSet.foreach(HashSet.scala:79) at org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:693) at org.apache.spark.scheduler.DAGSchedulerEventProcessActor.postStop(DAGScheduler.scala:1399) at akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:201) at akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:163) at akka.actor.ActorCell.terminate(ActorCell.scala:338) at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:431) at akka.actor.ActorCell.systemInvoke(ActorCell.scala:447) at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:262) at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:240) at akka.dispatch.Mailbox.run(Mailbox.scala:219) at akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386) at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260) at scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339) at scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979) at scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)