Hi,

I am trying to use spark to perform some basic text processing on news
articles
Recently I am facing issues on codes which ran perfectly well on the same
data before

I am pasting the last few lines, including the exception message. I am
using python.

Can anybody suggest a remedy

Thanks,
Sunandan



14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 10 on
beaker-3.news.cs.nyu.edu: remote Akka client disassociated
14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 10 from
TaskSet 1.0
14/10/21 01:52:23 WARN TaskSetManager: Lost task 8.2 in stage 1.0 (TID 29,
beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 11.2 in stage 1.0 (TID 31,
beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 9.2 in stage 1.0 (TID 28,
beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 10.2 in stage 1.0 (TID 30,
beaker-3.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 INFO DAGScheduler: Executor lost: 10 (epoch 12)
14/10/21 01:52:23 INFO BlockManagerMasterActor: Trying to remove executor
10 from BlockManagerMaster.
14/10/21 01:52:23 INFO BlockManagerMaster: Removed 10 successfully in
removeExecutor
14/10/21 01:52:23 INFO AppClient$ClientActor: Executor updated:
app-20141021015207-0000/10 is now EXITED (Command exited with code 53)
14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor
app-20141021015207-0000/10 removed: Command exited with code 53
14/10/21 01:52:23 ERROR SparkDeploySchedulerBackend: Application has been
killed. Reason: Master removed our application: FAILED
14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor 11
disconnected, so removing it
14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 11 on
beaker-12.news.cs.nyu.edu: remote Akka client disassociated
14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 11 from
TaskSet 1.0
14/10/21 01:52:23 WARN TaskSetManager: Lost task 7.2 in stage 1.0 (TID 32,
beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 5.2 in stage 1.0 (TID 35,
beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 6.2 in stage 1.0 (TID 34,
beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 4.2 in stage 1.0 (TID 33,
beaker-12.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Executor 12
disconnected, so removing it
14/10/21 01:52:23 ERROR TaskSchedulerImpl: Lost executor 12 on
beaker-13.news.cs.nyu.edu: remote Akka client disassociated
14/10/21 01:52:23 INFO TaskSetManager: Re-queueing tasks for 12 from
TaskSet 1.0
14/10/21 01:52:23 WARN TaskSetManager: Lost task 15.2 in stage 1.0 (TID 38,
beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 14.2 in stage 1.0 (TID 37,
beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 13.2 in stage 1.0 (TID 36,
beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 WARN TaskSetManager: Lost task 12.2 in stage 1.0 (TID 39,
beaker-13.news.cs.nyu.edu): ExecutorLostFailure (executor lost)
14/10/21 01:52:23 INFO SparkUI: Stopped Spark web UI at
http://beaker-2.news.cs.nyu.edu:4040
14/10/21 01:52:23 INFO DAGScheduler: Stopping DAGScheduler
14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Shutting down all
executors
14/10/21 01:52:23 INFO SparkDeploySchedulerBackend: Asking each executor to
shut down
14/10/21 01:52:24 INFO DAGScheduler: Failed to run saveAsTextFile at
NativeMethodAccessorImpl.java:
-2
Traceback (most recent call last):
  File "/home/sunandan/www/test.py", line 74, in <module>
    weekly.saveAsTextFile('part11')
  File "/home/sunandan/spark/spark-1.1.0/python/pyspark/rdd.py", line 1324,
in saveAsTextFile
    keyed._jrdd.map(self.ctx._jvm.BytesToString()).saveAsTextFile(path)
  File
"/home/sunandan/spark/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/java_gateway.py",
line 538, in __call__
  File
"/home/sunandan/spark/spark-1.1.0/python/lib/py4j-0.8.2.1-src.zip/py4j/protocol.py",
line 300, in get_return_value
py4j.protocol.Py4JJavaError: An error occurred while calling
o70.saveAsTextFile.
: org.apache.spark.SparkException: Job cancelled because SparkContext was
shut down
    at
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:694)
    at
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:693)
    at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
    at
org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:693)
    at
org.apache.spark.scheduler.DAGSchedulerEventProcessActor.postStop(DAGScheduler.scala:1399)
    at
akka.actor.dungeon.FaultHandling$class.akka$actor$dungeon$FaultHandling$$finishTerminate(FaultHandling.scala:201)
    at
akka.actor.dungeon.FaultHandling$class.terminate(FaultHandling.scala:163)
    at akka.actor.ActorCell.terminate(ActorCell.scala:338)
    at akka.actor.ActorCell.invokeAll$1(ActorCell.scala:431)
    at akka.actor.ActorCell.systemInvoke(ActorCell.scala:447)
    at akka.dispatch.Mailbox.processAllSystemMessages(Mailbox.scala:262)
    at akka.dispatch.Mailbox.processMailbox(Mailbox.scala:240)
    at akka.dispatch.Mailbox.run(Mailbox.scala:219)
    at
akka.dispatch.ForkJoinExecutorConfigurator$AkkaForkJoinTask.exec(AbstractDispatcher.scala:386)
    at scala.concurrent.forkjoin.ForkJoinTask.doExec(ForkJoinTask.java:260)
    at
scala.concurrent.forkjoin.ForkJoinPool$WorkQueue.runTask(ForkJoinPool.java:1339)
    at
scala.concurrent.forkjoin.ForkJoinPool.runWorker(ForkJoinPool.java:1979)
    at
scala.concurrent.forkjoin.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:107)

Reply via email to