Hi Team,

My Spark job fails with below error :

Could you please advice me what is the problem with my job.

Below is my error stack:

16/08/04 05:11:06 ERROR ActorSystemImpl: Uncaught fatal error from thread
[sparkDriver-akka.actor.default-dispatcher-14] shutting down ActorSystem
[sparkDriver]
java.lang.OutOfMemoryError: GC overhead limit exceeded
        at sun.reflect.ByteVectorImpl.trim(ByteVectorImpl.java:70)
        at
sun.reflect.MethodAccessorGenerator.generate(MethodAccessorGenerator.java:388)
        at
sun.reflect.MethodAccessorGenerator.generateMethod(MethodAccessorGenerator.java:77)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:46)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
java.io.ObjectStreamClass.invokeReadObject(ObjectStreamClass.java:1017)
        at
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1893)
        at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
        at
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
        at
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
        at
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
        at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
        at
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
        at
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
        at
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
        at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
        at
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
        at
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
        at
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
        at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
        at
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
        at
java.io.ObjectInputStream.defaultReadFields(ObjectInputStream.java:1990)
        at
java.io.ObjectInputStream.readSerialData(ObjectInputStream.java:1915)
        at
java.io.ObjectInputStream.readOrdinaryObject(ObjectInputStream.java:1798)
        at
java.io.ObjectInputStream.readObject0(ObjectInputStream.java:1350)
        at java.io.ObjectInputStream.readObject(ObjectInputStream.java:370)
        at
akka.serialization.JavaSerializer$$anonfun$1.apply(Serializer.scala:136)
        at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
        at
akka.serialization.JavaSerializer.fromBinary(Serializer.scala:136)
        at
akka.serialization.Serialization$$anonfun$deserialize$1.apply(Serialization.scala:104)
        at scala.util.Try$.apply(Try.scala:161)
16/08/04 05:11:06 INFO RemoteActorRefProvider$RemotingTerminator: Shutting
down remote daemon.
16/08/04 05:11:07 INFO RemoteActorRefProvider$RemotingTerminator: Remote
daemon shut down; proceeding with flushing remote transports.
16/08/04 05:11:07 INFO TaskSetManager: Finished task 18540.0 in stage 148.0
(TID 153058) in 190291 ms on lhrrhegapq005.enterprisenet.org (18536/32768)
16/08/04 05:11:07 INFO TaskSetManager: Finished task 18529.0 in stage 148.0
(TID 153044) in 190300 ms on lhrrhegapq008.enterprisenet.org (18537/32768)
16/08/04 05:11:07 INFO TaskSetManager: Finished task 18530.0 in stage 148.0
(TID 153049) in 190297 ms on lhrrhegapq005.enterprisenet.org (18538/32768)
16/08/04 05:11:07 INFO TaskSetManager: Finished task 18541.0 in stage 148.0
(TID 153062) in 190291 ms on lhrrhegapq006.enterprisenet.org (18539/32768)
16/08/04 05:11:09 INFO TaskSetManager: Finished task 18537.0 in stage 148.0
(TID 153057) in 191648 ms on lhrrhegapq003.enterprisenet.org (18540/32768)
16/08/04 05:11:10 INFO TaskSetManager: Finished task 18557.0 in stage 148.0
(TID 153073) in 193193 ms on lhrrhegapq003.enterprisenet.org (18541/32768)
16/08/04 05:11:10 INFO TaskSetManager: Finished task 18528.0 in stage 148.0
(TID 153045) in 193206 ms on lhrrhegapq007.enterprisenet.org (18542/32768)
16/08/04 05:11:10 INFO TaskSetManager: Finished task 18555.0 in stage 148.0
(TID 153072) in 193195 ms on lhrrhegapq002.enterprisenet.org (18543/32768)
16/08/04 05:11:10 ERROR YarnClientSchedulerBackend: Yarn application has
already exited with state FINISHED!
16/08/04 05:11:13 WARN QueuedThreadPool: 9 threads could not be stopped
16/08/04 05:11:13 INFO SparkUI: Stopped Spark web UI at
http://10.90.50.64:4043
16/08/04 05:11:15 INFO DAGScheduler: Stopping DAGScheduler
16/08/04 05:11:16 INFO DAGScheduler: Job 94 failed: save at
ndx_scala_util.scala:1264, took 232.788303 s
16/08/04 05:11:16 ERROR InsertIntoHadoopFsRelation: Aborting job.
org.apache.spark.SparkException: Job cancelled because SparkContext was
shut down
        at
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:703)
        at
org.apache.spark.scheduler.DAGScheduler$$anonfun$cleanUpAfterSchedulerStop$1.apply(DAGScheduler.scala:702)
        at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
        at
org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:702)
        at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onStop(DAGScheduler.scala:1525)
        at org.apache.spark.util.EventLoop.stop(EventLoop.scala:84)
        at
org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1449)
        at
org.apache.spark.SparkContext$$anonfun$stop$7.apply$mcV$sp(SparkContext.scala:1724)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1723)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
        at
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1824)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1837)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:1914)
        at
org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply$mcV$sp(InsertIntoHadoopFsRelation.scala:150)
        at
org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply(InsertIntoHadoopFsRelation.scala:108)
        at
org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation$$anonfun$run$1.apply(InsertIntoHadoopFsRelation.scala:108)
        at
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:56)
        at
org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelation.run(InsertIntoHadoopFsRelation.scala:108)
        at
org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult$lzycompute(commands.scala:57)
        at
org.apache.spark.sql.execution.ExecutedCommand.sideEffectResult(commands.scala:57)
        at
org.apache.spark.sql.execution.ExecutedCommand.doExecute(commands.scala:69)
        at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:140)
        at
org.apache.spark.sql.execution.SparkPlan$$anonfun$execute$5.apply(SparkPlan.scala:138)
        at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:147)
        at
org.apache.spark.sql.execution.SparkPlan.execute(SparkPlan.scala:138)
        at
org.apache.spark.sql.SQLContext$QueryExecution.toRdd$lzycompute(SQLContext.scala:933)
        at
org.apache.spark.sql.SQLContext$QueryExecution.toRdd(SQLContext.scala:933)
        at
org.apache.spark.sql.execution.datasources.ResolvedDataSource$.apply(ResolvedDataSource.scala:197)
        at
org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:146)
        at
org.apache.spark.sql.DataFrameWriter.save(DataFrameWriter.scala:137)
        at org.apache.spark.sql.DataFrame.save(DataFrame.scala:1808)
        at
com.acnielsen.madras.utils.ndx_scala_util$.newHiveTableData(ndx_scala_util.scala:1264)
        at
com.acnielsen.madras.utils.ndx_scala_util$.UPDATE(ndx_scala_util.scala:238)
        at
com.acnielsen.madras.pkgews_panel_extract$$anonfun$p_signed_rank_yago$1.apply(pkgews_panel_extract.scala:658)
        at
com.acnielsen.madras.pkgews_panel_extract$$anonfun$p_signed_rank_yago$1.apply(pkgews_panel_extract.scala:652)
        at
scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33)
        at
scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:108)
        at
com.acnielsen.madras.pkgews_panel_extract$.p_signed_rank_yago(pkgews_panel_extract.scala:652)
        at
com.acnielsen.madras.pkgews_panel_extract$.p_main(pkgews_panel_extract.scala:4844)
        at
com.acnielsen.madras.pkgews_panel_extract$.main(pkgews_panel_extract.scala:4655)
        at
com.acnielsen.madras.pkgews_panel_extract.main(pkgews_panel_extract.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
        at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
        at
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
16/08/04 05:11:17 INFO TaskSetManager: Finished task 18544.0 in stage 148.0
(TID 153064) in 200319 ms on lhrrhegapq005.enterprisenet.org (18544/32768)
16/08/04 05:11:17 INFO DAGScheduler: ShuffleMapStage 148 (save at
ndx_scala_util.scala:1264) failed in 233.531 s
16/08/04 05:11:17 INFO TaskSetManager: Finished task 18552.0 in stage 148.0
(TID 153067) in 200319 ms on lhrrhegapq008.enterprisenet.org (18545/32768)
16/08/04 05:11:17 INFO DAGScheduler: ShuffleMapStage 149 (save at
ndx_scala_util.scala:1264) failed in 233.277 s
16/08/04 05:11:17 INFO TaskSetManager: Finished task 18549.0 in stage 148.0
(TID 153056) in 200323 ms on lhrrhegapq002.enterprisenet.org (18546/32768)
16/08/04 05:11:17 INFO TaskSetManager: Finished task 18539.0 in stage 148.0
(TID 153061) in 200321 ms on lhrrhegapq006.enterprisenet.org (18547/32768)
16/08/04 05:11:17 INFO YarnClientSchedulerBackend: Shutting down all
executors
16/08/04 05:11:17 INFO Remoting: Remoting shut down
16/08/04 05:11:17 INFO RemoteActorRefProvider$RemotingTerminator: Remoting
shut down.
16/08/04 05:11:17 WARN AkkaRpcEndpointRef: Error sending message [message =
StopExecutors] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.. This timeout is controlled by
spark.rpc.askTimeout
        at org.apache.spark.rpc.RpcTimeout.org
$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcEnv.scala:214)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:229)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:225)
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
        at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:185)
        at scala.util.Try$.apply(Try.scala:161)
        at scala.util.Failure.recover(Try.scala:185)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
        at
org.spark-project.guava.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:293)
        at
scala.concurrent.impl.ExecutionContextImpl$$anon$1.execute(ExecutionContextImpl.scala:133)
        at
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
        at
scala.concurrent.impl.Promise$DefaultPromise.scala$concurrent$impl$Promise$DefaultPromise$$dispatchOrAddCallback(Promise.scala:280)
        at
scala.concurrent.impl.Promise$DefaultPromise.onComplete(Promise.scala:270)
        at scala.concurrent.Future$class.recover(Future.scala:324)
        at
scala.concurrent.impl.Promise$DefaultPromise.recover(Promise.scala:153)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:319)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:100)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:274)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:283)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.stop(YarnClientSchedulerBackend.scala:180)
        at
org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:439)
        at
org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1450)
        at
org.apache.spark.SparkContext$$anonfun$stop$7.apply$mcV$sp(SparkContext.scala:1724)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1723)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
Caused by: akka.pattern.AskTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.
        at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:307)
        ... 11 more
16/08/04 05:11:17 ERROR DefaultWriterContainer: Job job_201608040506_0000
aborted.
Failed to create dataset in HDFS location :
/user/hive/warehouse//TEWS_MS_TYPE_BASKET/1470301585850
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18533.0 in stage 148.0
(TID 153053) in 202457 ms on lhrrhegapq004.enterprisenet.org (18548/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18551.0 in stage 148.0
(TID 153070) in 202448 ms on lhrrhegapq002.enterprisenet.org (18549/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18553.0 in stage 148.0
(TID 153069) in 202449 ms on lhrrhegapq003.enterprisenet.org (18550/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18534.0 in stage 148.0
(TID 153047) in 202459 ms on lhrrhegapq007.enterprisenet.org (18551/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18554.0 in stage 148.0
(TID 153071) in 202958 ms on lhrrhegapq003.enterprisenet.org (18552/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18543.0 in stage 148.0
(TID 153065) in 202961 ms on lhrrhegapq004.enterprisenet.org (18553/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18538.0 in stage 148.0
(TID 153060) in 202963 ms on lhrrhegapq007.enterprisenet.org (18554/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18524.0 in stage 148.0
(TID 153048) in 202970 ms on lhrrhegapq005.enterprisenet.org (18555/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18584.0 in stage 148.0
(TID 153104) in 202166 ms on lhrrhegapq008.enterprisenet.org (18556/32768)
16/08/04 05:11:20 WARN AkkaRpcEndpointRef: Error sending message [message =
StopExecutors] in 2 attempts
org.apache.spark.rpc.RpcTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.. This timeout is controlled by
spark.rpc.askTimeout
        at org.apache.spark.rpc.RpcTimeout.org
$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcEnv.scala:214)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:229)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:225)
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
        at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:185)
        at scala.util.Try$.apply(Try.scala:161)
        at scala.util.Failure.recover(Try.scala:185)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
        at
org.spark-project.guava.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:293)
        at
scala.concurrent.impl.ExecutionContextImpl$$anon$1.execute(ExecutionContextImpl.scala:133)
        at
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
        at
scala.concurrent.impl.Promise$DefaultPromise.scala$concurrent$impl$Promise$DefaultPromise$$dispatchOrAddCallback(Promise.scala:280)
        at
scala.concurrent.impl.Promise$DefaultPromise.onComplete(Promise.scala:270)
        at scala.concurrent.Future$class.recover(Future.scala:324)
        at
scala.concurrent.impl.Promise$DefaultPromise.recover(Promise.scala:153)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:319)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:100)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:274)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:283)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.stop(YarnClientSchedulerBackend.scala:180)
        at
org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:439)
        at
org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1450)
        at
org.apache.spark.SparkContext$$anonfun$stop$7.apply$mcV$sp(SparkContext.scala:1724)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1723)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
Caused by: akka.pattern.AskTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.
        at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:307)
        ... 11 more
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18556.0 in stage 148.0
(TID 153074) in 203328 ms on lhrrhegapq008.enterprisenet.org (18557/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18571.0 in stage 148.0
(TID 153087) in 202531 ms on lhrrhegapq003.enterprisenet.org (18558/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18545.0 in stage 148.0
(TID 153066) in 203332 ms on lhrrhegapq006.enterprisenet.org (18559/32768)
16/08/04 05:11:20 INFO TaskSetManager: Finished task 18581.0 in stage 148.0
(TID 153100) in 202531 ms on lhrrhegapq004.enterprisenet.org (18560/32768)
Failed to fetch affected records
16/08/04 05:11:20 INFO pkgews_panel_extract$Log$: LOG,
PKGEWS_PANEL_EXTRACT, P_MAIN, CURRENT SCHEMA - mdr_devi1_supt_test, No of
records merged into TEWS_MS_TYPE_BASKET for msc_yago_trend > 03, LOG,
PKGEWS_PANEL_EXTRACT, p_signed_rank_yago,  Signed rank assignment for YAGO
Trend
16/08/04 05:11:21 INFO TaskSetManager: Finished task 18562.0 in stage 148.0
(TID 153081) in 203403 ms on lhrrhegapq004.enterprisenet.org (18561/32768)
16/08/04 05:11:21 INFO TaskSetManager: Finished task 18579.0 in stage 148.0
(TID 153084) in 203403 ms on lhrrhegapq008.enterprisenet.org (18562/32768)
16/08/04 05:11:21 INFO TaskSetManager: Finished task 18547.0 in stage 148.0
(TID 153077) in 203405 ms on lhrrhegapq007.enterprisenet.org (18563/32768)
16/08/04 05:11:21 INFO TaskSetManager: Finished task 18575.0 in stage 148.0
(TID 153098) in 203401 ms on lhrrhegapq002.enterprisenet.org (18564/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18559.0 in stage 148.0
(TID 153078) in 204359 ms on lhrrhegapq004.enterprisenet.org (18565/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18560.0 in stage 148.0
(TID 153082) in 204359 ms on lhrrhegapq006.enterprisenet.org (18566/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18583.0 in stage 148.0
(TID 153091) in 204357 ms on lhrrhegapq008.enterprisenet.org (18567/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18558.0 in stage 148.0
(TID 153075) in 204369 ms on lhrrhegapq008.enterprisenet.org (18568/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18591.0 in stage 148.0
(TID 153103) in 204358 ms on lhrrhegapq006.enterprisenet.org (18569/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18569.0 in stage 148.0
(TID 153089) in 204362 ms on lhrrhegapq002.enterprisenet.org (18570/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18561.0 in stage 148.0
(TID 153076) in 204364 ms on lhrrhegapq002.enterprisenet.org (18571/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18568.0 in stage 148.0
(TID 153088) in 204363 ms on lhrrhegapq007.enterprisenet.org (18572/32768)
16/08/04 05:11:22 INFO TaskSetManager: Finished task 18572.0 in stage 148.0
(TID 153096) in 204362 ms on lhrrhegapq004.enterprisenet.org (18573/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18574.0 in stage 148.0
(TID 153094) in 205321 ms on lhrrhegapq006.enterprisenet.org (18574/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18576.0 in stage 148.0
(TID 153090) in 205323 ms on lhrrhegapq003.enterprisenet.org (18575/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18567.0 in stage 148.0
(TID 153086) in 205324 ms on lhrrhegapq006.enterprisenet.org (18576/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18570.0 in stage 148.0
(TID 153092) in 205323 ms on lhrrhegapq002.enterprisenet.org (18577/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18582.0 in stage 148.0
(TID 153102) in 205324 ms on lhrrhegapq007.enterprisenet.org (18578/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18563.0 in stage 148.0
(TID 153079) in 205328 ms on lhrrhegapq002.enterprisenet.org (18579/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18578.0 in stage 148.0
(TID 153097) in 205325 ms on lhrrhegapq007.enterprisenet.org (18580/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18589.0 in stage 148.0
(TID 153101) in 205326 ms on lhrrhegapq005.enterprisenet.org (18581/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18566.0 in stage 148.0
(TID 153080) in 205330 ms on lhrrhegapq002.enterprisenet.org (18582/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18577.0 in stage 148.0
(TID 153095) in 205328 ms on lhrrhegapq003.enterprisenet.org (18583/32768)
16/08/04 05:11:23 INFO TaskSetManager: Finished task 18565.0 in stage 148.0
(TID 153085) in 205332 ms on lhrrhegapq005.enterprisenet.org (18584/32768)
16/08/04 05:11:24 WARN AkkaRpcEndpointRef: Error sending message [message =
StopExecutors] in 3 attempts
org.apache.spark.rpc.RpcTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.. This timeout is controlled by
spark.rpc.askTimeout
        at org.apache.spark.rpc.RpcTimeout.org
$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcEnv.scala:214)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:229)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:225)
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
        at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:185)
        at scala.util.Try$.apply(Try.scala:161)
        at scala.util.Failure.recover(Try.scala:185)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
        at
org.spark-project.guava.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:293)
        at
scala.concurrent.impl.ExecutionContextImpl$$anon$1.execute(ExecutionContextImpl.scala:133)
        at
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
        at
scala.concurrent.impl.Promise$DefaultPromise.scala$concurrent$impl$Promise$DefaultPromise$$dispatchOrAddCallback(Promise.scala:280)
        at
scala.concurrent.impl.Promise$DefaultPromise.onComplete(Promise.scala:270)
        at scala.concurrent.Future$class.recover(Future.scala:324)
        at
scala.concurrent.impl.Promise$DefaultPromise.recover(Promise.scala:153)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:319)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:100)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:274)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:283)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.stop(YarnClientSchedulerBackend.scala:180)
        at
org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:439)
        at
org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1450)
        at
org.apache.spark.SparkContext$$anonfun$stop$7.apply$mcV$sp(SparkContext.scala:1724)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1723)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
Caused by: akka.pattern.AskTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.
        at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:307)
        ... 11 more
16/08/04 05:11:24 INFO TaskSetManager: Finished task 18564.0 in stage 148.0
(TID 153083) in 206426 ms on lhrrhegapq005.enterprisenet.org (18585/32768)
16/08/04 05:11:24 ERROR Utils: Uncaught exception in thread Yarn
application state monitor
org.apache.spark.SparkException: Error asking standalone scheduler to shut
down executors
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:278)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stop(CoarseGrainedSchedulerBackend.scala:283)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.stop(YarnClientSchedulerBackend.scala:180)
        at
org.apache.spark.scheduler.TaskSchedulerImpl.stop(TaskSchedulerImpl.scala:439)
        at
org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:1450)
        at
org.apache.spark.SparkContext$$anonfun$stop$7.apply$mcV$sp(SparkContext.scala:1724)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1723)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
Caused by: org.apache.spark.SparkException: Error sending message [message
= StopExecutors]
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:118)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
        at
org.apache.spark.scheduler.cluster.CoarseGrainedSchedulerBackend.stopExecutors(CoarseGrainedSchedulerBackend.scala:274)
        ... 8 more
Caused by: org.apache.spark.rpc.RpcTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.. This timeout is controlled by
spark.rpc.askTimeout
        at org.apache.spark.rpc.RpcTimeout.org
$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcEnv.scala:214)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:229)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:225)
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
        at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:185)
        at scala.util.Try$.apply(Try.scala:161)
        at scala.util.Failure.recover(Try.scala:185)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
        at
org.spark-project.guava.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:293)
        at
scala.concurrent.impl.ExecutionContextImpl$$anon$1.execute(ExecutionContextImpl.scala:133)
        at
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
        at
scala.concurrent.impl.Promise$DefaultPromise.scala$concurrent$impl$Promise$DefaultPromise$$dispatchOrAddCallback(Promise.scala:280)
        at
scala.concurrent.impl.Promise$DefaultPromise.onComplete(Promise.scala:270)
        at scala.concurrent.Future$class.recover(Future.scala:324)
        at
scala.concurrent.impl.Promise$DefaultPromise.recover(Promise.scala:153)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:319)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:100)
        ... 10 more
Caused by: akka.pattern.AskTimeoutException:
Recipient[Actor[akka://sparkDriver/user/CoarseGrainedScheduler#-1206596405]]
had already been terminated.
        at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:307)
        ... 11 more
16/08/04 05:11:24 INFO TaskSetManager: Finished task 18573.0 in stage 148.0
(TID 153093) in 206426 ms on lhrrhegapq005.enterprisenet.org (18586/32768)
16/08/04 05:11:24 INFO TaskSetManager: Finished task 18580.0 in stage 148.0
(TID 153099) in 206425 ms on lhrrhegapq003.enterprisenet.org (18587/32768)
16/08/04 05:11:24 INFO TaskSetManager: Finished task 18585.0 in stage 148.0
(TID 153105) in 206418 ms on lhrrhegapq003.enterprisenet.org (18588/32768)
16/08/04 05:11:24 INFO TaskSetManager: Finished task 18586.0 in stage 148.0
(TID 153107) in 206431 ms on lhrrhegapq004.enterprisenet.org (18589/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18596.0 in stage 148.0
(TID 153112) in 207515 ms on lhrrhegapq008.enterprisenet.org (18590/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18587.0 in stage 148.0
(TID 153108) in 207517 ms on lhrrhegapq002.enterprisenet.org (18591/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18588.0 in stage 148.0
(TID 153106) in 207519 ms on lhrrhegapq008.enterprisenet.org (18592/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18590.0 in stage 148.0
(TID 153109) in 207519 ms on lhrrhegapq007.enterprisenet.org (18593/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18594.0 in stage 148.0
(TID 153114) in 207524 ms on lhrrhegapq004.enterprisenet.org (18594/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18593.0 in stage 148.0
(TID 153111) in 207526 ms on lhrrhegapq004.enterprisenet.org (18595/32768)
16/08/04 05:11:25 ERROR pkgews_panel_extract$Log$: PKGEWS_PANEL_EXTRACT,
p_signed_rank_yago, 2, Merge into TEWS_MS_TYPE_BASKET < 0  - NOT OK
java.lang.IllegalStateException: Cannot call methods on a stopped
SparkContext
        at org.apache.spark.SparkContext.org
$apache$spark$SparkContext$$assertNotStopped(SparkContext.scala:104)
        at
org.apache.spark.SparkContext.defaultParallelism(SparkContext.scala:2063)
        at
org.apache.spark.SparkContext.defaultMinPartitions(SparkContext.scala:2076)
        at
org.apache.spark.sql.hive.HadoopTableReader.<init>(TableReader.scala:70)
        at
org.apache.spark.sql.hive.execution.HiveTableScan.<init>(HiveTableScan.scala:77)
        at
org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$3.apply(HiveStrategies.scala:77)
        at
org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$$anonfun$3.apply(HiveStrategies.scala:77)
        at
org.apache.spark.sql.SQLContext$SparkPlanner.pruneFilterProject(SQLContext.scala:853)
        at
org.apache.spark.sql.hive.HiveStrategies$HiveTableScans$.apply(HiveStrategies.scala:73)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
        at
org.apache.spark.sql.execution.SparkStrategies$EquiJoinSelection$.makeBroadcastHashJoin(SparkStrategies.scala:92)
        at
org.apache.spark.sql.execution.SparkStrategies$EquiJoinSelection$.apply(SparkStrategies.scala:101)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
        at
org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:346)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
        at
org.apache.spark.sql.execution.SparkStrategies$EquiJoinSelection$.makeBroadcastHashJoin(SparkStrategies.scala:92)
        at
org.apache.spark.sql.execution.SparkStrategies$EquiJoinSelection$.apply(SparkStrategies.scala:101)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
        at
org.apache.spark.sql.execution.SparkStrategies$BasicOperators$.apply(SparkStrategies.scala:346)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.planLater(QueryPlanner.scala:54)
        at
org.apache.spark.sql.execution.SparkStrategies$Aggregation$.apply(SparkStrategies.scala:235)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner$$anonfun$1.apply(QueryPlanner.scala:58)
        at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:371)
        at
org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:59)
        at
org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan$lzycompute(SQLContext.scala:926)
        at
org.apache.spark.sql.SQLContext$QueryExecution.sparkPlan(SQLContext.scala:924)
        at
org.apache.spark.sql.SQLContext$QueryExecution.executedPlan$lzycompute(SQLContext.scala:930)
        at
org.apache.spark.sql.SQLContext$QueryExecution.executedPlan(SQLContext.scala:930)
        at
org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:53)
        at
org.apache.spark.sql.DataFrame.withNewExecutionId(DataFrame.scala:1904)
        at org.apache.spark.sql.DataFrame.collect(DataFrame.scala:1385)
        at
com.acnielsen.madras.pkgews_panel_extract$.p_signed_rank_yago(pkgews_panel_extract.scala:685)
        at
com.acnielsen.madras.pkgews_panel_extract$.p_main(pkgews_panel_extract.scala:4844)
        at
com.acnielsen.madras.pkgews_panel_extract$.main(pkgews_panel_extract.scala:4655)
        at
com.acnielsen.madras.pkgews_panel_extract.main(pkgews_panel_extract.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
        at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
        at
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18592.0 in stage 148.0
(TID 153110) in 207527 ms on lhrrhegapq008.enterprisenet.org (18596/32768)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18603.0 in stage 148.0
(TID 153116) in 207525 ms on lhrrhegapq006.enterprisenet.org (18597/32768)
16/08/04 05:11:25 ERROR pkgews_panel_extract$Log$: PKGEWS_PANEL_EXTRACT,
p_signed_rank_yago, 2, Merge into TEWS_MS_TYPE_BASKET < 0  - NOT OK ,
PKGEWS_PANEL_EXTRACT, p_signed_rank_yago, 2, Error in p_signed_rank_yago  -
NOT OK
Error >>>>> : null
Exception in thread "main" java.lang.Exception
        at
com.acnielsen.madras.pkgews_panel_extract$.p_main(pkgews_panel_extract.scala:4875)
        at
com.acnielsen.madras.pkgews_panel_extract$.main(pkgews_panel_extract.scala:4655)
        at
com.acnielsen.madras.pkgews_panel_extract.main(pkgews_panel_extract.scala)
        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
        at
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
        at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
        at java.lang.reflect.Method.invoke(Method.java:606)
        at
org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:672)
        at
org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
        at
org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
        at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120)
        at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
16/08/04 05:11:25 INFO TaskSetManager: Finished task 18602.0 in stage 148.0
(TID 153115) in 207526 ms on lhrrhegapq006.enterprisenet.org (18598/32768)
16/08/04 05:11:27 WARN QueuedThreadPool: 6 threads could not be stopped
16/08/04 05:11:27 INFO TaskSetManager: Finished task 18597.0 in stage 148.0
(TID 153113) in 208704 ms on lhrrhegapq008.enterprisenet.org (18599/32768)
16/08/04 05:11:27 INFO DiskBlockManager: Shutdown hook called
16/08/04 05:11:27 WARN AkkaRpcEndpointRef: Error sending message [message =
StopMapOutputTracker] in 1 attempts
org.apache.spark.rpc.RpcTimeoutException:
Recipient[Actor[akka://sparkDriver/user/MapOutputTracker#1304693884]] had
already been terminated.. This timeout is controlled by spark.rpc.askTimeout
        at org.apache.spark.rpc.RpcTimeout.org
$apache$spark$rpc$RpcTimeout$$createRpcTimeoutException(RpcEnv.scala:214)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:229)
        at
org.apache.spark.rpc.RpcTimeout$$anonfun$addMessageIfTimeout$1.applyOrElse(RpcEnv.scala:225)
        at
scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
        at scala.util.Failure$$anonfun$recover$1.apply(Try.scala:185)
        at scala.util.Try$.apply(Try.scala:161)
        at scala.util.Failure.recover(Try.scala:185)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at
scala.concurrent.Future$$anonfun$recover$1.apply(Future.scala:324)
        at scala.concurrent.impl.CallbackRunnable.run(Promise.scala:32)
        at
org.spark-project.guava.util.concurrent.MoreExecutors$SameThreadExecutorService.execute(MoreExecutors.java:293)
        at
scala.concurrent.impl.ExecutionContextImpl$$anon$1.execute(ExecutionContextImpl.scala:133)
        at
scala.concurrent.impl.CallbackRunnable.executeWithValue(Promise.scala:40)
        at
scala.concurrent.impl.Promise$DefaultPromise.scala$concurrent$impl$Promise$DefaultPromise$$dispatchOrAddCallback(Promise.scala:280)
        at
scala.concurrent.impl.Promise$DefaultPromise.onComplete(Promise.scala:270)
        at scala.concurrent.Future$class.recover(Future.scala:324)
        at
scala.concurrent.impl.Promise$DefaultPromise.recover(Promise.scala:153)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:319)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:100)
        at
org.apache.spark.rpc.RpcEndpointRef.askWithRetry(RpcEndpointRef.scala:77)
        at
org.apache.spark.MapOutputTracker.askTracker(MapOutputTracker.scala:109)
        at
org.apache.spark.MapOutputTracker.sendTracker(MapOutputTracker.scala:119)
        at
org.apache.spark.MapOutputTrackerMaster.stop(MapOutputTracker.scala:382)
        at org.apache.spark.SparkEnv.stop(SparkEnv.scala:97)
        at
org.apache.spark.SparkContext$$anonfun$stop$12.apply$mcV$sp(SparkContext.scala:1749)
        at
org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1184)
        at org.apache.spark.SparkContext.stop(SparkContext.scala:1748)
        at
org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend$MonitorThread.run(YarnClientSchedulerBackend.scala:146)
Caused by: akka.pattern.AskTimeoutException:
Recipient[Actor[akka://sparkDriver/user/MapOutputTracker#1304693884]] had
already been terminated.
        at akka.pattern.AskableActorRef$.ask$extension(AskSupport.scala:134)
        at
org.apache.spark.rpc.akka.AkkaRpcEndpointRef.ask(AkkaRpcEnv.scala:307)
        ... 10 more
16/08/04 05:11:27 INFO TaskSetManager: Finished task 18604.0 in stage 148.0
(TID 153127) in 208703 ms on lhrrhegapq003.enterprisenet.org (18600/32768)
16/08/04 05:11:27 INFO ZooKeeper: Session: 0x15590c0c2202d11 closed
16/08/04 05:11:27 INFO TaskSetManager: Finished task 18598.0 in stage 148.0
(TID 153117) in 208707 ms on lhrrhegapq002.enterprisenet.org (18601/32768)
16/08/04 05:11:27 INFO ClientCnxn: EventThread shut down
16/08/04 05:11:27 INFO CuratorFrameworkSingleton: Closing ZooKeeper client.
16/08/04 05:11:27 INFO TaskSetManager: Finished task 18599.0 in stage 148.0
(TID 153119) in 208708 ms on lhrrhegapq004.enterprisenet.org (18602/32768)
16/08/04 05:11:27 INFO ShutdownHookManager: Shutdown hook called
16/08/04 05:11:27 INFO ShutdownHookManager: Deleting directory
/tmp/spark-419c3d6c-a6d4-49c4-9f77-b79c5d3c8f37
16/08/04 05:11:27 INFO ShutdownHookManager: Deleting directory
/tmp/spark-579fc117-08a4-45a6-a2a1-2a6e90c1adcd/userFiles-f739d328-716d-4641-821f-a1b2af709182
16/08/04 05:11:27 INFO ShutdownHookManager: Deleting directory
/tmp/spark-579fc117-08a4-45a6-a2a1-2a6e90c1adcd
16/08/04 05:11:27 INFO TaskSetManager: Finished task 18608.0 in stage 148.0
(TID 153121) in 208715 ms on lhrrhegapq006.enterprisenet.org (18603/32768)

Best regards,
Vasu

Reply via email to