Just make sure your hadoop instances are functioning properly, (check for ResourceManager, NodeManager). How are you submitting the job? If that is getting submitted then you can look further in the yarn logs to see whats really going on.
Thanks Best Regards On Thu, Aug 6, 2015 at 6:59 PM, Clint McNeil <[email protected]> wrote: > Hi > > I am trying to launch a Spark application on a CM cluster and I get the > following error. > > Exception in thread "main" org.apache.spark.SparkException: Yarn > application has already ended! It might have been killed or unable to > launch application master. > > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.waitForApplication(YarnClientSchedulerBackend.scala:113) > > at > org.apache.spark.scheduler.cluster.YarnClientSchedulerBackend.start(YarnClientSchedulerBackend.scala:59) > > at > org.apache.spark.scheduler.TaskSchedulerImpl.start(TaskSchedulerImpl.scala:141) > > at org.apache.spark.SparkContext.<init>(SparkContext.scala:379) > > What is the remedy for this type of problem > > -- > > *Clint McNeil* > > BI & Data Science Engineer | Impact Radius > > 202 Suntyger, 313 Durban Road, Bellville, 7530 > > o: +2721 914-1764 <%2B2721%20910-3195> | m: +2782 4796 309 | > [email protected] > > *Learn more – Watch our 2 minute overview > <http://www.impactradius.com/?src=slsap>* > > www.impactradius.com | Twitter <http://twitter.com/impactradius> | > Facebook <https://www.facebook.com/pages/Impact-Radius/153376411365183> | > LinkedIn <http://www.linkedin.com/company/impact-radius-inc.> | YouTube > <https://www.youtube.com/user/ImpactRadius> > > Maximizing Return on Ad Spend > >
