Sure in local mode it works for me as well, the issue is that I run master only, I needed worker as well.
תודה רבה, משה בארי. 054-3133943 Email <moshe.be...@gmail.com> | linkedin <http://www.linkedin.com/in/mobee> On Mon, Sep 22, 2014 at 9:58 AM, Akhil Das-2 [via Apache Spark User List] < ml-node+s1001560n14785...@n3.nabble.com> wrote: > Hi Moshe, > > I ran the same code on my machine and it is working without any issues. > You can try running it in local mode and if that is working fine, then the > issue is with your configuration. > > [image: Inline image 1] > > Thanks > Best Regards > > On Sat, Sep 20, 2014 at 3:34 PM, Moshe Beeri <[hidden email] > <http://user/SendEmail.jtp?type=node&node=14785&i=0>> wrote: > >> Hi Sean, >> >> Thanks a lot for the answer , I loved your excellent book >> *Mahout in Action >> <http://www.amazon.com/Mahout-Action-Sean-Owen/dp/1935182684> *hope >> you'll keep on writing more books in the field of Big Data. >> The issue was with redundant Hadoop library, But now I am facing some >> other issue (see prev post in this thread) >> java.lang.ClassNotFoundException: com.example.scamel.Nizoz$$anonfun$3 >> >> But the class com.example.scamel.Nizoz (in fact Scala object) is the one >> under debugging. >> >> def main(args: Array[String]) { >> println(scala.tools.nsc.Properties.versionString) >> try { >> //Nizoz.connect >> val logFile = >> "/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" // Should >> be some file on your system >> val conf = new SparkConf().setAppName("spark >> town").setMaster("spark://nash:7077"); //spark://master:7077 >> val sc = new SparkContext(conf) >> val logData = sc.textFile(logFile, 2).cache() >> *val numAs = logData.filter(line => line.contains("a")).count() >> // <- here is where the exception thrown * >> >> Do you have any idea whats wrong? >> Thanks, >> Moshe Beeri. >> >> >> ** >> >> >> תודה רבה, >> משה בארי. >> 054-3133943 >> [hidden email] <http://user/SendEmail.jtp?type=node&node=14731&i=0> | >> linkedin <http://www.linkedin.com/in/mobee> >> >> >> >> On Sat, Sep 20, 2014 at 12:02 PM, sowen [via Apache Spark User List] <[hidden >> email] <http://user/SendEmail.jtp?type=node&node=14731&i=1>> wrote: >> >>> Spark does not require Hadoop 2 or YARN. This looks like a problem with >>> the Hadoop installation as it is not funding native libraries it needs to >>> make some security related system call. Check the installation. >>> On Sep 20, 2014 9:13 AM, "Manu Suryavansh" <[hidden email] >>> <http://user/SendEmail.jtp?type=node&node=14724&i=0>> wrote: >>> >>>> Hi Moshe, >>>> >>>> Spark needs a Hadoop 2.x/YARN cluster. Other wise you can run it >>>> without hadoop in the stand alone mode. >>>> >>>> Manu >>>> >>>> >>>> >>>> On Sat, Sep 20, 2014 at 12:55 AM, Moshe Beeri <[hidden email] >>>> <http://user/SendEmail.jtp?type=node&node=14724&i=1>> wrote: >>>> >>>>> object Nizoz { >>>>> >>>>> def connect(): Unit = { >>>>> val conf = new SparkConf().setAppName("nizoz").setMaster("master"); >>>>> val spark = new SparkContext(conf) >>>>> val lines = >>>>> >>>>> spark.textFile("file:///home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md") >>>>> val lineLengths = lines.map(s => s.length) >>>>> val totalLength = lineLengths.reduce((a, b) => a + b) >>>>> println("totalLength=" + totalLength) >>>>> >>>>> } >>>>> >>>>> def main(args: Array[String]) { >>>>> println(scala.tools.nsc.Properties.versionString) >>>>> try { >>>>> //Nizoz.connect >>>>> val logFile = >>>>> "/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" // >>>>> Should >>>>> be some file on your system >>>>> val conf = new SparkConf().setAppName("Simple >>>>> Application").setMaster("spark://master:7077") >>>>> val sc = new SparkContext(conf) >>>>> val logData = sc.textFile(logFile, 2).cache() >>>>> val numAs = logData.filter(line => line.contains("a")).count() >>>>> val numBs = logData.filter(line => line.contains("b")).count() >>>>> println("Lines with a: %s, Lines with b: %s".format(numAs, >>>>> numBs)) >>>>> >>>>> } catch { >>>>> case e => { >>>>> println(e.getCause()) >>>>> println("stack:") >>>>> e.printStackTrace() >>>>> } >>>>> } >>>>> } >>>>> } >>>>> Runs with Scala 2.10.4 >>>>> The problem is this [vogue] exception: >>>>> >>>>> at com.example.scamel.Nizoz.main(Nizoz.scala) >>>>> Caused by: java.lang.RuntimeException: >>>>> java.lang.reflect.InvocationTargetException >>>>> at >>>>> >>>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131) >>>>> at org.apache.hadoop.security.Groups.<init>(Groups.java:64) >>>>> at >>>>> >>>>> org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240) >>>>> ... >>>>> Caused by: java.lang.reflect.InvocationTargetException >>>>> at >>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >>>>> at >>>>> >>>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >>>>> at >>>>> >>>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >>>>> ... >>>>> ... 10 more >>>>> Caused by: java.lang.UnsatisfiedLinkError: >>>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V >>>>> at >>>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native >>>>> Method) >>>>> at >>>>> >>>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49) >>>>> >>>>> I have Hadoop 1.2.1 running on Ubuntu 14.04, the Scala console run as >>>>> expected. >>>>> >>>>> What am I doing wrong? >>>>> Any idea will be welcome >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> View this message in context: >>>>> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718.html >>>>> Sent from the Apache Spark User List mailing list archive at >>>>> Nabble.com. >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: [hidden email] >>>>> <http://user/SendEmail.jtp?type=node&node=14724&i=2> >>>>> For additional commands, e-mail: [hidden email] >>>>> <http://user/SendEmail.jtp?type=node&node=14724&i=3> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Manu Suryavansh >>>> >>> >>> >>> ------------------------------ >>> If you reply to this email, your message will be added to the >>> discussion below: >>> >>> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14724.html >>> To unsubscribe from Fails to run simple Spark (Hello World) scala >>> program, click here. >>> NAML >>> <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml> >>> >> >> >> ------------------------------ >> View this message in context: Re: Fails to run simple Spark (Hello >> World) scala program >> <http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14731.html> >> >> Sent from the Apache Spark User List mailing list archive >> <http://apache-spark-user-list.1001560.n3.nabble.com/> at Nabble.com. >> > > > > ------------------------------ > If you reply to this email, your message will be added to the discussion > below: > > http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14785.html > To unsubscribe from Fails to run simple Spark (Hello World) scala > program, click here > <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=14718&code=bW9zaGUuYmVlcmlAZ21haWwuY29tfDE0NzE4fDE0NzUwMDQ2Ng==> > . > NAML > <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml> > -- View this message in context: http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14907.html Sent from the Apache Spark User List mailing list archive at Nabble.com.