Spark does not require Hadoop 2 or YARN. This looks like a problem with the Hadoop installation as it is not funding native libraries it needs to make some security related system call. Check the installation. On Sep 20, 2014 9:13 AM, "Manu Suryavansh" <suryavanshi.m...@gmail.com> wrote:
> Hi Moshe, > > Spark needs a Hadoop 2.x/YARN cluster. Other wise you can run it without > hadoop in the stand alone mode. > > Manu > > > > On Sat, Sep 20, 2014 at 12:55 AM, Moshe Beeri <moshe.be...@gmail.com> > wrote: > >> object Nizoz { >> >> def connect(): Unit = { >> val conf = new SparkConf().setAppName("nizoz").setMaster("master"); >> val spark = new SparkContext(conf) >> val lines = >> >> spark.textFile("file:///home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md") >> val lineLengths = lines.map(s => s.length) >> val totalLength = lineLengths.reduce((a, b) => a + b) >> println("totalLength=" + totalLength) >> >> } >> >> def main(args: Array[String]) { >> println(scala.tools.nsc.Properties.versionString) >> try { >> //Nizoz.connect >> val logFile = >> "/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" // Should >> be some file on your system >> val conf = new SparkConf().setAppName("Simple >> Application").setMaster("spark://master:7077") >> val sc = new SparkContext(conf) >> val logData = sc.textFile(logFile, 2).cache() >> val numAs = logData.filter(line => line.contains("a")).count() >> val numBs = logData.filter(line => line.contains("b")).count() >> println("Lines with a: %s, Lines with b: %s".format(numAs, numBs)) >> >> } catch { >> case e => { >> println(e.getCause()) >> println("stack:") >> e.printStackTrace() >> } >> } >> } >> } >> Runs with Scala 2.10.4 >> The problem is this [vogue] exception: >> >> at com.example.scamel.Nizoz.main(Nizoz.scala) >> Caused by: java.lang.RuntimeException: >> java.lang.reflect.InvocationTargetException >> at >> >> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131) >> at org.apache.hadoop.security.Groups.<init>(Groups.java:64) >> at >> >> org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240) >> ... >> Caused by: java.lang.reflect.InvocationTargetException >> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native >> Method) >> at >> >> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62) >> at >> >> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) >> ... >> ... 10 more >> Caused by: java.lang.UnsatisfiedLinkError: >> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V >> at >> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native >> Method) >> at >> >> org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49) >> >> I have Hadoop 1.2.1 running on Ubuntu 14.04, the Scala console run as >> expected. >> >> What am I doing wrong? >> Any idea will be welcome >> >> >> >> >> >> -- >> View this message in context: >> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718.html >> Sent from the Apache Spark User List mailing list archive at Nabble.com. >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org >> For additional commands, e-mail: user-h...@spark.apache.org >> >> > > > -- > Manu Suryavansh >