Hi Moshe,

Spark needs a Hadoop 2.x/YARN cluster. Other wise you can run it without
hadoop in the stand alone mode.

Manu



On Sat, Sep 20, 2014 at 12:55 AM, Moshe Beeri <moshe.be...@gmail.com> wrote:

> object Nizoz {
>
>   def connect(): Unit = {
>     val conf = new SparkConf().setAppName("nizoz").setMaster("master");
>     val spark = new SparkContext(conf)
>     val lines =
>
> spark.textFile("file:///home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md")
>     val lineLengths = lines.map(s => s.length)
>     val totalLength = lineLengths.reduce((a, b) => a + b)
>     println("totalLength=" + totalLength)
>
>   }
>
>   def main(args: Array[String]) {
>     println(scala.tools.nsc.Properties.versionString)
>     try {
>       //Nizoz.connect
>       val logFile =
> "/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" // Should
> be some file on your system
>       val conf = new SparkConf().setAppName("Simple
> Application").setMaster("spark://master:7077")
>       val sc = new SparkContext(conf)
>       val logData = sc.textFile(logFile, 2).cache()
>       val numAs = logData.filter(line => line.contains("a")).count()
>       val numBs = logData.filter(line => line.contains("b")).count()
>       println("Lines with a: %s, Lines with b: %s".format(numAs, numBs))
>
>     } catch {
>       case e => {
>         println(e.getCause())
>         println("stack:")
>         e.printStackTrace()
>       }
>     }
>   }
> }
> Runs with Scala 2.10.4
> The problem is this [vogue] exception:
>
>         at com.example.scamel.Nizoz.main(Nizoz.scala)
> Caused by: java.lang.RuntimeException:
> java.lang.reflect.InvocationTargetException
>         at
>
> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131)
>         at org.apache.hadoop.security.Groups.<init>(Groups.java:64)
>         at
>
> org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240)
> ...
> Caused by: java.lang.reflect.InvocationTargetException
>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
> Method)
>         at
>
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>         at
>
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> ...
>         ... 10 more
> Caused by: java.lang.UnsatisfiedLinkError:
> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V
>         at
> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native
> Method)
>         at
>
> org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49)
>
> I have Hadoop 1.2.1 running on Ubuntu 14.04, the Scala console run as
> expected.
>
> What am I doing wrong?
> Any idea will be welcome
>
>
>
>
>
> --
> View this message in context:
> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718.html
> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: user-unsubscr...@spark.apache.org
> For additional commands, e-mail: user-h...@spark.apache.org
>
>


-- 
Manu Suryavansh

Reply via email to