Hi Sean,

Thanks a lot for the answer , I loved your excellent book
*​Mahout in Action
<http://www.amazon.com/Mahout-Action-Sean-Owen/dp/1935182684> *hope you'll
keep on writing more books in the field of Big Data.
The issue was with redundant Hadoop library, But now I am facing some other
issue (see prev post in this thread)
java.lang.ClassNotFoundException: com.example.scamel.Nizoz$$anonfun$3

But the class com.example.scamel.Nizoz (in fact Scala object) is the one
under debugging.

  def main(args: Array[String]) {
    println(scala.tools.nsc.Properties.versionString)
    try {
      //Nizoz.connect
      val logFile =
"/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" // Should
be some file on your system
      val conf = new SparkConf().setAppName("spark
town").setMaster("spark://nash:7077"); //spark://master:7077
      val sc = new SparkContext(conf)
      val logData = sc.textFile(logFile, 2).cache()
      *val numAs = logData.filter(line => line.contains("a")).count()    //
<- here is  where the exception thrown *

Do you have any idea whats wrong?
Thanks,
Moshe Beeri.


*​*


תודה רבה,
משה בארי.
054-3133943
Email <moshe.be...@gmail.com> | linkedin <http://www.linkedin.com/in/mobee>



On Sat, Sep 20, 2014 at 12:02 PM, sowen [via Apache Spark User List] <
ml-node+s1001560n14724...@n3.nabble.com> wrote:

> Spark does not require Hadoop 2 or YARN. This looks like a problem with
> the Hadoop installation as it is not funding native libraries it needs to
> make some security related system call. Check the installation.
> On Sep 20, 2014 9:13 AM, "Manu Suryavansh" <[hidden email]
> <http://user/SendEmail.jtp?type=node&node=14724&i=0>> wrote:
>
>> Hi Moshe,
>>
>> Spark needs a Hadoop 2.x/YARN cluster. Other wise you can run it without
>> hadoop in the stand alone mode.
>>
>> Manu
>>
>>
>>
>> On Sat, Sep 20, 2014 at 12:55 AM, Moshe Beeri <[hidden email]
>> <http://user/SendEmail.jtp?type=node&node=14724&i=1>> wrote:
>>
>>> object Nizoz {
>>>
>>>   def connect(): Unit = {
>>>     val conf = new SparkConf().setAppName("nizoz").setMaster("master");
>>>     val spark = new SparkContext(conf)
>>>     val lines =
>>>
>>> spark.textFile("file:///home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md")
>>>     val lineLengths = lines.map(s => s.length)
>>>     val totalLength = lineLengths.reduce((a, b) => a + b)
>>>     println("totalLength=" + totalLength)
>>>
>>>   }
>>>
>>>   def main(args: Array[String]) {
>>>     println(scala.tools.nsc.Properties.versionString)
>>>     try {
>>>       //Nizoz.connect
>>>       val logFile =
>>> "/home/moshe/store/frameworks/spark-1.1.0-bin-hadoop1/README.md" //
>>> Should
>>> be some file on your system
>>>       val conf = new SparkConf().setAppName("Simple
>>> Application").setMaster("spark://master:7077")
>>>       val sc = new SparkContext(conf)
>>>       val logData = sc.textFile(logFile, 2).cache()
>>>       val numAs = logData.filter(line => line.contains("a")).count()
>>>       val numBs = logData.filter(line => line.contains("b")).count()
>>>       println("Lines with a: %s, Lines with b: %s".format(numAs, numBs))
>>>
>>>     } catch {
>>>       case e => {
>>>         println(e.getCause())
>>>         println("stack:")
>>>         e.printStackTrace()
>>>       }
>>>     }
>>>   }
>>> }
>>> Runs with Scala 2.10.4
>>> The problem is this [vogue] exception:
>>>
>>>         at com.example.scamel.Nizoz.main(Nizoz.scala)
>>> Caused by: java.lang.RuntimeException:
>>> java.lang.reflect.InvocationTargetException
>>>         at
>>>
>>> org.apache.hadoop.util.ReflectionUtils.newInstance(ReflectionUtils.java:131)
>>>         at org.apache.hadoop.security.Groups.<init>(Groups.java:64)
>>>         at
>>>
>>> org.apache.hadoop.security.Groups.getUserToGroupsMappingService(Groups.java:240)
>>> ...
>>> Caused by: java.lang.reflect.InvocationTargetException
>>>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>>> Method)
>>>         at
>>>
>>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>>>         at
>>>
>>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>>> ...
>>>         ... 10 more
>>> Caused by: java.lang.UnsatisfiedLinkError:
>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative()V
>>>         at
>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.anchorNative(Native
>>> Method)
>>>         at
>>>
>>> org.apache.hadoop.security.JniBasedUnixGroupsMapping.<clinit>(JniBasedUnixGroupsMapping.java:49)
>>>
>>> I have Hadoop 1.2.1 running on Ubuntu 14.04, the Scala console run as
>>> expected.
>>>
>>> What am I doing wrong?
>>> Any idea will be welcome
>>>
>>>
>>>
>>>
>>>
>>> --
>>> View this message in context:
>>> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718.html
>>> Sent from the Apache Spark User List mailing list archive at Nabble.com.
>>>
>>> ---------------------------------------------------------------------
>>> To unsubscribe, e-mail: [hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=14724&i=2>
>>> For additional commands, e-mail: [hidden email]
>>> <http://user/SendEmail.jtp?type=node&node=14724&i=3>
>>>
>>>
>>
>>
>> --
>> Manu Suryavansh
>>
>
>
> ------------------------------
>  If you reply to this email, your message will be added to the discussion
> below:
>
> http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14724.html
>  To unsubscribe from Fails to run simple Spark (Hello World) scala
> program, click here
> <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=unsubscribe_by_code&node=14718&code=bW9zaGUuYmVlcmlAZ21haWwuY29tfDE0NzE4fDE0NzUwMDQ2Ng==>
> .
> NAML
> <http://apache-spark-user-list.1001560.n3.nabble.com/template/NamlServlet.jtp?macro=macro_viewer&id=instant_html%21nabble%3Aemail.naml&base=nabble.naml.namespaces.BasicNamespace-nabble.view.web.template.NabbleNamespace-nabble.view.web.template.NodeNamespace&breadcrumbs=notify_subscribers%21nabble%3Aemail.naml-instant_emails%21nabble%3Aemail.naml-send_instant_email%21nabble%3Aemail.naml>
>




--
View this message in context: 
http://apache-spark-user-list.1001560.n3.nabble.com/Fails-to-run-simple-Spark-Hello-World-scala-program-tp14718p14731.html
Sent from the Apache Spark User List mailing list archive at Nabble.com.

Reply via email to