ok, i see
i imported wrong jar files which only work well on default hadoop version
2014-06-05
bluejoe2008
From: prabeesh k
Date: 2014-06-05 16:14
To: user
Subject: Re: Re: mismatched hdfs protocol
If you are not setting the Spark hadoop version, Spark built using default
hadoop version("1.0.4").
Before importing Spark-1.0.0 libraries ,
build Spark using SPARK_HADOOP_VERSION=2.4.0 sbt/sbt assembly command.
On Thu, Jun 5, 2014 at 12:28 PM, bluejoe2008 <[email protected]> wrote:
thank you!
i am developping a java project in Eclipse IDE on Windows
in which spark 1.0.0 libraries are imported
and now i want to open HDFS files as input
the hadoop version of HDFS is 2.4.0
2014-06-05
bluejoe2008
From: prabeesh k
Date: 2014-06-05 13:23
To: user
Subject: Re: mismatched hdfs protocol
For building Spark for particular version of Hadoop
Refer http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html
On Thu, Jun 5, 2014 at 8:14 AM, Koert Kuipers <[email protected]> wrote:
you have to build spark against the version of hadoop your are using
On Wed, Jun 4, 2014 at 10:25 PM, bluejoe2008 <[email protected]> wrote:
hi, all
when my spark program accessed hdfs files
an error happened:
Exception in thread "main" org.apache.hadoop.ipc.RemoteException: Server IPC
version 9 cannot communicate with client version 4
it seems the client was trying to connect hadoop2 via an old hadoop protocol
so my question is:
how to specify the version of hadoop on connection?
thank you!
bluejoe
2014-06-05