If you are not setting the Spark hadoop version, Spark built using default
hadoop version("1.0.4").

Before importing Spark-1.0.0 libraries ,
build Spark using *SPARK_HADOOP_VERSION=2.4.0 sbt/sbt assembly *command.



On Thu, Jun 5, 2014 at 12:28 PM, bluejoe2008 <bluejoe2...@gmail.com> wrote:

>  thank you!
>
> i am developping a java project in Eclipse IDE on Windows
> in which spark 1.0.0 libraries are imported
> and now i want to open HDFS files as input
> the hadoop version of HDFS is 2.4.0
>
> 2014-06-05
> ------------------------------
> bluejoe2008
>
>  *From:* prabeesh k <prabsma...@gmail.com>
> *Date:* 2014-06-05 13:23
> *To:* user <user@spark.apache.org>
> *Subject:* Re: mismatched hdfs protocol
>   For building Spark for particular version of Hadoop
> Refer
> http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html
>
>
> On Thu, Jun 5, 2014 at 8:14 AM, Koert Kuipers <ko...@tresata.com> wrote:
>
>>  you have to build spark against the version of hadoop your are using
>>
>>
>> On Wed, Jun 4, 2014 at 10:25 PM, bluejoe2008 <bluejoe2...@gmail.com>
>> wrote:
>>
>>>  hi, all
>>> when my spark program accessed hdfs files
>>> an error happened:
>>>
>>> Exception in thread "main" org.apache.hadoop.ipc.RemoteException: Server 
>>> IPC version 9 cannot communicate with client version 4
>>>
>>>
>>> it seems the client was trying to connect hadoop2 via an old hadoop
>>> protocol
>>>
>>> so my question is:
>>> how to specify the version of hadoop on connection?
>>>
>>> thank you!
>>>
>>>  bluejoe
>>>
>>> 2014-06-05
>>> ------------------------------
>>>
>>>
>>
>>
>

Reply via email to