If you are not setting the Spark hadoop version, Spark built using default hadoop version("1.0.4").
Before importing Spark-1.0.0 libraries , build Spark using *SPARK_HADOOP_VERSION=2.4.0 sbt/sbt assembly *command. On Thu, Jun 5, 2014 at 12:28 PM, bluejoe2008 <bluejoe2...@gmail.com> wrote: > thank you! > > i am developping a java project in Eclipse IDE on Windows > in which spark 1.0.0 libraries are imported > and now i want to open HDFS files as input > the hadoop version of HDFS is 2.4.0 > > 2014-06-05 > ------------------------------ > bluejoe2008 > > *From:* prabeesh k <prabsma...@gmail.com> > *Date:* 2014-06-05 13:23 > *To:* user <user@spark.apache.org> > *Subject:* Re: mismatched hdfs protocol > For building Spark for particular version of Hadoop > Refer > http://spark.apache.org/docs/latest/hadoop-third-party-distributions.html > > > On Thu, Jun 5, 2014 at 8:14 AM, Koert Kuipers <ko...@tresata.com> wrote: > >> you have to build spark against the version of hadoop your are using >> >> >> On Wed, Jun 4, 2014 at 10:25 PM, bluejoe2008 <bluejoe2...@gmail.com> >> wrote: >> >>> hi, all >>> when my spark program accessed hdfs files >>> an error happened: >>> >>> Exception in thread "main" org.apache.hadoop.ipc.RemoteException: Server >>> IPC version 9 cannot communicate with client version 4 >>> >>> >>> it seems the client was trying to connect hadoop2 via an old hadoop >>> protocol >>> >>> so my question is: >>> how to specify the version of hadoop on connection? >>> >>> thank you! >>> >>> bluejoe >>> >>> 2014-06-05 >>> ------------------------------ >>> >>> >> >> >