Looking at the lib folder revealed the problem. The lib folder on one of
the nodes had libraries for both hadoop 1 and 2. I’m not sure how I ended
up with that but it must have happened while I was copying the dependency
jars to each node. I removed all the jars and started with a fresh copy
and reran the test. It worked this time.

Thank you so much for being patient with me and sorry if I wasted your
time.

-Ali

On 2016-01-04, 3:17 PM, "Robert Metzger" <rmetz...@apache.org> wrote:

>Okay, then, my previous statement is false. From the stack trace, it seems
>that Flink is using an older Hadoop version. The DFSClient and RPC classes
>look different in Hadoop 2.6.0.
>Max checked already for some environment variables. Is $CLASSPATH set? Did
>you install Hadoop only by downloading the binaries or did you use
>something like the ubuntu package manager? (those packages sometimes place
>jar files in some lib/ folders)
>Did you put something additionally into the lib/ folder of Flink?
>
>I think when Flink is starting up, its also logging the classpath of the
>JVM. Can you post it here? (Maybe you need to start with DEBUG log level)
>
>
>On Mon, Jan 4, 2016 at 8:46 PM, Kashmar, Ali <ali.kash...@emc.com> wrote:
>
>> Hi Robert,
>>
>> On the ubuntu host, the port 9000 is up:
>>
>> ubuntu@ubuntu-171:~$ netstat -tulpn
>> ....
>> tcp        0      0 10.13.182.171:9000      0.0.0.0:*
>>LISTEN
>>      8849/java
>> ....
>>
>> And the process using this port is:
>>
>> ubuntu@ubuntu-171:~$ ps -ef | grep 8849
>> ubuntu    8849     1  0  2015 ?        00:25:00
>> /usr/lib/jvm/java-8-oracle/bin/java -Dproc_namenode -Xmx1000m
>> -Djava.net.preferIPv4Stack=true
>> -Dhadoop.log.dir=/home/ubuntu/Documents/hadoop-2.6.0/logs
>> -Dhadoop.log.file=hadoop.log
>> -Dhadoop.home.dir=/home/ubuntu/Documents/hadoop-2.6.0
>> -Dhadoop.id.str=ubuntu -Dhadoop.root.logger=INFO,console
>> -Djava.library.path=/home/ubuntu/Documents/hadoop-2.6.0/lib/native
>> -Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true
>> -Djava.net.preferIPv4Stack=true -Djava.net.preferIPv4Stack=true
>> -Dhadoop.log.dir=/home/ubuntu/Documents/hadoop-2.6.0/logs
>> -Dhadoop.log.file=hadoop-ubuntu-namenode-ubuntu-171.log
>> -Dhadoop.home.dir=/home/ubuntu/Documents/hadoop-2.6.0
>> -Dhadoop.id.str=ubuntu -Dhadoop.root.logger=INFO,RFA
>> -Djava.library.path=/home/ubuntu/Documents/hadoop-2.6.0/lib/native
>> -Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true
>> -Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
>> -Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
>> -Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
>> -Dhadoop.security.logger=INFO,RFAS
>> org.apache.hadoop.hdfs.server.namenode.NameNode
>>
>> And the hadoop version is:
>>
>> ubuntu@ubuntu-171:~$ Documents/hadoop-2.6.0/bin/hadoop version
>> Hadoop 2.6.0
>> Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r
>> e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
>> Compiled by jenkins on 2014-11-13T21:10Z
>> Compiled with protoc 2.5.0
>> From source with checksum 18e43357c8f927c0695f1e9522859d6a
>> This command was run using
>> 
>>/home/ubuntu/Documents/hadoop-2.6.0/share/hadoop/common/hadoop-common-2.6
>>.0
>> .jar
>>
>>
>>
>> -Ali
>>
>>
>>
>> On 2016-01-04, 2:20 PM, "Robert Metzger" <rmetz...@apache.org> wrote:
>>
>> >Your Flink installation has Hadoop 2.6.0 included, on the other
>>machine,
>> >there is a Hadoop version installed, which is most likely a 1.x or
>>even a
>> >0.x version.
>> >Are you sure that the host "ubuntu-171" has the ip 10.13.182.171 and
>>that
>> >the hadoop installation in the "/home/ubuntu/Documents/hadoop-2.6.0/"
>> >directory is listening on port 9000 ?
>> >
>> >On Mon, Jan 4, 2016 at 3:03 PM, Kashmar, Ali <ali.kash...@emc.com>
>>wrote:
>> >
>> >> Hi Max,
>> >>
>> >> Both commands return nothing. Those variables aren’t set.
>> >>
>> >> The only software I installed on these machines is Flink and Java.
>> >>
>> >> -Ali
>> >>
>> >> On 2015-12-28, 6:42 AM, "Maximilian Michels" <m...@apache.org> wrote:
>> >>
>> >> >Hi Ali,
>> >> >
>> >> >The warning about the native Hadoop libraries is nothing to worry
>> >> >about. The native modules are platform-optimized modules which may
>>be
>> >> >used to improve performance. They are not necessary for Hadoop to
>> >> >function correctly.
>> >> >
>> >> >The exception message implies that you are using is a very old
>>version
>> >> >of Hadoop. Do you have other Hadoop versions installed on the same
>> >> >machine? We have had people using Flink 0.10.0 with Hadoop 2.6.0
>> >> >without any problems.
>> >> >
>> >> >On the cluster machines, what is the output of these commands?
>> >> >
>> >> >echo $HADOOP_CLASSPATH
>> >> >echo $HADOOP_CONF_DIR
>> >> >
>> >> >
>> >> >Thanks,
>> >> >Max
>> >> >
>> >> >On Wed, Dec 23, 2015 at 3:53 PM, Kashmar, Ali <ali.kash...@emc.com>
>> >> wrote:
>> >> >> Hi Max,
>> >> >>
>> >> >> I have the same output for the Task Manager:
>> >> >>
>> >> >> 11:25:04,274 INFO
>>org.apache.flink.runtime.taskmanager.TaskManager
>> >> >>       -  Hadoop version: 2.6.0
>> >> >>
>> >> >> I do get this line at the beginning of both job and task manager
>>log
>> >> >>files:
>> >> >>
>> >> >> 11:25:04,100 WARN  org.apache.hadoop.util.NativeCodeLoader
>> >> >>       - Unable to load native-hadoop library for your platform...
>> >>using
>> >> >> builtin-java classes where applicable
>> >> >>
>> >> >> Do you think it has anything to do with it?
>> >> >>
>> >> >> Thanks,
>> >> >> Ali
>> >> >>
>> >> >> On 2015-12-23, 7:30 AM, "Maximilian Michels" <m...@apache.org>
>>wrote:
>> >> >>
>> >> >>>Hi Ali,
>> >> >>>
>> >> >>>Could you please also post the Hadoop version output of the task
>> >> >>>manager log files? It looks like the task managers are running a
>> >> >>>different Hadoop version.
>> >> >>>
>> >> >>>Thanks,
>> >> >>>Max
>> >> >>>
>> >> >>>On Tue, Dec 22, 2015 at 4:28 PM, Kashmar, Ali
>><ali.kash...@emc.com>
>> >> >>>wrote:
>> >> >>>> Hi Robert,
>> >> >>>>
>> >> >>>> I found the version in the job manager log file:
>> >> >>>>
>> >> >>>> 17:33:49,636 INFO
>>org.apache.flink.runtime.jobmanager.JobManager
>> >> >>>>       -  Hadoop version: 2.6.0
>> >> >>>>
>> >> >>>> But the Hadoop installation I have is saying this:
>> >> >>>>
>> >> >>>> ubuntu@ubuntu-171:~/Documents/hadoop-2.6.0$ bin/hadoop version
>> >> >>>> Hadoop 2.6.0
>> >> >>>> Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r
>> >> >>>> e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
>> >> >>>> Compiled by jenkins on 2014-11-13T21:10Z
>> >> >>>> Compiled with protoc 2.5.0
>> >> >>>> From source with checksum 18e43357c8f927c0695f1e9522859d6a
>> >> >>>> This command was run using
>> >> >>>>
>> >>
>> 
>>>>>>>>/home/ubuntu/Documents/hadoop-2.6.0/share/hadoop/common/hadoop-comm
>>>>>>>>on
>> >>>>>>-2
>> >> >>>>.6
>> >> >>>>.0
>> >> >>>> .jar
>> >> >>>>
>> >> >>>>
>> >> >>>> So one of them is lying to me? :)
>> >> >>>>
>> >> >>>> Ali
>> >> >>>>
>> >> >>>> On 2015-12-22, 10:16 AM, "Robert Metzger" <rmetz...@apache.org>
>> >> wrote:
>> >> >>>>
>> >> >>>>>Hi Ali,
>> >> >>>>>
>> >> >>>>>the TaskManagers and the JobManager is logging the Hadoop
>>version
>> >>on
>> >> >>>>>startup.
>> >> >>>>>
>> >> >>>>>On Tue, Dec 22, 2015 at 4:10 PM, Kashmar, Ali
>><ali.kash...@emc.com
>> >
>> >> >>>>>wrote:
>> >> >>>>>
>> >> >>>>>> Hello,
>> >> >>>>>>
>> >> >>>>>> I¹m trying to use HDFS as store for Flink checkpoints so I
>> >> >>>>>>downloaded
>> >> >>>>>>the
>> >> >>>>>> Hadoop 2.6.0/Scala 2.10 version of Flink and installed it. I
>>also
>> >> >>>>>> downloaded Hadoop 2.6.0 separately from the Hadoop website and
>> >>set
>> >> >>>>>>up
>> >> >>>>>>HDFS
>> >> >>>>>> on a separate machine. When I start Flink I get the following
>> >>error:
>> >> >>>>>>
>> >> >>>>>> 17:34:13,047 INFO
>>org.apache.flink.runtime.jobmanager.JobManager
>> >> >>>>>>       - Status of job 9ba32a08bc0ec02810bf5d2710842f72
>>(Protocol
>> >> >>>>>>Event
>> >> >>>>>> Processing) changed to FAILED.
>> >> >>>>>> java.lang.Exception: Call to registerInputOutput() of
>>invokable
>> >> >>>>>>failed
>> >> >>>>>>         at
>> >> >>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:529)
>> >> >>>>>>         at java.lang.Thread.run(Thread.java:745)
>> >> >>>>>> Caused by: java.io.IOException: The given file URI (hdfs://
>> >> >>>>>> 10.13.182.171:9000/user/flink/checkpoints) points to the HDFS
>> >> >>>>>>NameNode
>> >> >>>>>>at
>> >> >>>>>> 10.13.182.171:9000, but the File System could not be
>>initialized
>> >> >>>>>>with
>> >> >>>>>> that address: Server IPC version 9 cannot communicate with
>>client
>> >> >>>>>>version 4
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(Hado
>>>>>>>>>>op
>> >>>>>>>>Fi
>> >> >>>>>>le
>> >> >>>>>>Sy
>> >> >>>>>>stem.java:337)
>> >> >>>>>>         at
>> >> >>>>>>org.apache.flink.core.fs.FileSystem.get(FileSystem.java:253)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(F
>>>>>>>>>>sS
>> >>>>>>>>ta
>> >> >>>>>>te
>> >> >>>>>>Ba
>> >> >>>>>>ckend.java:142)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(F
>>>>>>>>>>sS
>> >>>>>>>>ta
>> >> >>>>>>te
>> >> >>>>>>Ba
>> >> >>>>>>ckend.java:101)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackendFactory.c
>>>>>>>>>>re
>> >>>>>>>>at
>> >> >>>>>>eF
>> >> >>>>>>ro
>> >> >>>>>>mConfig(FsStateBackendFactory.java:48)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.createStateBa
>>>>>>>>>>ck
>> >>>>>>>>en
>> >> >>>>>>d(
>> >> >>>>>>St
>> >> >>>>>>reamTask.java:517)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.registerInput
>>>>>>>>>>Ou
>> >>>>>>>>tp
>> >> >>>>>>ut
>> >> >>>>>>(S
>> >> >>>>>>treamTask.java:171)
>> >> >>>>>>         at
>> >> >>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:526)
>> >> >>>>>>         ... 1 more
>> >> >>>>>> Caused by:
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.RPC$V
>>>>>>>>>>er
>> >>>>>>>>si
>> >> >>>>>>on
>> >> >>>>>>Mi
>> >> >>>>>>smatch):
>> >> >>>>>> Server IPC version 9 cannot communicate with client version 4
>> >> >>>>>>         at org.apache.hadoop.ipc.Client.call(Client.java:1113)
>> >> >>>>>>         at
>>org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229)
>> >> >>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown
>> >>Source)
>> >> >>>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> >> >>>>>>Method)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorI
>>>>>>>>>>mp
>> >>>>>>>>l.
>> >> >>>>>>ja
>> >> >>>>>>va
>> >> >>>>>>:62)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodA
>>>>>>>>>>cc
>> >>>>>>>>es
>> >> >>>>>>so
>> >> >>>>>>rI
>> >> >>>>>>mpl.java:43)
>> >> >>>>>>         at java.lang.reflect.Method.invoke(Method.java:497)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(Re
>>>>>>>>>>tr
>> >>>>>>>>yI
>> >> >>>>>>nv
>> >> >>>>>>oc
>> >> >>>>>>ationHandler.java:85)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInv
>>>>>>>>>>oc
>> >>>>>>>>at
>> >> >>>>>>io
>> >> >>>>>>nH
>> >> >>>>>>andler.java:62)
>> >> >>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown
>> >>Source)
>> >> >>>>>>         at
>>org.apache.hadoop.ipc.RPC.checkVersion(RPC.java:422)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>org.apache.hadoop.hdfs.DFSClient.createNamenode(DFSClient.java:183)
>> >> >>>>>>         at
>> >> >>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:281)
>> >> >>>>>>         at
>> >> >>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.hadoop.hdfs.DistributedFileSystem.initialize(Distribut
>>>>>>>>>>ed
>> >>>>>>>>Fi
>> >> >>>>>>le
>> >> >>>>>>Sy
>> >> >>>>>>stem.java:100)
>> >> >>>>>>         at
>> >> >>>>>>
>> >>
>> 
>>>>>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(Hado
>>>>>>>>>>op
>> >>>>>>>>Fi
>> >> >>>>>>le
>> >> >>>>>>Sy
>> >> >>>>>>stem.java:321)
>> >> >>>>>>         ... 8 more
>> >> >>>>>>
>> >> >>>>>> I searched for this error online and it indicates that the
>>client
>> >> >>>>>>which
>> >> >>>>>>is
>> >> >>>>>> Flink in this case is at a much lower version. Is there a way
>>to
>> >> >>>>>>check
>> >> >>>>>>the
>> >> >>>>>> version of Hadoop packaged with my Flink installation?
>> >> >>>>>>
>> >> >>>>>> Thanks,
>> >> >>>>>> Ali
>> >> >>>>>>
>> >> >>>>
>> >> >>
>> >>
>> >>
>>
>>

Reply via email to