Hi Robert,

On the ubuntu host, the port 9000 is up:

ubuntu@ubuntu-171:~$ netstat -tulpn
....
tcp        0      0 10.13.182.171:9000      0.0.0.0:*               LISTEN
     8849/java
....

And the process using this port is:

ubuntu@ubuntu-171:~$ ps -ef | grep 8849
ubuntu    8849     1  0  2015 ?        00:25:00
/usr/lib/jvm/java-8-oracle/bin/java -Dproc_namenode -Xmx1000m
-Djava.net.preferIPv4Stack=true
-Dhadoop.log.dir=/home/ubuntu/Documents/hadoop-2.6.0/logs
-Dhadoop.log.file=hadoop.log
-Dhadoop.home.dir=/home/ubuntu/Documents/hadoop-2.6.0
-Dhadoop.id.str=ubuntu -Dhadoop.root.logger=INFO,console
-Djava.library.path=/home/ubuntu/Documents/hadoop-2.6.0/lib/native
-Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true
-Djava.net.preferIPv4Stack=true -Djava.net.preferIPv4Stack=true
-Dhadoop.log.dir=/home/ubuntu/Documents/hadoop-2.6.0/logs
-Dhadoop.log.file=hadoop-ubuntu-namenode-ubuntu-171.log
-Dhadoop.home.dir=/home/ubuntu/Documents/hadoop-2.6.0
-Dhadoop.id.str=ubuntu -Dhadoop.root.logger=INFO,RFA
-Djava.library.path=/home/ubuntu/Documents/hadoop-2.6.0/lib/native
-Dhadoop.policy.file=hadoop-policy.xml -Djava.net.preferIPv4Stack=true
-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
-Dhadoop.security.logger=INFO,RFAS -Dhdfs.audit.logger=INFO,NullAppender
-Dhadoop.security.logger=INFO,RFAS
org.apache.hadoop.hdfs.server.namenode.NameNode

And the hadoop version is:

ubuntu@ubuntu-171:~$ Documents/hadoop-2.6.0/bin/hadoop version
Hadoop 2.6.0
Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r
e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
Compiled by jenkins on 2014-11-13T21:10Z
Compiled with protoc 2.5.0
From source with checksum 18e43357c8f927c0695f1e9522859d6a
This command was run using
/home/ubuntu/Documents/hadoop-2.6.0/share/hadoop/common/hadoop-common-2.6.0
.jar



-Ali



On 2016-01-04, 2:20 PM, "Robert Metzger" <rmetz...@apache.org> wrote:

>Your Flink installation has Hadoop 2.6.0 included, on the other machine,
>there is a Hadoop version installed, which is most likely a 1.x or even a
>0.x version.
>Are you sure that the host "ubuntu-171" has the ip 10.13.182.171 and that
>the hadoop installation in the "/home/ubuntu/Documents/hadoop-2.6.0/"
>directory is listening on port 9000 ?
>
>On Mon, Jan 4, 2016 at 3:03 PM, Kashmar, Ali <ali.kash...@emc.com> wrote:
>
>> Hi Max,
>>
>> Both commands return nothing. Those variables aren’t set.
>>
>> The only software I installed on these machines is Flink and Java.
>>
>> -Ali
>>
>> On 2015-12-28, 6:42 AM, "Maximilian Michels" <m...@apache.org> wrote:
>>
>> >Hi Ali,
>> >
>> >The warning about the native Hadoop libraries is nothing to worry
>> >about. The native modules are platform-optimized modules which may be
>> >used to improve performance. They are not necessary for Hadoop to
>> >function correctly.
>> >
>> >The exception message implies that you are using is a very old version
>> >of Hadoop. Do you have other Hadoop versions installed on the same
>> >machine? We have had people using Flink 0.10.0 with Hadoop 2.6.0
>> >without any problems.
>> >
>> >On the cluster machines, what is the output of these commands?
>> >
>> >echo $HADOOP_CLASSPATH
>> >echo $HADOOP_CONF_DIR
>> >
>> >
>> >Thanks,
>> >Max
>> >
>> >On Wed, Dec 23, 2015 at 3:53 PM, Kashmar, Ali <ali.kash...@emc.com>
>> wrote:
>> >> Hi Max,
>> >>
>> >> I have the same output for the Task Manager:
>> >>
>> >> 11:25:04,274 INFO  org.apache.flink.runtime.taskmanager.TaskManager
>> >>       -  Hadoop version: 2.6.0
>> >>
>> >> I do get this line at the beginning of both job and task manager log
>> >>files:
>> >>
>> >> 11:25:04,100 WARN  org.apache.hadoop.util.NativeCodeLoader
>> >>       - Unable to load native-hadoop library for your platform...
>>using
>> >> builtin-java classes where applicable
>> >>
>> >> Do you think it has anything to do with it?
>> >>
>> >> Thanks,
>> >> Ali
>> >>
>> >> On 2015-12-23, 7:30 AM, "Maximilian Michels" <m...@apache.org> wrote:
>> >>
>> >>>Hi Ali,
>> >>>
>> >>>Could you please also post the Hadoop version output of the task
>> >>>manager log files? It looks like the task managers are running a
>> >>>different Hadoop version.
>> >>>
>> >>>Thanks,
>> >>>Max
>> >>>
>> >>>On Tue, Dec 22, 2015 at 4:28 PM, Kashmar, Ali <ali.kash...@emc.com>
>> >>>wrote:
>> >>>> Hi Robert,
>> >>>>
>> >>>> I found the version in the job manager log file:
>> >>>>
>> >>>> 17:33:49,636 INFO  org.apache.flink.runtime.jobmanager.JobManager
>> >>>>       -  Hadoop version: 2.6.0
>> >>>>
>> >>>> But the Hadoop installation I have is saying this:
>> >>>>
>> >>>> ubuntu@ubuntu-171:~/Documents/hadoop-2.6.0$ bin/hadoop version
>> >>>> Hadoop 2.6.0
>> >>>> Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r
>> >>>> e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
>> >>>> Compiled by jenkins on 2014-11-13T21:10Z
>> >>>> Compiled with protoc 2.5.0
>> >>>> From source with checksum 18e43357c8f927c0695f1e9522859d6a
>> >>>> This command was run using
>> >>>>
>> 
>>>>>>/home/ubuntu/Documents/hadoop-2.6.0/share/hadoop/common/hadoop-common
>>>>>>-2
>> >>>>.6
>> >>>>.0
>> >>>> .jar
>> >>>>
>> >>>>
>> >>>> So one of them is lying to me? :)
>> >>>>
>> >>>> Ali
>> >>>>
>> >>>> On 2015-12-22, 10:16 AM, "Robert Metzger" <rmetz...@apache.org>
>> wrote:
>> >>>>
>> >>>>>Hi Ali,
>> >>>>>
>> >>>>>the TaskManagers and the JobManager is logging the Hadoop version
>>on
>> >>>>>startup.
>> >>>>>
>> >>>>>On Tue, Dec 22, 2015 at 4:10 PM, Kashmar, Ali <ali.kash...@emc.com>
>> >>>>>wrote:
>> >>>>>
>> >>>>>> Hello,
>> >>>>>>
>> >>>>>> I¹m trying to use HDFS as store for Flink checkpoints so I
>> >>>>>>downloaded
>> >>>>>>the
>> >>>>>> Hadoop 2.6.0/Scala 2.10 version of Flink and installed it. I also
>> >>>>>> downloaded Hadoop 2.6.0 separately from the Hadoop website and
>>set
>> >>>>>>up
>> >>>>>>HDFS
>> >>>>>> on a separate machine. When I start Flink I get the following
>>error:
>> >>>>>>
>> >>>>>> 17:34:13,047 INFO  org.apache.flink.runtime.jobmanager.JobManager
>> >>>>>>       - Status of job 9ba32a08bc0ec02810bf5d2710842f72 (Protocol
>> >>>>>>Event
>> >>>>>> Processing) changed to FAILED.
>> >>>>>> java.lang.Exception: Call to registerInputOutput() of invokable
>> >>>>>>failed
>> >>>>>>         at
>> >>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:529)
>> >>>>>>         at java.lang.Thread.run(Thread.java:745)
>> >>>>>> Caused by: java.io.IOException: The given file URI (hdfs://
>> >>>>>> 10.13.182.171:9000/user/flink/checkpoints) points to the HDFS
>> >>>>>>NameNode
>> >>>>>>at
>> >>>>>> 10.13.182.171:9000, but the File System could not be initialized
>> >>>>>>with
>> >>>>>> that address: Server IPC version 9 cannot communicate with client
>> >>>>>>version 4
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(Hadoop
>>>>>>>>Fi
>> >>>>>>le
>> >>>>>>Sy
>> >>>>>>stem.java:337)
>> >>>>>>         at
>> >>>>>>org.apache.flink.core.fs.FileSystem.get(FileSystem.java:253)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsS
>>>>>>>>ta
>> >>>>>>te
>> >>>>>>Ba
>> >>>>>>ckend.java:142)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsS
>>>>>>>>ta
>> >>>>>>te
>> >>>>>>Ba
>> >>>>>>ckend.java:101)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackendFactory.cre
>>>>>>>>at
>> >>>>>>eF
>> >>>>>>ro
>> >>>>>>mConfig(FsStateBackendFactory.java:48)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.createStateBack
>>>>>>>>en
>> >>>>>>d(
>> >>>>>>St
>> >>>>>>reamTask.java:517)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.registerInputOu
>>>>>>>>tp
>> >>>>>>ut
>> >>>>>>(S
>> >>>>>>treamTask.java:171)
>> >>>>>>         at
>> >>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:526)
>> >>>>>>         ... 1 more
>> >>>>>> Caused by:
>> >>>>>>
>> 
>>>>>>>>org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.RPC$Ver
>>>>>>>>si
>> >>>>>>on
>> >>>>>>Mi
>> >>>>>>smatch):
>> >>>>>> Server IPC version 9 cannot communicate with client version 4
>> >>>>>>         at org.apache.hadoop.ipc.Client.call(Client.java:1113)
>> >>>>>>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229)
>> >>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown
>>Source)
>> >>>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>> >>>>>>Method)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImp
>>>>>>>>l.
>> >>>>>>ja
>> >>>>>>va
>> >>>>>>:62)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcc
>>>>>>>>es
>> >>>>>>so
>> >>>>>>rI
>> >>>>>>mpl.java:43)
>> >>>>>>         at java.lang.reflect.Method.invoke(Method.java:497)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(Retr
>>>>>>>>yI
>> >>>>>>nv
>> >>>>>>oc
>> >>>>>>ationHandler.java:85)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvoc
>>>>>>>>at
>> >>>>>>io
>> >>>>>>nH
>> >>>>>>andler.java:62)
>> >>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown
>>Source)
>> >>>>>>         at org.apache.hadoop.ipc.RPC.checkVersion(RPC.java:422)
>> >>>>>>         at
>> >>>>>> 
>>org.apache.hadoop.hdfs.DFSClient.createNamenode(DFSClient.java:183)
>> >>>>>>         at
>> >>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:281)
>> >>>>>>         at
>> >>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.hadoop.hdfs.DistributedFileSystem.initialize(Distributed
>>>>>>>>Fi
>> >>>>>>le
>> >>>>>>Sy
>> >>>>>>stem.java:100)
>> >>>>>>         at
>> >>>>>>
>> 
>>>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(Hadoop
>>>>>>>>Fi
>> >>>>>>le
>> >>>>>>Sy
>> >>>>>>stem.java:321)
>> >>>>>>         ... 8 more
>> >>>>>>
>> >>>>>> I searched for this error online and it indicates that the client
>> >>>>>>which
>> >>>>>>is
>> >>>>>> Flink in this case is at a much lower version. Is there a way to
>> >>>>>>check
>> >>>>>>the
>> >>>>>> version of Hadoop packaged with my Flink installation?
>> >>>>>>
>> >>>>>> Thanks,
>> >>>>>> Ali
>> >>>>>>
>> >>>>
>> >>
>>
>>

Reply via email to