Hi Max,

Both commands return nothing. Those variables aren’t set.

The only software I installed on these machines is Flink and Java.

-Ali

On 2015-12-28, 6:42 AM, "Maximilian Michels" <m...@apache.org> wrote:

>Hi Ali,
>
>The warning about the native Hadoop libraries is nothing to worry
>about. The native modules are platform-optimized modules which may be
>used to improve performance. They are not necessary for Hadoop to
>function correctly.
>
>The exception message implies that you are using is a very old version
>of Hadoop. Do you have other Hadoop versions installed on the same
>machine? We have had people using Flink 0.10.0 with Hadoop 2.6.0
>without any problems.
>
>On the cluster machines, what is the output of these commands?
>
>echo $HADOOP_CLASSPATH
>echo $HADOOP_CONF_DIR
>
>
>Thanks,
>Max
>
>On Wed, Dec 23, 2015 at 3:53 PM, Kashmar, Ali <ali.kash...@emc.com> wrote:
>> Hi Max,
>>
>> I have the same output for the Task Manager:
>>
>> 11:25:04,274 INFO  org.apache.flink.runtime.taskmanager.TaskManager
>>       -  Hadoop version: 2.6.0
>>
>> I do get this line at the beginning of both job and task manager log
>>files:
>>
>> 11:25:04,100 WARN  org.apache.hadoop.util.NativeCodeLoader
>>       - Unable to load native-hadoop library for your platform... using
>> builtin-java classes where applicable
>>
>> Do you think it has anything to do with it?
>>
>> Thanks,
>> Ali
>>
>> On 2015-12-23, 7:30 AM, "Maximilian Michels" <m...@apache.org> wrote:
>>
>>>Hi Ali,
>>>
>>>Could you please also post the Hadoop version output of the task
>>>manager log files? It looks like the task managers are running a
>>>different Hadoop version.
>>>
>>>Thanks,
>>>Max
>>>
>>>On Tue, Dec 22, 2015 at 4:28 PM, Kashmar, Ali <ali.kash...@emc.com>
>>>wrote:
>>>> Hi Robert,
>>>>
>>>> I found the version in the job manager log file:
>>>>
>>>> 17:33:49,636 INFO  org.apache.flink.runtime.jobmanager.JobManager
>>>>       -  Hadoop version: 2.6.0
>>>>
>>>> But the Hadoop installation I have is saying this:
>>>>
>>>> ubuntu@ubuntu-171:~/Documents/hadoop-2.6.0$ bin/hadoop version
>>>> Hadoop 2.6.0
>>>> Subversion https://git-wip-us.apache.org/repos/asf/hadoop.git -r
>>>> e3496499ecb8d220fba99dc5ed4c99c8f9e33bb1
>>>> Compiled by jenkins on 2014-11-13T21:10Z
>>>> Compiled with protoc 2.5.0
>>>> From source with checksum 18e43357c8f927c0695f1e9522859d6a
>>>> This command was run using
>>>>
>>>>/home/ubuntu/Documents/hadoop-2.6.0/share/hadoop/common/hadoop-common-2
>>>>.6
>>>>.0
>>>> .jar
>>>>
>>>>
>>>> So one of them is lying to me? :)
>>>>
>>>> Ali
>>>>
>>>> On 2015-12-22, 10:16 AM, "Robert Metzger" <rmetz...@apache.org> wrote:
>>>>
>>>>>Hi Ali,
>>>>>
>>>>>the TaskManagers and the JobManager is logging the Hadoop version on
>>>>>startup.
>>>>>
>>>>>On Tue, Dec 22, 2015 at 4:10 PM, Kashmar, Ali <ali.kash...@emc.com>
>>>>>wrote:
>>>>>
>>>>>> Hello,
>>>>>>
>>>>>> I¹m trying to use HDFS as store for Flink checkpoints so I
>>>>>>downloaded
>>>>>>the
>>>>>> Hadoop 2.6.0/Scala 2.10 version of Flink and installed it. I also
>>>>>> downloaded Hadoop 2.6.0 separately from the Hadoop website and set
>>>>>>up
>>>>>>HDFS
>>>>>> on a separate machine. When I start Flink I get the following error:
>>>>>>
>>>>>> 17:34:13,047 INFO  org.apache.flink.runtime.jobmanager.JobManager
>>>>>>       - Status of job 9ba32a08bc0ec02810bf5d2710842f72 (Protocol
>>>>>>Event
>>>>>> Processing) changed to FAILED.
>>>>>> java.lang.Exception: Call to registerInputOutput() of invokable
>>>>>>failed
>>>>>>         at
>>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:529)
>>>>>>         at java.lang.Thread.run(Thread.java:745)
>>>>>> Caused by: java.io.IOException: The given file URI (hdfs://
>>>>>> 10.13.182.171:9000/user/flink/checkpoints) points to the HDFS
>>>>>>NameNode
>>>>>>at
>>>>>> 10.13.182.171:9000, but the File System could not be initialized
>>>>>>with
>>>>>> that address: Server IPC version 9 cannot communicate with client
>>>>>>version 4
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(HadoopFi
>>>>>>le
>>>>>>Sy
>>>>>>stem.java:337)
>>>>>>         at
>>>>>>org.apache.flink.core.fs.FileSystem.get(FileSystem.java:253)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsSta
>>>>>>te
>>>>>>Ba
>>>>>>ckend.java:142)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsSta
>>>>>>te
>>>>>>Ba
>>>>>>ckend.java:101)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.runtime.state.filesystem.FsStateBackendFactory.creat
>>>>>>eF
>>>>>>ro
>>>>>>mConfig(FsStateBackendFactory.java:48)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.createStateBacken
>>>>>>d(
>>>>>>St
>>>>>>reamTask.java:517)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.streaming.runtime.tasks.StreamTask.registerInputOutp
>>>>>>ut
>>>>>>(S
>>>>>>treamTask.java:171)
>>>>>>         at
>>>>>>org.apache.flink.runtime.taskmanager.Task.run(Task.java:526)
>>>>>>         ... 1 more
>>>>>> Caused by:
>>>>>>
>>>>>>org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.RPC$Versi
>>>>>>on
>>>>>>Mi
>>>>>>smatch):
>>>>>> Server IPC version 9 cannot communicate with client version 4
>>>>>>         at org.apache.hadoop.ipc.Client.call(Client.java:1113)
>>>>>>         at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229)
>>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown Source)
>>>>>>         at sun.reflect.NativeMethodAccessorImpl.invoke0(Native
>>>>>>Method)
>>>>>>         at
>>>>>>
>>>>>>sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.
>>>>>>ja
>>>>>>va
>>>>>>:62)
>>>>>>         at
>>>>>>
>>>>>>sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces
>>>>>>so
>>>>>>rI
>>>>>>mpl.java:43)
>>>>>>         at java.lang.reflect.Method.invoke(Method.java:497)
>>>>>>         at
>>>>>>
>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryI
>>>>>>nv
>>>>>>oc
>>>>>>ationHandler.java:85)
>>>>>>         at
>>>>>>
>>>>>>org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocat
>>>>>>io
>>>>>>nH
>>>>>>andler.java:62)
>>>>>>         at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown Source)
>>>>>>         at org.apache.hadoop.ipc.RPC.checkVersion(RPC.java:422)
>>>>>>         at
>>>>>> org.apache.hadoop.hdfs.DFSClient.createNamenode(DFSClient.java:183)
>>>>>>         at 
>>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:281)
>>>>>>         at 
>>>>>>org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245)
>>>>>>         at
>>>>>>
>>>>>>org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFi
>>>>>>le
>>>>>>Sy
>>>>>>stem.java:100)
>>>>>>         at
>>>>>>
>>>>>>org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(HadoopFi
>>>>>>le
>>>>>>Sy
>>>>>>stem.java:321)
>>>>>>         ... 8 more
>>>>>>
>>>>>> I searched for this error online and it indicates that the client
>>>>>>which
>>>>>>is
>>>>>> Flink in this case is at a much lower version. Is there a way to
>>>>>>check
>>>>>>the
>>>>>> version of Hadoop packaged with my Flink installation?
>>>>>>
>>>>>> Thanks,
>>>>>> Ali
>>>>>>
>>>>
>>

Reply via email to