Hi Ali, the TaskManagers and the JobManager is logging the Hadoop version on startup.
On Tue, Dec 22, 2015 at 4:10 PM, Kashmar, Ali <ali.kash...@emc.com> wrote: > Hello, > > I’m trying to use HDFS as store for Flink checkpoints so I downloaded the > Hadoop 2.6.0/Scala 2.10 version of Flink and installed it. I also > downloaded Hadoop 2.6.0 separately from the Hadoop website and set up HDFS > on a separate machine. When I start Flink I get the following error: > > 17:34:13,047 INFO org.apache.flink.runtime.jobmanager.JobManager > - Status of job 9ba32a08bc0ec02810bf5d2710842f72 (Protocol Event > Processing) changed to FAILED. > java.lang.Exception: Call to registerInputOutput() of invokable failed > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:529) > at java.lang.Thread.run(Thread.java:745) > Caused by: java.io.IOException: The given file URI (hdfs:// > 10.13.182.171:9000/user/flink/checkpoints) points to the HDFS NameNode at > 10.13.182.171:9000, but the File System could not be initialized with > that address: Server IPC version 9 cannot communicate with client version 4 > at > org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(HadoopFileSystem.java:337) > at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:253) > at > org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsStateBackend.java:142) > at > org.apache.flink.runtime.state.filesystem.FsStateBackend.<init>(FsStateBackend.java:101) > at > org.apache.flink.runtime.state.filesystem.FsStateBackendFactory.createFromConfig(FsStateBackendFactory.java:48) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.createStateBackend(StreamTask.java:517) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.registerInputOutput(StreamTask.java:171) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:526) > ... 1 more > Caused by: > org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.RPC$VersionMismatch): > Server IPC version 9 cannot communicate with client version 4 > at org.apache.hadoop.ipc.Client.call(Client.java:1113) > at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229) > at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown Source) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:497) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:85) > at > org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:62) > at com.sun.proxy.$Proxy6.getProtocolVersion(Unknown Source) > at org.apache.hadoop.ipc.RPC.checkVersion(RPC.java:422) > at > org.apache.hadoop.hdfs.DFSClient.createNamenode(DFSClient.java:183) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:281) > at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:245) > at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:100) > at > org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.initialize(HadoopFileSystem.java:321) > ... 8 more > > I searched for this error online and it indicates that the client which is > Flink in this case is at a much lower version. Is there a way to check the > version of Hadoop packaged with my Flink installation? > > Thanks, > Ali >