Thank You Piotre. I moved *flink-s3-fs-hadoop* library to plugin. Now it's good.
On Wed, Oct 14, 2020 at 6:23 AM Piotr Nowojski <pnowoj...@apache.org> wrote: > Hi, > > Are you sure you are loading the filesystems correctly? Are you using the > plugin mechanism? [1] Since Flink 1.10 plugins can only be loaded in this > way [2], while there were some changes to plug some holes in Flink 1.11 [3]. > > Best, > Piotrek > > [1] > https://ci.apache.org/projects/flink/flink-docs-stable/ops/plugins.html > [2] FileSystems should be loaded via Plugin Architecture ( > https://ci.apache.org/projects/flink/flink-docs-release-1.10/release-notes/flink-1.10.html > ) > [3] Refined fallback filesystems to only handle specific filesystems ( > https://ci.apache.org/projects/flink/flink-docs-stable/release-notes/flink-1.11.html > ) > > śr., 14 paź 2020 o 06:49 Vijayendra Yadav <contact....@gmail.com> > napisał(a): > >> Hi Team, >> >> I have upgraded to Flink 1.11 (EMR 5.31) from Flink 1.10 (5.30.1). >> >> I am facing following Error while running *flink streaming *Job from >> command line. >> run command like: */usr/lib/flink/bin/flink run* >> >> *What dependency I might be missing or conflicting ?* >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> >> *04:46:51.669 [main] ERROR org.apache.flink.client.cli.CliFrontend - >> Fatal error while running command line >> interface.java.lang.NoSuchMethodError: >> org.apache.hadoop.ipc.RPC.getProtocolProxy(Ljava/lang/Class;JLjava/net/InetSocketAddress;Lorg/apache/hadoop/security/UserGroupInformation;Lorg/apache/hadoop/conf/Configuration;Ljavax/net/SocketFactory;ILorg/apache/hadoop/io/retry/RetryPolicy;Ljava/util/concurrent/atomic/AtomicBoolean;Lorg/apache/hadoop/ipc/AlignmentContext;)Lorg/apache/hadoop/ipc/ProtocolProxy; >> at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithAlignmentContext(NameNodeProxiesClient.java:400) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createNonHAProxyWithClientProtocol(NameNodeProxiesClient.java:351) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithClientProtocol(NameNodeProxiesClient.java:143) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:353) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:287) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:159) >> ~[hadoop-hdfs-client-2.10.0-amzn-0.jar:?] at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3354) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:226) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.flink.yarn.YarnClusterDescriptor.startAppMaster(YarnClusterDescriptor.java:661) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.yarn.YarnClusterDescriptor.deployInternal(YarnClusterDescriptor.java:524) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:424) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.deployment.executors.AbstractJobClusterExecutor.execute(AbstractJobClusterExecutor.java:70) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1812) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:128) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1785) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.scala:752) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> com.att.vdcs.StreamingJobKafkaFlink$.main(StreamingJobKafkaFlink.scala:196) >> ~[?:?] at >> com.att.vdcs.StreamingJobKafkaFlink.main(StreamingJobKafkaFlink.scala) >> ~[?:?] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native >> Method) ~[?:1.8.0_265] at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> ~[?:1.8.0_265] at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> ~[?:1.8.0_265] at java.lang.reflect.Method.invoke(Method.java:498) >> ~[?:1.8.0_265] at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:288) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:198) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:149) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:699) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:232) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:916) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:992) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> java.security.AccessController.doPrivileged(Native Method) ~[?:1.8.0_265] >> at javax.security.auth.Subject.doAs(Subject.java:422) ~[?:1.8.0_265] >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) >> ~[flink-s3-fs-hadoop-1.11.0.jar:1.11.0] at >> org.apache.flink.runtime.security.contexts.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >> ~[flink-dist_2.11-1.11.0.jar:1.11.0] at >> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:992) >> [flink-dist_2.11-1.11.0.jar:1.11.0]java.lang.NoSuchMethodError: >> org.apache.hadoop.ipc.RPC.getProtocolProxy(Ljava/lang/Class;JLjava/net/InetSocketAddress;Lorg/apache/hadoop/security/UserGroupInformation;Lorg/apache/hadoop/conf/Configuration;Ljavax/net/SocketFactory;ILorg/apache/hadoop/io/retry/RetryPolicy;Ljava/util/concurrent/atomic/AtomicBoolean;Lorg/apache/hadoop/ipc/AlignmentContext;)Lorg/apache/hadoop/ipc/ProtocolProxy; >> at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithAlignmentContext(NameNodeProxiesClient.java:400) >> at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createNonHAProxyWithClientProtocol(NameNodeProxiesClient.java:351) >> at >> org.apache.hadoop.hdfs.NameNodeProxiesClient.createProxyWithClientProtocol(NameNodeProxiesClient.java:143) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:353) >> at org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:287) at >> org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:159) >> at >> org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:3354) >> at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:124) >> at org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:3403) >> at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:3371) >> at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:477) at >> org.apache.hadoop.fs.FileSystem.get(FileSystem.java:226) at >> org.apache.flink.yarn.YarnClusterDescriptor.startAppMaster(YarnClusterDescriptor.java:661) >> at >> org.apache.flink.yarn.YarnClusterDescriptor.deployInternal(YarnClusterDescriptor.java:524) >> at >> org.apache.flink.yarn.YarnClusterDescriptor.deployJobCluster(YarnClusterDescriptor.java:424) >> at >> org.apache.flink.client.deployment.executors.AbstractJobClusterExecutor.execute(AbstractJobClusterExecutor.java:70) >> at >> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1812) >> at >> org.apache.flink.client.program.StreamContextEnvironment.executeAsync(StreamContextEnvironment.java:128) >> at >> org.apache.flink.streaming.api.environment.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.java:1785) >> at >> org.apache.flink.streaming.api.scala.StreamExecutionEnvironment.executeAsync(StreamExecutionEnvironment.scala:752) >> at >> com.att.vdcs.StreamingJobKafkaFlink$.main(StreamingJobKafkaFlink.scala:196) >> at >> com.att.vdcs.StreamingJobKafkaFlink.main(StreamingJobKafkaFlink.scala) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:498) at >> org.apache.flink.client.program.PackagedProgram.callMainMethod(PackagedProgram.java:288) >> at >> org.apache.flink.client.program.PackagedProgram.invokeInteractiveModeForExecution(PackagedProgram.java:198) >> at >> org.apache.flink.client.ClientUtils.executeProgram(ClientUtils.java:149) >> at >> org.apache.flink.client.cli.CliFrontend.executeProgram(CliFrontend.java:699) >> at org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:232) >> at >> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:916) >> at >> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:992) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:422) at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) >> at >> org.apache.flink.runtime.security.contexts.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41) >> at org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:992)* >> >> >> Regards, >> Vijay >> >