Flink 1.11 will support Hadoop 3. EMR 6 requires Hadoop 3, that's why Flink was not included anymore. Amazon will add Flink back to EMR 6.0 soon.
On Thu, May 14, 2020 at 7:11 PM aj <ajainje...@gmail.com> wrote: > Hi Yang, > > I am able to resolve the issue by removing Hadoop dependency as you > mentioned. > > 1. Removed hadoop-common dependency and > > <groupId>org.apache.flink</groupId> > <artifactId>flink-streaming-java_${scala.binary.version}</artifactId> > <version>${flink.version}</version> > <exclusions> > <exclusion> > <groupId>org.apache.flink</groupId> > <artifactId>flink-hadoop-fs</artifactId> > </exclusion> > </exclusions> > > </dependency> > > > 2. After the above changes, I am able to submit job on yarn but facing > issues with s3 plugin that I resolved by putting * > flink-s3-fs-hadoop-1.9.0.jar *i n the plugins/s3-fs-hadoop directory. > > Thanks for your support. > > Any update when will flink.10 officially supported in EMR. Even in new > EMR version(emr 6.0) flink has been removed. > > > > On Sat, May 9, 2020 at 1:36 PM aj <ajainje...@gmail.com> wrote: > >> Hello Yang, >> >> I have attached my pom file and I did not see that I am using any Hadoop >> dependency. Can you please help me. >> >> On Wed, May 6, 2020 at 1:22 PM Yang Wang <danrtsey...@gmail.com> wrote: >> >>> Hi aj, >>> >>> From the logs you have provided, the hadoop version is still 2.4.1. >>> Could you check the user jar(i.e. events-processor-1.0-SNAPSHOT.jar) >>> have some >>> hadoop classes? If it is, you need to exclude the hadoop dependency. >>> >>> >>> Best, >>> Yang >>> >>> aj <ajainje...@gmail.com> 于2020年5月6日周三 下午3:38写道: >>> >>>> Hello, >>>> >>>> Please help me upgrade to 1.10 in AWS EMR. >>>> >>>> On Fri, May 1, 2020 at 4:05 PM aj <ajainje...@gmail.com> wrote: >>>> >>>>> Hi Yang, >>>>> >>>>> I am attaching the logs for your reference, please help me what i am >>>>> doing wrong. >>>>> >>>>> Thanks, >>>>> Anuj >>>>> >>>>> On Wed, Apr 29, 2020 at 9:06 AM Yang Wang <danrtsey...@gmail.com> >>>>> wrote: >>>>> >>>>>> Hi Anuj, >>>>>> >>>>>> I think the exception you come across still because the hadoop version >>>>>> is 2.4.1. I have checked the hadoop code, the code line are exactly >>>>>> same. >>>>>> For 2.8.1, i also have checked the ruleParse. It could work. >>>>>> >>>>>> /** >>>>>> * A pattern for parsing a auth_to_local rule. >>>>>> */ >>>>>> private static final Pattern ruleParser = >>>>>> >>>>>> Pattern.compile("\\s*((DEFAULT)|(RULE:\\[(\\d*):([^\\]]*)](\\(([^)]*)\\))?"+ >>>>>> "(s/([^/]*)/([^/]*)/(g)?)?))/?(L)?"); >>>>>> >>>>>> >>>>>> Could you share the jobmanager logs so that i could check the >>>>>> classpath and hadoop version? >>>>>> >>>>>> Best, >>>>>> Yang >>>>>> >>>>>> aj <ajainje...@gmail.com> 于2020年4月28日周二 上午1:01写道: >>>>>> >>>>>>> Hello Yang, >>>>>>> My Hadoop version is Hadoop 3.2.1-amzn-0 >>>>>>> and I have put in flink/lib. >>>>>>> flink-shaded-hadoop-2-uber-2.8.3-10.0.jar >>>>>>> >>>>>>> then I am getting below error : >>>>>>> >>>>>>> SLF4J: Class path contains multiple SLF4J bindings. >>>>>>> SLF4J: Found binding in >>>>>>> [jar:file:/mnt/yarn/usercache/hadoop/appcache/application_1587983834922_0002/filecache/10/slf4j-log4j12-1.7.15.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>> SLF4J: Found binding in >>>>>>> [jar:file:/usr/lib/hadoop/lib/slf4j-log4j12-1.7.25.jar!/org/slf4j/impl/StaticLoggerBinder.class] >>>>>>> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >>>>>>> explanation. >>>>>>> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] >>>>>>> Exception in thread "main" java.lang.IllegalArgumentException: >>>>>>> Invalid rule: /L >>>>>>> RULE:[2:$1@$0](.*@)s/@.*///L >>>>>>> DEFAULT >>>>>>> at >>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.parseRules(KerberosName.java:321) >>>>>>> at >>>>>>> org.apache.hadoop.security.authentication.util.KerberosName.setRules(KerberosName.java:386) >>>>>>> at >>>>>>> org.apache.hadoop.security.HadoopKerberosName.setConfiguration(HadoopKerberosName.java:75) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.initialize(UserGroupInformation.java:247) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.ensureInitialized(UserGroupInformation.java:232) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.loginUserFromSubject(UserGroupInformation.java:718) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.getLoginUser(UserGroupInformation.java:703) >>>>>>> at >>>>>>> org.apache.hadoop.security.UserGroupInformation.getCurrentUser(UserGroupInformation.java:605) >>>>>>> at >>>>>>> org.apache.flink.yarn.entrypoint.YarnEntrypointUtils.logYarnEnvironmentInformation(YarnEntrypointUtils.java:136) >>>>>>> at >>>>>>> org.apache.flink.yarn.entrypoint.YarnJobClusterEntrypoint.main(YarnJobClusterEntrypoint.java:109) >>>>>>> >>>>>>> >>>>>>> if I remove the flink-shaded-hadoop-2-uber-2.8.3-10.0.jar from >>>>>>> lib then i get below error: >>>>>>> >>>>>>> 2020-04-27 16:59:37,293 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend - >>>>>>> Classpath: >>>>>>> /usr/lib/flink/lib/flink-table-blink_2.11-1.10.0.jar:/usr/lib/flink/lib/flink-table_2.11-1.10.0.jar:/usr/lib/flink/lib/log4j-1.2.17.jar:/usr/lib/flink/lib/slf4j-log4j12-1.7.15.jar:/usr/lib/flink/lib/flink-dist_2.11-1.10.0.jar::/etc/hadoop/conf:/etc/hadoop/conf >>>>>>> 2020-04-27 16:59:37,293 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend - >>>>>>> -------------------------------------------------------------------------------- >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: jobmanager.heap.size, 1024m >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: taskmanager.memory.process.size, 1568m >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: taskmanager.numberOfTaskSlots, 1 >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: parallelism.default, 1 >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: env.yarn.conf.dir, /etc/hadoop/conf >>>>>>> 2020-04-27 16:59:37,300 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: env.hadoop.conf.dir, /etc/hadoop/conf >>>>>>> 2020-04-27 16:59:37,301 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: jobmanager.execution.failover-strategy, region >>>>>>> 2020-04-27 16:59:37,301 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: classloader.resolve-order, parent-first >>>>>>> 2020-04-27 16:59:37,301 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: s3.access-key, AKIA52DD5QA5FC7HPKXG >>>>>>> 2020-04-27 16:59:37,301 INFO >>>>>>> org.apache.flink.configuration.GlobalConfiguration - Loading >>>>>>> configuration property: s3.secret-key, ****** >>>>>>> 2020-04-27 16:59:37,305 WARN >>>>>>> org.apache.flink.client.cli.CliFrontend - Could >>>>>>> not >>>>>>> load CLI class org.apache.flink.yarn.cli.FlinkYarnSessionCli. >>>>>>> java.lang.NoClassDefFoundError: >>>>>>> org/apache/hadoop/yarn/exceptions/YarnException >>>>>>> at java.lang.Class.forName0(Native Method) >>>>>>> at java.lang.Class.forName(Class.java:264) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLine(CliFrontend.java:1076) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.loadCustomCommandLines(CliFrontend.java:1030) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:957) >>>>>>> Caused by: java.lang.ClassNotFoundException: >>>>>>> org.apache.hadoop.yarn.exceptions.YarnException >>>>>>> at java.net.URLClassLoader.findClass(URLClassLoader.java:382) >>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:419) >>>>>>> at >>>>>>> sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) >>>>>>> at java.lang.ClassLoader.loadClass(ClassLoader.java:352) >>>>>>> ... 5 more >>>>>>> 2020-04-27 16:59:37,406 INFO org.apache.flink.core.fs.FileSystem >>>>>>> - Hadoop is not in the classpath/dependencies. >>>>>>> The >>>>>>> extended set of supported File Systems via Hadoop is not available. >>>>>>> 2020-04-27 16:59:37,458 INFO >>>>>>> org.apache.flink.runtime.security.modules.HadoopModuleFactory - Cannot >>>>>>> create Hadoop Security Module because Hadoop cannot be found in the >>>>>>> Classpath. >>>>>>> 2020-04-27 16:59:37,476 INFO >>>>>>> org.apache.flink.runtime.security.modules.JaasModule - Jaas >>>>>>> file >>>>>>> will be created as /tmp/jaas-7054453135321774613.conf. >>>>>>> 2020-04-27 16:59:37,480 INFO >>>>>>> org.apache.flink.runtime.security.SecurityUtils - Cannot >>>>>>> install HadoopSecurityContext because Hadoop cannot be found in the >>>>>>> Classpath. >>>>>>> 2020-04-27 16:59:37,481 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend - Running >>>>>>> 'run' command. >>>>>>> 2020-04-27 16:59:37,488 INFO >>>>>>> org.apache.flink.client.cli.CliFrontend - >>>>>>> Building >>>>>>> program from JAR file >>>>>>> 2020-04-27 16:59:37,488 ERROR >>>>>>> org.apache.flink.client.cli.CliFrontend - Invalid >>>>>>> command line arguments. >>>>>>> org.apache.flink.client.cli.CliArgsException: Could not build the >>>>>>> program from JAR file. >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:203) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.parseParameters(CliFrontend.java:895) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.lambda$main$10(CliFrontend.java:968) >>>>>>> at >>>>>>> org.apache.flink.runtime.security.NoOpSecurityContext.runSecured(NoOpSecurityContext.java:30) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.main(CliFrontend.java:968) >>>>>>> Caused by: java.io.FileNotFoundException: JAR file does not exist: >>>>>>> -ynm >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.getJarFile(CliFrontend.java:719) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.buildProgram(CliFrontend.java:695) >>>>>>> at >>>>>>> org.apache.flink.client.cli.CliFrontend.run(CliFrontend.java:200) >>>>>>> ... 4 more >>>>>>> >>>>>>> >>>>>>> Can you please help. >>>>>>> >>>>>>> Thanks, >>>>>>> Anuj >>>>>>> >>>>>>> >>>>>>> On Mon, Apr 13, 2020 at 7:43 AM Yang Wang <danrtsey...@gmail.com> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Anuj, >>>>>>>> >>>>>>>> It seems that you are using hadoop version 2.4.1. I think "L" could >>>>>>>> not be supported in >>>>>>>> this version. Could you upgrade your hadoop version to 2.8 and have >>>>>>>> a try? If your >>>>>>>> YARN cluster version is 2.8+, then you could directly remove the >>>>>>>> flink-shaded-hadoop >>>>>>>> in your lib directory. Otherwise, you need to download the >>>>>>>> flink-shaded-hadoop with >>>>>>>> version 2.8 here[1]. >>>>>>>> >>>>>>>> >>>>>>>> [1]. https://flink.apache.org/downloads.html#additional-components >>>>>>>> >>>>>>>> Best, >>>>>>>> Yang >>>>>>>> >>>>>>>> aj <ajainje...@gmail.com> 于2020年4月11日周六 上午4:21写道: >>>>>>>> >>>>>>>>> Hi Robert, >>>>>>>>> attached the full application log file. >>>>>>>>> >>>>>>>>> Thanks, >>>>>>>>> Anuj >>>>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/> >>>>>>>>> >>>>>>>> >>>>>>> >>>>>>> -- >>>>>>> Thanks & Regards, >>>>>>> Anuj Jain >>>>>>> Mob. : +91- 8588817877 >>>>>>> Skype : anuj.jain07 >>>>>>> <http://www.oracle.com/> >>>>>>> >>>>>>> >>>>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/> >>>>>>> >>>>>> >>>>> >>>>> -- >>>>> Thanks & Regards, >>>>> Anuj Jain >>>>> Mob. : +91- 8588817877 >>>>> Skype : anuj.jain07 >>>>> <http://www.oracle.com/> >>>>> >>>>> >>>>> <http://www.cse.iitm.ac.in/%7Eanujjain/> >>>>> >>>> >>>> >>>> -- >>>> Thanks & Regards, >>>> Anuj Jain >>>> Mob. : +91- 8588817877 >>>> Skype : anuj.jain07 >>>> <http://www.oracle.com/> >>>> >>>> >>>> <http://www.cse.iitm.ac.in/%7Eanujjain/> >>>> >>> >> >> -- >> Thanks & Regards, >> Anuj Jain >> Mob. : +91- 8588817877 >> Skype : anuj.jain07 >> <http://www.oracle.com/> >> >> >> <http://www.cse.iitm.ac.in/%7Eanujjain/> >> > > > -- > Thanks & Regards, > Anuj Jain > Mob. : +91- 8588817877 > Skype : anuj.jain07 > <http://www.oracle.com/> > > > <http://www.cse.iitm.ac.in/%7Eanujjain/> >