For now, i have solved this issue by adding the following in filink config :
classloader.resolve-order: parent-first So it will ignore the duplicate classes from uber jar. I will work on the dependencies. One quick question I am using SBT for the building. Do you have any example sbt file for dependencies? I am bit confused. Should I set all others too "provided", So that won't be included in fat jar. But as the docs CEP etc. are not part of flink-dist. I haven't checked the content of flink-dist yet but just asking a quick question. val flinkDependencies = Seq( "org.slf4j" % "slf4j-log4j12" % "1.7.21", "org.apache.flink" %% "flink-scala" % flinkVersion % "provided", "org.apache.flink" %% "flink-streaming-scala" % flinkVersion % "provided", "org.apache.flink" %% "flink-table" % flinkVersion, "org.apache.flink" %% "flink-cep-scala" % flinkVersion, "org.apache.flink" %% "flink-connector-kafka-0.10" % flinkVersion, "org.apache.flink" %% "flink-connector-filesystem" % flinkVersion, "org.apache.flink" %% "flink-statebackend-rocksdb" % flinkVersion, "org.apache.flink" %% "flink-connector-cassandra" % flinkVersion, "org.apache.flink" % "flink-shaded-hadoop2" % flinkVersion, On Wed, Dec 20, 2017 at 7:48 PM, Timo Walther <twal...@apache.org> wrote: > Libraries such as CEP or Table API should have the "compile" scope and > should be in the both the fat and non-fat jar. > > The non-fat jar should contain everything that is not in flink-dist or > your lib directory. > > Regards, > Timo > > > Am 12/20/17 um 3:07 PM schrieb shashank agarwal: > > Hi, > > In that case, it won't find the dependencies. Cause I have other > dependencies also and what about CEP etc. cause that is not part of > flink-dist. > > Best > Shashank > > > > > > On Wed, Dec 20, 2017 at 3:16 PM, Aljoscha Krettek <aljos...@apache.org> > wrote: > >> Hi, >> >> That jar file looks like it has too much stuff in there that shouldn't be >> there. This can explain the errors you seeing because of classloading >> conflicts. >> >> Could you try not building a fat-jar and have only your code in your jar? >> >> Best, >> Aljoscha >> >> >> On 20. Dec 2017, at 10:15, shashank agarwal <shashank...@gmail.com> >> wrote: >> >> One more thing when i submit the job ir start yarn session it prints >> following logs : >> >> Using the result of 'hadoop classpath' to augment the Hadoop classpath: >> /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/ >> lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0.3- >> 8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/ >> hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop- >> yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6. >> 0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//* >> SLF4J: Class path contains multiple SLF4J bindings. >> SLF4J: Found binding in [jar:file:/opt/flink/lib/slf4j >> -log4j12-1.7.7.jar!/org/slf4j/impl/StaticLoggerBinder.class] >> SLF4J: Found binding in [jar:file:/usr/hdp/2.6.0.3-8/h >> adoop/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLog >> gerBinder.class] >> SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an >> explanation. >> SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory] >> >> >> So i think it's adding Hadoop libs in classpath too cause it's able to >> create the checkpointing directories from flink-conf file to HDFS. >> >> >> >> >> >> >> >> >> On Wed, Dec 20, 2017 at 2:31 PM, shashank agarwal <shashank...@gmail.com> >> wrote: >> >>> Hi, >>> >>> Please find attached list of jar file contents and flink/lib/ contents. >>> I have removed my class files list from jar list and I have >>> added flink-hadoop-compatibility_2.11-1.4.0.jar later in flink/lib/ but >>> no success. >>> >>> I have tried by removing flink-shaded-hadoop2 from my project but still >>> no success. >>> >>> >>> Thanks >>> Shashank >>> >>> >>> On Wed, Dec 20, 2017 at 2:14 PM, Aljoscha Krettek <aljos...@apache.org> >>> wrote: >>> >>>> Hi, >>>> >>>> Could you please list what exactly is in your submitted jar file, for >>>> example using "jar tf my-jar-file.jar"? And also what files exactly are in >>>> your Flink lib directory. >>>> >>>> Best, >>>> Aljoscha >>>> >>>> >>>> On 19. Dec 2017, at 20:10, shashank agarwal <shashank...@gmail.com> >>>> wrote: >>>> >>>> Hi Timo, >>>> >>>> I am using Rocksdbstatebackend with hdfs path. I have following flink >>>> dependencies in my sbt : >>>> >>>> "org.slf4j" % "slf4j-log4j12" % "1.7.21", >>>> "org.apache.flink" %% "flink-scala" % flinkVersion % "provided", >>>> "org.apache.flink" %% "flink-streaming-scala" % flinkVersion % >>>> "provided", >>>> "org.apache.flink" %% "flink-cep-scala" % flinkVersion, >>>> "org.apache.flink" %% "flink-connector-kafka-0.10" % flinkVersion, >>>> "org.apache.flink" %% "flink-connector-filesystem" % flinkVersion, >>>> "org.apache.flink" %% "flink-statebackend-rocksdb" % flinkVersion, >>>> "org.apache.flink" %% "flink-connector-cassandra" % "1.3.2", >>>> "org.apache.flink" % "flink-shaded-hadoop2" % flinkVersion, >>>> >>>> when i start flink yarn session it's working fine even it's creating >>>> flink checkpointing directory and copying libs into hdfs. >>>> >>>> But when I submit the application to this yarn session it prints >>>> following logs : >>>> >>>> Using the result of 'hadoop classpath' to augment the Hadoop classpath: >>>> /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0. >>>> >>>> <http://2.6.0.0/>3-8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop-yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//* >>>> >>>> But application fails contuniously with logs which i have sent earlier. >>>> >>>> >>>> >>>> I have tried to add flink- hadoop-compability*.jar as suggested by >>>> Jorn but it's not working. >>>> >>>> >>>> >>>> On Tue, Dec 19, 2017 at 5:08 PM, shashank agarwal < >>>> shashank...@gmail.com> wrote: >>>> >>>>> yes, it's working fine. now not getting compile time error. >>>>> >>>>> But when i trying to run this on cluster or yarn, getting following >>>>> runtime error : >>>>> >>>>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not >>>>> find a file system implementation for scheme 'hdfs'. The scheme is not >>>>> directly supported by Flink and no Hadoop file system to support this >>>>> scheme could be loaded. >>>>> at >>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:405) >>>>> at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:320) >>>>> at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293) >>>>> at >>>>> org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.<init>(FsCheckpointStreamFactory.java:99) >>>>> at >>>>> org.apache.flink.runtime.state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277) >>>>> at >>>>> org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createStreamFactory(RocksDBStateBackend.java:273) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.createCheckpointStreamFactory(StreamTask.java:787) >>>>> at >>>>> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:247) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:682) >>>>> at >>>>> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253) >>>>> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:718) >>>>> at java.lang.Thread.run(Thread.java:745) >>>>> Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: >>>>> Hadoop File System abstraction does not support scheme 'hdfs'. Either no >>>>> file system implementation exists for that scheme, or the relevant >>>>> classes are missing from the classpath. >>>>> at >>>>> org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:102) >>>>> at >>>>> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:401) >>>>> ... 12 more >>>>> Caused by: java.io.IOException: No FileSystem for scheme: hdfs >>>>> at >>>>> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2786) >>>>> at >>>>> org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:99) >>>>> ... 13 more >>>>> >>>>> while submitting job it's printing following logs so i think it's >>>>> including hdoop libs : >>>>> >>>>> Using the result of 'hadoop classpath' to augment the Hadoop classpath: >>>>> /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0. >>>>> >>>>> <http://2.6.0.0/>3-8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop-yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//* >>>>> >>>>> >>>>> On Fri, Dec 8, 2017 at 9:24 PM, shashank agarwal < >>>>> shashank...@gmail.com> wrote: >>>>> >>>>>> Sure i’ll Try that. Thanks >>>>>> >>>>>> On Fri, 8 Dec 2017 at 9:18 PM, Stephan Ewen <se...@apache.org> wrote: >>>>>> >>>>>>> I would recommend to add "flink-shaded-hadoop2". That is a bundle of >>>>>>> all Hadoop dependencies used by Flink. >>>>>>> >>>>>>> >>>>>>> On Fri, Dec 8, 2017 at 3:44 PM, Aljoscha Krettek < >>>>>>> aljos...@apache.org> wrote: >>>>>>> >>>>>>>> I see, thanks for letting us know! >>>>>>>> >>>>>>>> >>>>>>>> On 8. Dec 2017, at 15:42, shashank agarwal <shashank...@gmail.com> >>>>>>>> wrote: >>>>>>>> >>>>>>>> I had to include two dependencies. >>>>>>>> >>>>>>>> hadoop-hdfs (this for HDFS configuration) >>>>>>>> hadoop-common (this for Path) >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> On Fri, Dec 8, 2017 at 7:38 PM, Aljoscha Krettek < >>>>>>>> aljos...@apache.org> wrote: >>>>>>>> >>>>>>>>> I think hadoop-hdfs might be sufficient. >>>>>>>>> >>>>>>>>> >>>>>>>>> On 8. Dec 2017, at 14:48, shashank agarwal <shashank...@gmail.com> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>> Can you specifically guide which dependencies should I add to >>>>>>>>> extend this : >>>>>>>>> >>>>>>>>> https://github.com/apache/flink/blob/release-1.4.0-rc3/flink >>>>>>>>> -connectors/flink-connector-filesystem/src/main/java/org/apa >>>>>>>>> che/flink/streaming/connectors/fs/Bucketer.java >>>>>>>>> >>>>>>>>> is the Hadoop-core is sufficient. >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Fri, Dec 8, 2017 at 6:58 PM, shashank agarwal < >>>>>>>>> shashank...@gmail.com> wrote: >>>>>>>>> >>>>>>>>>> It's a compilation error. I think I have to include the Hadoop >>>>>>>>>> dependencies. >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Fri, Dec 8, 2017 at 6:54 PM, Aljoscha Krettek < >>>>>>>>>> aljos...@apache.org> wrote: >>>>>>>>>> >>>>>>>>>>> Hi, >>>>>>>>>>> >>>>>>>>>>> Is this a compilation error or at runtime. But in general, yes >>>>>>>>>>> you have to include the Hadoop dependencies if they're not there. >>>>>>>>>>> >>>>>>>>>>> Best, >>>>>>>>>>> Aljoscha >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On 8. Dec 2017, at 14:10, shashank agarwal < >>>>>>>>>>> shashank...@gmail.com> wrote: >>>>>>>>>>> >>>>>>>>>>> Hello, >>>>>>>>>>> >>>>>>>>>>> I am trying to test 1.4.0-RC3, Hadoop libraries removed in this >>>>>>>>>>> version. Actually, i have created custom Bucketer for the bucketing >>>>>>>>>>> sink. >>>>>>>>>>> I am extending >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> org.apache.flink.streaming.connectors.fs.bucketing.Bucketer >>>>>>>>>>> >>>>>>>>>>> in the class, i have to use org.apache.hadoop.fs.Path but as >>>>>>>>>>> hadoop libraries removed it's giving error >>>>>>>>>>> >>>>>>>>>>> "object hadoop is not a member of package org.apache" >>>>>>>>>>> >>>>>>>>>>> Should i have to include Hadoop client libs in >>>>>>>>>>> build.sbt dependencies. >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> Thanks >>>>>>>>>>> Shashank >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> -- >>>>>>>>>> Thanks Regards >>>>>>>>>> >>>>>>>>>> SHASHANK AGARWAL >>>>>>>>>> --- Trying to mobilize the things.... >>>>>>>>>> >>>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> Thanks Regards >>>>>>>>> >>>>>>>>> SHASHANK AGARWAL >>>>>>>>> --- Trying to mobilize the things.... >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> Thanks Regards >>>>>>>> >>>>>>>> SHASHANK AGARWAL >>>>>>>> --- Trying to mobilize the things.... >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>> -- >>>>>> Sent from iPhone 5 >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> Thanks Regards >>>>> >>>>> SHASHANK AGARWAL >>>>> --- Trying to mobilize the things.... >>>>> >>>>> >>>> >>>> >>>> -- >>>> Thanks Regards >>>> >>>> SHASHANK AGARWAL >>>> --- Trying to mobilize the things.... >>>> >>>> >>>> >>> >>> >>> -- >>> Thanks Regards >>> >>> SHASHANK AGARWAL >>> --- Trying to mobilize the things.... >>> >>> >> >> >> -- >> Thanks Regards >> >> SHASHANK AGARWAL >> --- Trying to mobilize the things.... >> >> >> > > > -- > Thanks Regards > > SHASHANK AGARWAL > --- Trying to mobilize the things.... > > > -- Thanks Regards SHASHANK AGARWAL --- Trying to mobilize the things....