Hi, What is the exact problem you're seeing? Could you please post your POM file and also a listing of all the files in your user jar?
Best, Aljoscha > On 15. Feb 2018, at 11:12, Or Sher <or.sh...@gmail.com> wrote: > > Hi, > Did you ever get to solve this issue? I'm getting the same error. > > On 1.3.2 I used to run the fat jar as a standalone without any job submission > and it worked just fine > It looked like it used an embedded MiniCluster. > > After just changing the dependencies to 1.4.0 we started getting this errors. > > Does this execution type should also be supported in 1.4.0? > Couldn't find anything in the docs about it. > It always says to run the job using "flink run" which depends on an already > running cluster. > > On Fri, Jan 26, 2018 at 12:59 PM Aljoscha Krettek <aljos...@apache.org > <mailto:aljos...@apache.org>> wrote: > Hi, > > It seems you are using Akka in your user program and the reference.conf in > your Jar is clashing with the same File from Flink. What is the reason for > having Akka (or reference.conf) in your user jar? > > Best, > Aljoscha > > >> On 22. Jan 2018, at 11:09, 韩宁宁 <453673...@qq.com <mailto:453673...@qq.com>> >> wrote: >> >> Dear All >> I have a question about Flink&Hadoop. >> I want to read the files on HDFS by flink,but I encountered an error >> as follows,can you please advise the solution about this problem. It will be >> much appreciated.: >> SLF4J: Failed to load class "org.slf4j.impl.StaticLoggerBinder". >> SLF4J: Defaulting to no-operation (NOP) logger implementation >> SLF4J: See http://www.slf4j.org/codes.html#StaticLoggerBinder >> <http://www.slf4j.org/codes.html#StaticLoggerBinder> for further details. >> Exception in thread "main" >> com.typesafe.config.ConfigException$UnresolvedSubstitution: reference.conf @ >> jar:file:/data/home/fbi/hanningning/flink-hdfs/target/flink-hdfs.jar!/reference.conf: >> 804: Could not resolve substitution to a value: ${akka.stream.materializer} >> at >> com.typesafe.config.impl.ConfigReference.resolveSubstitutions(ConfigReference.java:108) >> at >> com.typesafe.config.impl.ResolveContext.realResolve(ResolveContext.java:179) >> at >> com.typesafe.config.impl.ResolveContext.resolve(ResolveContext.java:142) >> at >> com.typesafe.config.impl.SimpleConfigObject$ResolveModifier.modifyChildMayThrow(SimpleConfigObject.java:379) >> at >> com.typesafe.config.impl.SimpleConfigObject.modifyMayThrow(SimpleConfigObject.java:312) >> at >> com.typesafe.config.impl.SimpleConfigObject.resolveSubstitutions(SimpleConfigObject.java:398) >> at >> com.typesafe.config.impl.ResolveContext.realResolve(ResolveContext.java:179) >> at >> com.typesafe.config.impl.ResolveContext.resolve(ResolveContext.java:142) >> at >> com.typesafe.config.impl.SimpleConfigObject$ResolveModifier.modifyChildMayThrow(SimpleConfigObject.java:379) >> at >> com.typesafe.config.impl.SimpleConfigObject.modifyMayThrow(SimpleConfigObject.java:312) >> at >> com.typesafe.config.impl.SimpleConfigObject.resolveSubstitutions(SimpleConfigObject.java:398) >> at >> com.typesafe.config.impl.ResolveContext.realResolve(ResolveContext.java:179) >> at >> com.typesafe.config.impl.ResolveContext.resolve(ResolveContext.java:142) >> at >> com.typesafe.config.impl.SimpleConfigObject$ResolveModifier.modifyChildMayThrow(SimpleConfigObject.java:379) >> at >> com.typesafe.config.impl.SimpleConfigObject.modifyMayThrow(SimpleConfigObject.java:312) >> at >> com.typesafe.config.impl.SimpleConfigObject.resolveSubstitutions(SimpleConfigObject.java:398) >> at >> com.typesafe.config.impl.ResolveContext.realResolve(ResolveContext.java:179) >> at >> com.typesafe.config.impl.ResolveContext.resolve(ResolveContext.java:142) >> at >> com.typesafe.config.impl.SimpleConfigObject$ResolveModifier.modifyChildMayThrow(SimpleConfigObject.java:379) >> at >> com.typesafe.config.impl.SimpleConfigObject.modifyMayThrow(SimpleConfigObject.java:312) >> at >> com.typesafe.config.impl.SimpleConfigObject.resolveSubstitutions(SimpleConfigObject.java:398) >> at >> com.typesafe.config.impl.ResolveContext.realResolve(ResolveContext.java:179) >> at >> com.typesafe.config.impl.ResolveContext.resolve(ResolveContext.java:142) >> at >> com.typesafe.config.impl.SimpleConfigObject$ResolveModifier.modifyChildMayThrow(SimpleConfigObject.java:379) >> ..... >> >> >> ====================my code=========================== >> public class App { >> >> public static void main(String[] args) throws Exception { >> >> final String inputPath = args[0]//hdfs file path; >> final ExecutionEnvironment env = >> ExecutionEnvironment.getExecutionEnvironment(); >> >> HadoopInputFormat<LongWritable,Text> hadoopInputFormat = >> new HadoopInputFormat<LongWritable, Text>(new >> TextInputFormat(),LongWritable.class, >> Text.class,new JobConf()); >> TextInputFormat.addInputPath(hadoopInputFormat.getJobConf(), new >> Path(inputPath)); >> >> DataSet<Tuple2<LongWritable, Text>> text = >> env.createInput(hadoopInputFormat); >> >> text.print(); >> >> env.execute("read hdfs by flink test"); >> } >> >> } >> ==========================maven >> dependencies======================================= >> <properties> >> <flink.version>1.4.0</flink.version> >> <log4j.version>1.2.17</log4j.version> >> <akka.version>2.4.20</akka.version> >> </properties> >> >> <dependencies> >> <dependency> >> <groupId>org.apache.flink</groupId> >> <artifactId>flink-java</artifactId> >> <version>${flink.version}</version> >> </dependency> >> <dependency> >> <groupId>org.apache.flink</groupId> >> <artifactId>flink-clients_2.11</artifactId> >> <version>${flink.version}</version> >> </dependency> >> <dependency> >> <groupId>org.apache.flink</groupId> >> <artifactId>flink-hadoop-compatibility_2.11</artifactId> >> <version>${flink.version}</version> >> </dependency> >> >> <dependency> >> <groupId>log4j</groupId> >> <artifactId>log4j</artifactId> >> <version>${log4j.version}</version> >> </dependency> >> </dependencies> >> Best wishes >> >> Thanks >