Hi,

Could you please list what exactly is in your submitted jar file, for example 
using "jar tf my-jar-file.jar"? And also what files exactly are in your Flink 
lib directory.

Best,
Aljoscha

> On 19. Dec 2017, at 20:10, shashank agarwal <shashank...@gmail.com> wrote:
> 
> Hi Timo,
> 
> I am using Rocksdbstatebackend with hdfs path. I have following flink 
> dependencies in my sbt :
> 
> "org.slf4j" % "slf4j-log4j12" % "1.7.21",
>   "org.apache.flink" %% "flink-scala" % flinkVersion % "provided",
>   "org.apache.flink" %% "flink-streaming-scala" % flinkVersion % "provided",
>   "org.apache.flink" %% "flink-cep-scala" % flinkVersion,
>   "org.apache.flink" %% "flink-connector-kafka-0.10" % flinkVersion,
>   "org.apache.flink" %% "flink-connector-filesystem" % flinkVersion,
>   "org.apache.flink" %% "flink-statebackend-rocksdb" % flinkVersion,
>   "org.apache.flink" %% "flink-connector-cassandra" % "1.3.2",
>   "org.apache.flink" % "flink-shaded-hadoop2" % flinkVersion,
> 
> when i start flink yarn session  it's working fine even it's creating flink 
> checkpointing directory and copying libs into hdfs.
> 
> But when I submit the application to this yarn session it prints following 
> logs :
> 
>> Using the result of 'hadoop classpath' to augment the Hadoop classpath: 
>> /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0.
>>  
>> <http://2.6.0.0/>3-8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop-yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//*
> 
> But application fails contuniously with logs which i have sent earlier.
> 
> 
> ‌I have tried to add flink- hadoop-compability*.jar as suggested by Jorn but 
> it's not working.
> 
> 
> 
> On Tue, Dec 19, 2017 at 5:08 PM, shashank agarwal <shashank...@gmail.com 
> <mailto:shashank...@gmail.com>> wrote:
> yes, it's working fine. now not getting compile time error.
> 
> But when i trying to run this on cluster or yarn, getting following runtime 
> error :
> 
> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find 
> a file system implementation for scheme 'hdfs'. The scheme is not directly 
> supported by Flink and no Hadoop file system to support this scheme could be 
> loaded.
>       at 
> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:405)
>       at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:320)
>       at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293)
>       at 
> org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.<init>(FsCheckpointStreamFactory.java:99)
>       at 
> org.apache.flink.runtime.state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277)
>       at 
> org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createStreamFactory(RocksDBStateBackend.java:273)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.createCheckpointStreamFactory(StreamTask.java:787)
>       at 
> org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:247)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:682)
>       at 
> org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253)
>       at org.apache.flink.runtime.taskmanager.Task.run(Task.java:718)
>       at java.lang.Thread.run(Thread.java:745)
> Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: 
> Hadoop File System abstraction does not support scheme 'hdfs'. Either no file 
> system implementation exists for that scheme, or the relevant classes are 
> missing from the classpath.
>       at 
> org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:102)
>       at 
> org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:401)
>       ... 12 more
> Caused by: java.io.IOException: No FileSystem for scheme: hdfs
>       at 
> org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2786)
>       at 
> org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:99)
>       ... 13 more
> 
> 
> 
> 
> while submitting job it's printing following logs so i think it's including 
> hdoop libs :
> 
> Using the result of 'hadoop classpath' to augment the Hadoop classpath: 
> /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0.
>  
> <http://2.6.0.0/>3-8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop-yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//*
> 
> On Fri, Dec 8, 2017 at 9:24 PM, shashank agarwal <shashank...@gmail.com 
> <mailto:shashank...@gmail.com>> wrote:
> Sure i’ll Try that. Thanks
> 
> On Fri, 8 Dec 2017 at 9:18 PM, Stephan Ewen <se...@apache.org 
> <mailto:se...@apache.org>> wrote:
> I would recommend to add "flink-shaded-hadoop2". That is a bundle of all 
> Hadoop dependencies used by Flink.
> 
> 
> On Fri, Dec 8, 2017 at 3:44 PM, Aljoscha Krettek <aljos...@apache.org 
> <mailto:aljos...@apache.org>> wrote:
> I see, thanks for letting us know!
> 
> 
>> On 8. Dec 2017, at 15:42, shashank agarwal <shashank...@gmail.com 
>> <mailto:shashank...@gmail.com>> wrote:
>> 
>> I had to include two dependencies.
>> 
>> hadoop-hdfs (this for HDFS configuration) 
>> hadoop-common (this for Path)
>> 
>> 
>> ‌
>> 
>> On Fri, Dec 8, 2017 at 7:38 PM, Aljoscha Krettek <aljos...@apache.org 
>> <mailto:aljos...@apache.org>> wrote:
>> I think hadoop-hdfs might be sufficient.
>> 
>> 
>>> On 8. Dec 2017, at 14:48, shashank agarwal <shashank...@gmail.com 
>>> <mailto:shashank...@gmail.com>> wrote:
>>> 
>>> Can you specifically guide which dependencies should I add to extend this :
>>> 
>>> https://github.com/apache/flink/blob/release-1.4.0-rc3/flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/Bucketer.java
>>>  
>>> <https://github.com/apache/flink/blob/release-1.4.0-rc3/flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/Bucketer.java>
>>> 
>>> is the Hadoop-core is sufficient. 
>>> 
>>> 
>>> ‌
>>> 
>>> On Fri, Dec 8, 2017 at 6:58 PM, shashank agarwal <shashank...@gmail.com 
>>> <mailto:shashank...@gmail.com>> wrote:
>>> It's a compilation error. I think I have to include the Hadoop dependencies.
>>> 
>>> 
>>> 
>>> ‌
>>> 
>>> On Fri, Dec 8, 2017 at 6:54 PM, Aljoscha Krettek <aljos...@apache.org 
>>> <mailto:aljos...@apache.org>> wrote:
>>> Hi,
>>> 
>>> Is this a compilation error or at runtime. But in general, yes you have to 
>>> include the Hadoop dependencies if they're not there.
>>> 
>>> Best,
>>> Aljoscha
>>> 
>>> 
>>>> On 8. Dec 2017, at 14:10, shashank agarwal <shashank...@gmail.com 
>>>> <mailto:shashank...@gmail.com>> wrote:
>>>> 
>>>> Hello,
>>>> 
>>>> I am trying to test 1.4.0-RC3, Hadoop libraries removed in this version. 
>>>> Actually, i have created custom Bucketer for the bucketing sink.  I am 
>>>> extending 
>>>> 
>>>> 
>>>> org.apache.flink.streaming.connectors.fs.bucketing.Bucketer
>>>> 
>>>> in the class, i have to use org.apache.hadoop.fs.Path  but as hadoop 
>>>> libraries removed it's giving error 
>>>> 
>>>> "object hadoop is not a member of package org.apache"
>>>> 
>>>> Should i have to include Hadoop client libs in build.sbt dependencies.
>>>> 
>>>> 
>>>> Thanks
>>>> Shashank
>>>> ‌
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> Thanks Regards
>>> 
>>> SHASHANK AGARWAL
>>>  ---  Trying to mobilize the things....
>>> 
>>> 
>>> 
>>> 
>>> -- 
>>> Thanks Regards
>>> 
>>> SHASHANK AGARWAL
>>>  ---  Trying to mobilize the things....
>> 
>> 
>> 
>> 
>> -- 
>> Thanks Regards
>> 
>> SHASHANK AGARWAL
>>  ---  Trying to mobilize the things....
> 
> 
> -- 
> Sent from iPhone 5
> 
> 
> 
> -- 
> Thanks Regards
> 
> SHASHANK AGARWAL
>  ---  Trying to mobilize the things....
> 
> 
> 
> 
> -- 
> Thanks Regards
> 
> SHASHANK AGARWAL
>  ---  Trying to mobilize the things....

Reply via email to