Hi Shashank,

it seems that HDFS is still not in classpath. Could you quickly explain how I can reproduce the error?

Regards,
Timo



Am 12/19/17 um 12:38 PM schrieb shashank agarwal:
yes, it's working fine. now not getting compile time error.

But when i trying to run this on cluster or yarn, getting following runtime error :

org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could not find a 
file system implementation for scheme 'hdfs'. The scheme is not directly 
supported by Flink and no Hadoop file system to support this scheme could be 
loaded.
        at 
org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:405)
        at org.apache.flink.core.fs.FileSystem.get(FileSystem.java:320)
        at org.apache.flink.core.fs.Path.getFileSystem(Path.java:293)
        at 
org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.<init>(FsCheckpointStreamFactory.java:99)
        at 
org.apache.flink.runtime.state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277)
        at 
org.apache.flink.contrib.streaming.state.RocksDBStateBackend.createStreamFactory(RocksDBStateBackend.java:273)
        at 
org.apache.flink.streaming.runtime.tasks.StreamTask.createCheckpointStreamFactory(StreamTask.java:787)
        at 
org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:247)
        at 
org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694)
        at 
org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:682)
        at 
org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:253)
        at org.apache.flink.runtime.taskmanager.Task.run(Task.java:718)
        at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: 
Hadoop File System abstraction does not support scheme 'hdfs'. Either no file 
system implementation exists for that scheme, or the relevant classes are 
missing from the classpath.
        at 
org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:102)
        at 
org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:401)
        ... 12 more
Caused by: java.io.IOException: No FileSystem for scheme: hdfs
        at 
org.apache.hadoop.fs.FileSystem.getFileSystemClass(FileSystem.java:2786)
        at 
org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:99)
        ... 13 more
while submitting job it's printing following logs so i think it's including 
hdoop libs :
Using the result of 'hadoop classpath' to augment the Hadoop classpath: /usr/hdp/2.6.0.3-8/hadoop/conf:/usr/hdp/2.6.0.3-8/hadoop/lib/*:/usr/hdp/2.6.0.3-8/hadoop/.//*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/./:/usr/hdp/2.6.0.3-8/hadoop-hdfs/lib/*:/usr/hdp/2.6.0.3-8/hadoop-hdfs/.//*:/usr/hdp/2.6.0.3-8/hadoop-yarn/lib/*:/usr/hdp/2.6.0.3-8/hadoop-yarn/.//*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/lib/*:/usr/hdp/2.6.0.3-8/hadoop-mapreduce/.//*

On Fri, Dec 8, 2017 at 9:24 PM, shashank agarwal <shashank...@gmail.com <mailto:shashank...@gmail.com>> wrote:

    Sure i’ll Try that. Thanks

    On Fri, 8 Dec 2017 at 9:18 PM, Stephan Ewen <se...@apache.org
    <mailto:se...@apache.org>> wrote:

        I would recommend to add "flink-shaded-hadoop2". That is a
        bundle of all Hadoop dependencies used by Flink.


        On Fri, Dec 8, 2017 at 3:44 PM, Aljoscha Krettek
        <aljos...@apache.org <mailto:aljos...@apache.org>> wrote:

            I see, thanks for letting us know!


            On 8. Dec 2017, at 15:42, shashank agarwal
            <shashank...@gmail.com <mailto:shashank...@gmail.com>> wrote:

            I had to include two dependencies.

            hadoop-hdfs (this for HDFS configuration)
            hadoop-common (this for Path)


            ‌

            On Fri, Dec 8, 2017 at 7:38 PM, Aljoscha
            Krettek<aljos...@apache.org
            <mailto:aljos...@apache.org>>wrote:

                I think hadoop-hdfs might be sufficient.


                On 8. Dec 2017, at 14:48, shashank agarwal
                <shashank...@gmail.com
                <mailto:shashank...@gmail.com>> wrote:

                Can you specifically guide which dependencies should
                I add to extend this :

                
https://github.com/apache/flink/blob/release-1.4.0-rc3/flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/Bucketer.java
                
<https://github.com/apache/flink/blob/release-1.4.0-rc3/flink-connectors/flink-connector-filesystem/src/main/java/org/apache/flink/streaming/connectors/fs/Bucketer.java>

                is the Hadoop-core is sufficient.


                ‌

                On Fri, Dec 8, 2017 at 6:58 PM, shashank
                agarwal<shashank...@gmail.com
                <mailto:shashank...@gmail.com>>wrote:

                    It's a compilation error. I think I have to
                    include the Hadoop dependencies.



                    ‌

                    On Fri, Dec 8, 2017 at 6:54 PM, Aljoscha
                    Krettek<aljos...@apache.org
                    <mailto:aljos...@apache.org>>wrote:

                        Hi,

                        Is this a compilation error or at runtime.
                        But in general, yes you have to include the
                        Hadoop dependencies if they're not there.

                        Best,
                        Aljoscha


                        On 8. Dec 2017, at 14:10, shashank agarwal
                        <shashank...@gmail.com
                        <mailto:shashank...@gmail.com>> wrote:

                        Hello,

                        I am trying to test 1.4.0-RC3, Hadoop
                        libraries removed in this version.
                        Actually, i have created custom
                        Bucketer for the bucketing sink.  I am
                        extending


                        
org.apache.flink.streaming.connectors.fs.bucketing.Bucketer

                        in the class, i have to
                        use org.apache.hadoop.fs.Path but as hadoop
                        libraries removed it's giving error

                        "object hadoop is not a member of package
                        org.apache"

                        Should i have to include Hadoop client libs
                        in build.sbt dependencies.


                        Thanks
                        Shashank
                        ‌




                    --
                    Thanks Regards

                    SHASHANK AGARWAL
                     ---  Trying to mobilize the things....




                --
                Thanks Regards

                SHASHANK AGARWAL
                 ---  Trying to mobilize the things....




            --
            Thanks Regards

            SHASHANK AGARWAL
             ---  Trying to mobilize the things....


-- Sent from iPhone 5




--
Thanks Regards

SHASHANK AGARWAL
 ---  Trying to mobilize the things....


Reply via email to