Hi! Here is the issue: https://issues.apache.org/jira/browse/FLINK-8196
Fix coming up... Stephan On Mon, Dec 4, 2017 at 6:45 PM, Vijay Srinivasaraghavan < vijikar...@yahoo.com> wrote: > Hi Aljoscha, > > >> Regarding the issue you discovered, I think we have to revert a commit > that added exclusions that were too aggressive. > > Is there a JIRA for this? > > Regards > Vijay > > On Monday, December 4, 2017 1:16 AM, Aljoscha Krettek <aljos...@apache.org> > wrote: > > > Hi, > > The behaviour with Flink 1.4 is that we will always include Hadoop > dependencies unless we explicitly exclude them via "-DwithoutHadoop". If > you build without any flags you will get the default Hadoop version which > currently is 2.4.1. > > Regarding the issue you discovered, I think we have to revert a commit > that added exclusions that were too aggressive. > > Best, > Aljoscha > > > On 3. Dec 2017, at 21:58, Vijay Srinivasaraghavan < > vijikar...@yahoo.com.INVALID> wrote: > > > > The issue is reproducible with 2.7.1 as well. > > > > My understanding is from 1.4.0 we don't include Hadoop dependencies by > default but only when we explicitly enable the particular Hadoop version. > However, when I build it normally using "mvn clean install -DskipTests", > the Hadoop dependencies are getting included and I did not see the error > that I have reported earlier. > > > > Regards, > > Vijay > > > > Sent from my iPhone > > > >> On Dec 3, 2017, at 11:23 AM, Stephan Ewen <se...@apache.org> wrote: > >> > >> Hmmm, this is caused by a missing dependency (javax.servlet). > >> Could be that this dependency is newly excluded, or that this issue > only exists with Hadoop 2.7.0 (HDFS release tests have worked with both > newer and older Hadoop versions). > >> > >> Here is what I think we should do > >> > >> - Can you check whether this works with other Hadoop versions (for > example Hadoop 2.7.1 or so?) > >> - To run this in practice, you can also just drop the bundled hadoop > dependency and use whatever Hadoop dependency you have already on the > cluster (for example via yarn) > >> - Flink needs to adjust some dependencies / exclusions for the > pre-bundled Hadoop to make this build error go away > >> > >> > >> In any case, the fact that Hadoop seriously needs the Servlet > dependency to initialize HDFS is just ridiculous. Hadoop seriously has the > worst dependency management I have ever seen. > >> > >> > >>> On Fri, Dec 1, 2017 at 1:41 PM, Vijay Srinivasaraghavan < > vijikar...@yahoo.com.invalid> wrote: > >>> Hello, > >>> I am trying to build and run Flink from 1.4.0-rc2 branch with hadoop > binary 2.7.0 compatibility. > >>> Here are the steps I followed to build (I have maven 3.3.9). > >>> ===========================================================================cd > $FLINK_HOMEmvn clean install -DskipTests -Dhadoop.version=2.7.0cd > $FLINK_HOME/flink-distmvn clean install -Dhadoop.version=2.7.0 > >>> Running Flink from $FLINK_HOME/flink-dist/target/ > flink-1.4.0-bin============================================= > ============================= > >>> I am seeing below error error messages from the logs and it suggests > something to do with Hadoop dependency not available? > >>> Could someone please confirm the build steps? > >>> org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: Could > not find a file system implementation for scheme 'hdfs'. The scheme is not > directly supported by Flink and no Hadoop file system to support this > scheme could be loaded. at org.apache.flink.core.fs.FileSystem. > getUnguardedFileSystem(FileSystem.java:403) at org.apache.flink.core.fs. > FileSystem.get(FileSystem.java:318) at > org.apache.flink.core.fs.Path.getFileSystem(Path.java:293) > at org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.< > init>(FsCheckpointStreamFactory.java:99) at org.apache.flink.runtime. > state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > createCheckpointStreamFactory(StreamTask.java:787) at > org.apache.flink.streaming.api.operators.AbstractStreamOperator. > initializeState(AbstractStreamOperator.java:247) at > org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > initializeState(StreamTask.java:682) at org.apache.flink.streaming. > runtime.tasks.StreamTask.invoke(StreamTask.java:253) at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:718) at > java.lang.Thread.run(Thread.java:748)Caused by: org.apache.flink.core.fs. > UnsupportedFileSystemSchemeException: Cannot support file system for > 'hdfs' via Hadoop, because Hadoop is not in the classpath, or some classes > are missing from the classpath. at org.apache.flink.runtime.fs. > hdfs.HadoopFsFactory.create(HadoopFsFactory.java:179) at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:399) > ... 11 moreCaused by: java.lang.NoClassDefFoundError: Could not > initialize class org.apache.hadoop.hdfs.DFSConfigKeys at > org.apache.hadoop.hdfs.DFSClient$Conf.<init>(DFSClient.java:509) at > org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:637) at > org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize( > DistributedFileSystem.java:149) at org.apache.flink.runtime.fs. > hdfs.HadoopFsFactory.create(HadoopFsFactory.java:159) ... 12 > more2017-12-01 04:28:09,274 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph > - Source: Collection Source -> Map (1/1) ( > 74e8a3f21c86e8dec3c55988fff42e5d) switched from RUNNING to > FAILED.org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Could not find a file system implementation for scheme 'hdfs'. The scheme > is not directly supported by Flink and no Hadoop file system to support > this scheme could be loaded. at org.apache.flink.core.fs.FileSystem. > getUnguardedFileSystem(FileSystem.java:403) at org.apache.flink.core.fs. > FileSystem.get(FileSystem.java:318) at > org.apache.flink.core.fs.Path.getFileSystem(Path.java:293) > at org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.< > init>(FsCheckpointStreamFactory.java:99) at org.apache.flink.runtime. > state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > createCheckpointStreamFactory(StreamTask.java:787) at > org.apache.flink.streaming.api.operators.AbstractStreamOperator. > initializeState(AbstractStreamOperator.java:247) at > org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > initializeState(StreamTask.java:682) at org.apache.flink.streaming. > runtime.tasks.StreamTask.invoke(StreamTask.java:253) at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:718) at > java.lang.Thread.run(Thread.java:748)Caused by: org.apache.flink.core.fs. > UnsupportedFileSystemSchemeException: Cannot support file system for > 'hdfs' via Hadoop, because Hadoop is not in the classpath, or some classes > are missing from the classpath. at org.apache.flink.runtime.fs. > hdfs.HadoopFsFactory.create(HadoopFsFactory.java:179) at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:399) > ... 11 moreCaused by: java.lang.NoClassDefFoundError: Could not > initialize class org.apache.hadoop.hdfs.DFSConfigKeys at > org.apache.hadoop.hdfs.DFSClient$Conf.<init>(DFSClient.java:509) at > org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:637) at > org.apache.hadoop.hdfs.DFSClient.<init>(DFSClient.java:619) at > org.apache.hadoop.hdfs.DistributedFileSystem.initialize( > DistributedFileSystem.java:149) at org.apache.flink.runtime.fs. > hdfs.HadoopFsFactory.create(HadoopFsFactory.java:159) ... 12 > more2017-12-01 04:28:09,274 INFO > org.apache.flink.runtime.executiongraph.ExecutionGraph > - Source: Collection Source -> Map (1/1) ( > 18fcafbd7bc31a9257de7728b46cee7b) switched from RUNNING to > FAILED.org.apache.flink.core.fs.UnsupportedFileSystemSchemeException: > Could not find a file system implementation for scheme 'hdfs'. The scheme > is not directly supported by Flink and no Hadoop file system to support > this scheme could be loaded. at org.apache.flink.core.fs.FileSystem. > getUnguardedFileSystem(FileSystem.java:403) at org.apache.flink.core.fs. > FileSystem.get(FileSystem.java:318) at > org.apache.flink.core.fs.Path.getFileSystem(Path.java:293) > at org.apache.flink.runtime.state.filesystem.FsCheckpointStreamFactory.< > init>(FsCheckpointStreamFactory.java:99) at org.apache.flink.runtime. > state.filesystem.FsStateBackend.createStreamFactory(FsStateBackend.java:277) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > createCheckpointStreamFactory(StreamTask.java:787) at > org.apache.flink.streaming.api.operators.AbstractStreamOperator. > initializeState(AbstractStreamOperator.java:247) at > org.apache.flink.streaming.runtime.tasks.StreamTask.initializeOperators(StreamTask.java:694) > at org.apache.flink.streaming.runtime.tasks.StreamTask. > initializeState(StreamTask.java:682) at org.apache.flink.streaming. > runtime.tasks.StreamTask.invoke(StreamTask.java:253) at > org.apache.flink.runtime.taskmanager.Task.run(Task.java:718) at > java.lang.Thread.run(Thread.java:748)Caused by: org.apache.flink.core.fs. > UnsupportedFileSystemSchemeException: Cannot support file system for > 'hdfs' via Hadoop, because Hadoop is not in the classpath, or some classes > are missing from the classpath. at org.apache.flink.runtime.fs. > hdfs.HadoopFsFactory.create(HadoopFsFactory.java:179) at > org.apache.flink.core.fs.FileSystem.getUnguardedFileSystem(FileSystem.java:399) > ... 11 moreCaused by: java.lang.NoClassDefFoundError: > javax/servlet/Filter at java.lang.ClassLoader.defineClass1(Native Method) > at java.lang.ClassLoader.defineClass(ClassLoader.java:763) at > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) at > java.net.URLClassLoader.access$100(URLClassLoader.java:73) at > java.net.URLClassLoader$1.run(URLClassLoader.java:368) at > java.net.URLClassLoader$1.run(URLClassLoader.java:362) at java.security. > AccessController.doPrivileged(Native Method) at java.net.URLClassLoader. > findClass(URLClassLoader.java:361) at java.lang.ClassLoader. > loadClass(ClassLoader.java:424) at sun.misc.Launcher$ > AppClassLoader.loadClass(Launcher.java:335) at java.lang.ClassLoader. > loadClass(ClassLoader.java:357) at java.lang.ClassLoader.defineClass1(Native > Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:763) at > java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) > at java.net.URLClassLoader.defineClass(URLClassLoader.java:467) at > java.net.URLClassLoader.access$100(URLClassLoader.java:73) at > java.net.URLClassLoader$1.run(URLClassLoader.java:368) at > java.net.URLClassLoader$1.run(URLClassLoader.java:362) at java.security. > AccessController.doPrivileged(Native Method) at java.net.URLClassLoader. > findClass(URLClassLoader.java:361) at java.lang.ClassLoader. > loadClass(ClassLoader.java:424) at sun.misc.Launcher$ > AppClassLoader.loadClass(Launcher.java:335) at java.lang.ClassLoader. > loadClass(ClassLoader.java:357) at org.apache.hadoop.hdfs. > DFSConfigKeys.<clinit>(DFSConfigKeys.java:230) at org.apache.hadoop.hdfs. > DFSClient$Conf.<init>(DFSClient.java:509) at org.apache.hadoop.hdfs. > DFSClient.<init>(DFSClient.java:637) at org.apache.hadoop.hdfs. > DFSClient.<init>(DFSClient.java:619) at org.apache.hadoop.hdfs. > DistributedFileSystem.initialize(DistributedFileSystem.java:149) at > org.apache.flink.runtime.fs.hdfs.HadoopFsFactory.create(HadoopFsFactory.java:159) > ... 12 moreCaused by: java.lang.ClassNotFoundException: > javax.servlet.Filter at java.net.URLClassLoader. > findClass(URLClassLoader.java:381) at java.lang.ClassLoader. > loadClass(ClassLoader.java:424) at sun.misc.Launcher$ > AppClassLoader.loadClass(Launcher.java:335) at java.lang.ClassLoader. > loadClass(ClassLoader.java:357) ... 42 more > >>> RegardsVijay > >> > > >