FWIW I also tried this on Flink 1.6.2 today and got the same error. This is my full stack trace:
java.lang.UnsupportedOperationException: Recoverable writers on Hadoop are only supported for HDFS and for Hadoop version 2.7 or newer at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57) at org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:202) at org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69) at org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111) at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242) at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327) at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178) at org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160) at org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96) at org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:277) at org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738) at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289) at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711) at java.lang.Thread.run(Thread.java:748) On Tue, Oct 30, 2018 at 4:12 PM Till Rohrmann <trohrm...@apache.org> wrote: > Hi Addison, > > I think the idea was to also backport this feature to 1.6 since we > considered it a bug that S3 was not supported in 1.6. I've pulled in Kostas > who worked on the S3 writer. @Klou did we intentionally not backport this > feature? > > I think there should be nothing special about backporting this feature and > building your own version of Flink. > > Cheers, > Till > > On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <addis...@gmail.com> > wrote: > > > Hi all, > > > > Been hitting my head against a wall for the last few hours. The release > > notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 as > > resolved in 1.6.2. I am trying to upgrade and switch some things to use > the > > StreamingFileSink against s3. However, when doing so, I get the following > > error: > > > > Recoverable writers on Hadoop are only supported for HDFS and for > > Hadoop version 2.7 or newer > > > > > > I started digging into the code and looking deeper, I don't see the code > at > > https://github.com/apache/flink/pull/6795 as being backported to 1.6.2? > > > > Was the Fix Version erroneous? If so, are there plans to backport it? If > > not, seems like that should be fixed in the release notes. > > > > I have been waiting for that functionality and we already build our own > > flink, so I am tempted to backport it onto 1.6.2... anything tricky about > > that? > > > > Thanks! > > >