Hi Addison, This is definitely an error on my end, as the feature is going to be available from Flink 1.7 onwards. I forgot to correctly update the Flink version when closing the JIRA issue.
I will update the release notes accordingly. Sorry for the miscommunication, Kostas > On Oct 31, 2018, at 5:45 AM, Addison Higham <addis...@gmail.com> wrote: > > I have backported this at > https://github.com/instructure/flink/tree/s3_recover_backport > <https://github.com/instructure/flink/tree/s3_recover_backport> by > cherry-picking all the relevant code, I am not sure how backports are usually > done with Flink (if you squash and merge) but there were a few minor > conflicts and involved quite a few changes from master. > > Going to try with my branch tomorrow and will report any issues. > > > > On Tue, Oct 30, 2018 at 8:44 PM Mike Mintz <mikemi...@gmail.com > <mailto:mikemi...@gmail.com>> wrote: > FWIW I also tried this on Flink 1.6.2 today and got the same error. This is > my full stack trace: > > java.lang.UnsupportedOperationException: Recoverable writers on Hadoop > are only supported for HDFS and for Hadoop version 2.7 or newer > at > org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopRecoverableWriter.<init>(HadoopRecoverableWriter.java:57) > at > org.apache.flink.fs.s3hadoop.shaded.org.apache.flink.runtime.fs.hdfs.HadoopFileSystem.createRecoverableWriter(HadoopFileSystem.java:202) > at > org.apache.flink.core.fs.SafetyNetWrapperFileSystem.createRecoverableWriter(SafetyNetWrapperFileSystem.java:69) > at > org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.<init>(Buckets.java:111) > at > org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink$RowFormatBuilder.createBuckets(StreamingFileSink.java:242) > at > org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSink.initializeState(StreamingFileSink.java:327) > at > org.apache.flink.streaming.util.functions.StreamingFunctionUtils.tryRestoreFunction(StreamingFunctionUtils.java:178) > at > org.apache.flink.streaming.util.functions.StreamingFunctionUtils.restoreFunctionState(StreamingFunctionUtils.java:160) > at > org.apache.flink.streaming.api.operators.AbstractUdfStreamOperator.initializeState(AbstractUdfStreamOperator.java:96) > at > org.apache.flink.streaming.api.operators.AbstractStreamOperator.initializeState(AbstractStreamOperator.java:277) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.initializeState(StreamTask.java:738) > at > org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:289) > at org.apache.flink.runtime.taskmanager.Task.run(Task.java:711) > at java.lang.Thread.run(Thread.java:748) > > > On Tue, Oct 30, 2018 at 4:12 PM Till Rohrmann <trohrm...@apache.org > <mailto:trohrm...@apache.org>> wrote: > > > Hi Addison, > > > > I think the idea was to also backport this feature to 1.6 since we > > considered it a bug that S3 was not supported in 1.6. I've pulled in Kostas > > who worked on the S3 writer. @Klou did we intentionally not backport this > > feature? > > > > I think there should be nothing special about backporting this feature and > > building your own version of Flink. > > > > Cheers, > > Till > > > > On Tue, Oct 30, 2018 at 10:54 PM Addison Higham <addis...@gmail.com > > <mailto:addis...@gmail.com>> > > wrote: > > > > > Hi all, > > > > > > Been hitting my head against a wall for the last few hours. The release > > > notes for 1.6.2 show https://issues.apache.org/jira/browse/FLINK-9752 > > > <https://issues.apache.org/jira/browse/FLINK-9752> as > > > resolved in 1.6.2. I am trying to upgrade and switch some things to use > > the > > > StreamingFileSink against s3. However, when doing so, I get the following > > > error: > > > > > > Recoverable writers on Hadoop are only supported for HDFS and for > > > Hadoop version 2.7 or newer > > > > > > > > > I started digging into the code and looking deeper, I don't see the code > > at > > > https://github.com/apache/flink/pull/6795 > > > <https://github.com/apache/flink/pull/6795> as being backported to 1.6.2? > > > > > > Was the Fix Version erroneous? If so, are there plans to backport it? If > > > not, seems like that should be fixed in the release notes. > > > > > > I have been waiting for that functionality and we already build our own > > > flink, so I am tempted to backport it onto 1.6.2... anything tricky about > > > that? > > > > > > Thanks! > > > > >