I've created a JIRA issue for the the Hadoop 2.9.0 build problem [1].

Best, Fabian

[1] https://issues.apache.org/jira/browse/FLINK-8177

2017-11-30 4:35 GMT+01:00 Eron Wright <eronwri...@gmail.com>:

> Unfortunately we've identified a blocker bug for Flink on Mesos -
> FLINK-8174.    We'll have a patch ready on Thursday.
>
> Thanks,
> Eron
>
> On Wed, Nov 29, 2017 at 3:40 PM, Eron Wright <eronwri...@gmail.com> wrote:
>
> > On Dell EMC side, we're testing the RC2 on DCOS 1.10.0.   Seeing a
> > potential issue with offer acceptance and we'll update the thread with a
> +1
> > or with a more concrete issue within 24 hours.
> >
> > Thanks,
> > Eron
> >
> > On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <ches...@apache.org>
> > wrote:
> >
> >> I don't think anyone has taken a look yet, nor was there a discussion as
> >> to postponing it.
> >>
> >> It just slipped through the cracks i guess...
> >>
> >>
> >> On 29.11.2017 15:47, Gyula Fóra wrote:
> >>
> >>> Hi guys,
> >>> I ran into this again while playing with savepoint/restore parallelism:
> >>>
> >>> https://issues.apache.org/jira/browse/FLINK-7595
> >>> https://github.com/apache/flink/pull/4651
> >>>
> >>> Anyone has some idea about the status of this PR or were we planning to
> >>> postpone this to 1.5?
> >>>
> >>> Thanks,
> >>> Gyula
> >>>
> >>>
> >>> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov. 29.,
> >>> Sze,
> >>> 13:10):
> >>>
> >>> OK, the situation is the following:
> >>>>
> >>>> The test class (org.apache.flink.yarn.UtilsTest) implements a Hadoop
> >>>> interface (Container) that was extended in Hadoop 2.9.0 by a getter
> and
> >>>> setter.
> >>>> By adding the methods, we can compile Flink for Hadoop 2.9.0. However,
> >>>> the
> >>>> getter/setter add a dependency on a class that was also added in
> Hadoop
> >>>> 2.9.0.
> >>>> Therefore, the implementation is not backwards compatible with Hadoop
> >>>> versions < 2.9.0.
> >>>>
> >>>> Not sure how we can fix the problem. We would need two version of the
> >>>> class
> >>>> that are chosen based on the Hadoop version. Do we have something like
> >>>> that
> >>>> somewhere else?
> >>>>
> >>>> Since this is only a problem in a test class, Flink 1.4.0 might still
> >>>> work
> >>>> very well with Hadoop 2.9.0.
> >>>> However, this has not been tested AFAIK.
> >>>>
> >>>> Cheers, Fabian
> >>>>
> >>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>:
> >>>>
> >>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0
> (released
> >>>>> a
> >>>>> few days ago) and got a compilation failure in a test class.
> >>>>>
> >>>>> Right now, I'm assessing how much we need to fix to support Hadoop
> >>>>> 2.9.0.
> >>>>> I'll report later.
> >>>>>
> >>>>> Best, Fabian
> >>>>>
> >>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org>:
> >>>>>
> >>>>> Agreed, this is a regression compared to the previous functionality.
> I
> >>>>>> updated the issue to "Blocker".
> >>>>>>
> >>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com> wrote:
> >>>>>>>
> >>>>>>> Hi all,
> >>>>>>>
> >>>>>>> I have found the following issue:
> >>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
> >>>>>>>
> >>>>>>> I would say this is a blocker (I personally pass the ParameterTool
> >>>>>>> all
> >>>>>>>
> >>>>>> over
> >>>>>>
> >>>>>>> the place in my production apps), but a pretty trivial issue to
> fix,
> >>>>>>>
> >>>>>> we
> >>>>
> >>>>> can
> >>>>>>
> >>>>>>> wait a little to find other potential problems.
> >>>>>>>
> >>>>>>> I can submit a fix in a little bit.
> >>>>>>>
> >>>>>>> Cheers,
> >>>>>>> Gyula
> >>>>>>>
> >>>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont: 2017.
> >>>>>>>
> >>>>>> nov.
> >>>>
> >>>>> 29., Sze, 9:23):
> >>>>>>>
> >>>>>>> +1
> >>>>>>>>
> >>>>>>>> Verified:
> >>>>>>>> - No missing release Maven artifacts
> >>>>>>>> - Staged Apache source & binary convenience releases looks good
> >>>>>>>> - NOTICE / LICENSE is correct, README is sane
> >>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop 2.8)
> >>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
> >>>>>>>>
> >>>>>>> configuration
> >>>>
> >>>>> details)
> >>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency clashes
> on
> >>>>>>>> cluster execution, works locally in IDE)
> >>>>>>>>
> >>>>>>>> Thanks a lot for managing the release Aljoscha!
> >>>>>>>>
> >>>>>>>> Cheers,
> >>>>>>>> Gordon
> >>>>>>>>
> >>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
> >>>>>>>> s.rich...@data-artisans.com) wrote:
> >>>>>>>>
> >>>>>>>> +1 (non-binding)
> >>>>>>>>
> >>>>>>>> I tested Flink in a cluster setup on Google Cloud, YARN-per-job,
> >>>>>>>>
> >>>>>>> checked
> >>>>>>
> >>>>>>> that for all backends that HA, recovery, at-least-once, end-to-end
> >>>>>>>>
> >>>>>>> exactly
> >>>>>>
> >>>>>>> once (with Kafka11 Producer), savepoints, externalized checkpoints,
> >>>>>>>>
> >>>>>>> and
> >>>>
> >>>>> rescaling work correctly.
> >>>>>>>>
> >>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
> >>>>>>>>>
> >>>>>>>> aljos...@apache.org
> >>>>
> >>>>> :
> >>>>>>>
> >>>>>>>> +1
> >>>>>>>>>
> >>>>>>>>> Verified:
> >>>>>>>>> - NOTICE and LICENSE are correct
> >>>>>>>>> - source doesn't contain binaries
> >>>>>>>>> - verified signatures
> >>>>>>>>> - verified hashes
> >>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
> >>>>>>>>>
> >>>>>>>> release-testing doc)
> >>>>>>>>
> >>>>>>>>>
> >>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <aljos...@apache.org
> >
> >>>>>>>>>>
> >>>>>>>>> wrote:
> >>>>>>>>
> >>>>>>>>> Phew, thanks for the update!
> >>>>>>>>>>
> >>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org>
> wrote:
> >>>>>>>>>>>
> >>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was
> corrupted
> >>>>>>>>>>>
> >>>>>>>>>> (makes
> >>>>>>
> >>>>>>> total sense) :P
> >>>>>>>>>>>
> >>>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017. nov.
> >>>>>>>>>>> 28.,
> >>>>>>>>>>>
> >>>>>>>>>> K,
> >>>>>>
> >>>>>>> 11:13):
> >>>>>>>>
> >>>>>>>>> Hi Aljoscha,
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
> >>>>>>>>>>>>
> >>>>>>>>>>> building
> >>>>
> >>>>> the rc,
> >>>>>>>>
> >>>>>>>>> I seem to get this error no matter what I do:
> >>>>>>>>>>>>
> >>>>>>>>>>>> [ERROR] Failed to execute goal
> >>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
> >>>>>>>>>>>>
> >>>>>>>>>>> (shade-hadoop) on
> >>>>>>>>
> >>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
> >>>>>>>>>>>>
> >>>>>>>>>>> invalid
> >>>>>>
> >>>>>>> LOC
> >>>>>>>>
> >>>>>>>>> header (bad signature) -> [Help 1]
> >>>>>>>>>>>>
> >>>>>>>>>>>> (Apache Maven 3.3.9)
> >>>>>>>>>>>>
> >>>>>>>>>>>> Any idea what I am missing?
> >>>>>>>>>>>>
> >>>>>>>>>>>> Thanks,
> >>>>>>>>>>>> Gyula
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont:
> 2017.
> >>>>>>>>>>>>
> >>>>>>>>>>> nov.
> >>>>>>
> >>>>>>> 27.,
> >>>>>>>>
> >>>>>>>>> H, 19:35):
> >>>>>>>>>>>>
> >>>>>>>>>>>> Hi everyone,
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Please review and vote on release candidate #2 for the
> version
> >>>>>>>>>>>>>
> >>>>>>>>>>>> 1.4.0, as
> >>>>>>>>
> >>>>>>>>> follows:
> >>>>>>>>>>>>> [ ] +1, Approve the release
> >>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide specific
> >>>>>>>>>>>>>
> >>>>>>>>>>>> comments)
> >>>>>>
> >>>>>>>
> >>>>>>>>>>>>> The complete staging area is available for your review, which
> >>>>>>>>>>>>>
> >>>>>>>>>>>> includes:
> >>>>>>>>
> >>>>>>>>> * JIRA release notes [1],
> >>>>>>>>>>>>> * the official Apache source release and binary convenience
> >>>>>>>>>>>>>
> >>>>>>>>>>>> releases
> >>>>>>
> >>>>>>> to
> >>>>>>>>
> >>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
> >>>>>>>>>>>>>
> >>>>>>>>>>>> key
> >>>>
> >>>>> with
> >>>>>>>>
> >>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
> >>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central
> Repository
> >>>>>>>>>>>>>
> >>>>>>>>>>>> [4],
> >>>>>>
> >>>>>>> * source code tag "release-1.4.0-rc1" [5],
> >>>>>>>>>>>>> * website pull request listing the new release [6].
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Please have a careful look at the website PR because I
> changed
> >>>>>>>>>>>>>
> >>>>>>>>>>>> some
> >>>>>>
> >>>>>>> wording and we're now also releasing a binary without Hadoop
> >>>>>>>>>>>>>
> >>>>>>>>>>>> dependencies.
> >>>>>>>>
> >>>>>>>>> Please use this document for coordinating testing efforts: [7]
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> The only change between RC1 and this RC2 is that the source
> >>>>>>>>>>>>>
> >>>>>>>>>>>> release
> >>>>>>
> >>>>>>> package does not include the erroneously included binary Ruby
> >>>>>>>>>>>>>
> >>>>>>>>>>>> dependencies
> >>>>>>>>
> >>>>>>>>> of the documentation anymore. Because of this I would like to
> >>>>>>>>>>>>>
> >>>>>>>>>>>> propose a
> >>>>>>>>
> >>>>>>>>> shorter voting time and close the vote around the time that RC1
> >>>>>>>>>>>>>
> >>>>>>>>>>>> would have
> >>>>>>>>
> >>>>>>>>> closed. This would mean closing by end of Wednesday. Please let
> >>>>>>>>>>>>>
> >>>>>>>>>>>> me
> >>>>
> >>>>> know if
> >>>>>>>>
> >>>>>>>>> you disagree with this. The vote is adopted by majority
> >>>>>>>>>>>>>
> >>>>>>>>>>>> approval,
> >>>>
> >>>>> with at
> >>>>>>>>
> >>>>>>>>> least 3 PMC affirmative votes.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> Thanks,
> >>>>>>>>>>>>> Your friendly Release Manager
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> [1]
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?proje
> >>>>>>>>
> >>>>>>> ctId=12315522&version=12340533
> >>>>>>
> >>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
> >>>>>>>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> >>>>>>>>>>>>> [4]
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
> >>>> flink-1140
> >>>>
> >>>>> [5]
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://git-wip-us.apache.org/repos/asf?p=flink.git;a=tag;h=
> >>>>>>>>
> >>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
> >>>>>>
> >>>>>>> [6] https://github.com/apache/flink-web/pull/95
> >>>>>>>>>>>>> [7]
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://docs.google.com/document/d/1HqYyrNoMSXwo8zBpZj7s39Uz
> >>>>>>>>
> >>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
> >>>>>>
> >>>>>>> Pro-tip: you can create a settings.xml file with these contents:
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> <settings>
> >>>>>>>>>>>>> <activeProfiles>
> >>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
> >>>>>>>>>>>>> </activeProfiles>
> >>>>>>>>>>>>> <profiles>
> >>>>>>>>>>>>> <profile>
> >>>>>>>>>>>>> <id>flink-1.4.0</id>
> >>>>>>>>>>>>> <repositories>
> >>>>>>>>>>>>> <repository>
> >>>>>>>>>>>>> <id>flink-1.4.0</id>
> >>>>>>>>>>>>> <url>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
> >>>>>>>>
> >>>>>>> flink-1140/
> >>>>>>
> >>>>>>> </url>
> >>>>>>>>>>>>> </repository>
> >>>>>>>>>>>>> <repository>
> >>>>>>>>>>>>> <id>archetype</id>
> >>>>>>>>>>>>> <url>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> https://repository.apache.org/content/repositories/orgapache
> >>>>>>>>
> >>>>>>> flink-1140/
> >>>>>>
> >>>>>>> </url>
> >>>>>>>>>>>>> </repository>
> >>>>>>>>>>>>> </repositories>
> >>>>>>>>>>>>> </profile>
> >>>>>>>>>>>>> </profiles>
> >>>>>>>>>>>>> </settings>
> >>>>>>>>>>>>>
> >>>>>>>>>>>>> And reference that in you maven commands via --settings
> >>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a
> quickstart
> >>>>>>>>>>>>>
> >>>>>>>>>>>> based
> >>>>>>
> >>>>>>> on the
> >>>>>>>>
> >>>>>>>>> staged release and for building against the staged jars.
> >>>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>>>>>
> >>>>>>>>
> >>>>>>
> >>
> >
>

Reply via email to