Neat! I wil try and make sure that we get the fix in ASAP.
> On 3. Dec 2017, at 12:34, Eron Wright <eronwri...@gmail.com> wrote:
>
> Update: PR for "FLINK-8174 - Mesos RM unable to accept offers for
> unreserved resources" is ready for review. We were able to solve (3),
> supporting any combination of reserved and unreserved resources.
>
> On Fri, Dec 1, 2017 at 2:24 PM, Eron Wright <eronwri...@gmail.com> wrote:
>
>> There are three levels of support we could land on.
>> 1. Flink works with unreserved resources (revert FLINK-7294).
>> 2. Flink works with unreserved resources, and correctly ignores reserved
>> resources (revert FLINK-7294 and mitigate Fenzo bug).
>> 3. Flink works with unreserved resources and reserved resources.
>>
>> 3 is a moon shot. Striving for 2. Fallback on 1.
>>
>>
>>
>> On Fri, Dec 1, 2017 at 2:10 PM, Aljoscha Krettek <aljos...@apache.org>
>> wrote:
>>
>>> Thanks for the update!
>>>
>>> Just to be clear, you're proposing going forward with the "simple fix" of
>>> reverting FLINK-7294?
>>>
>>>> On 1. Dec 2017, at 18:39, Eron Wright <eronwri...@gmail.com> wrote:
>>>>
>>>> Update on reported Mesos issue (FLINK-8174):
>>>>
>>>> TLDR; a PR will be ready within 24 hours that will undo reservation
>>> support.
>>>>
>>>> A couple of months ago, a fix (FLINK-7294) was merged related to how
>>> Flink
>>>> accepts Mesos resource offers. The intention was to allow Flink to make
>>>> use of so-called +reserved+ resources, a Mesos feature which makes it
>>>> possible to reserve hosts for use by a specific framework/role. The fix
>>>> inadvertently regressed the ability to use +unreserved+ resources.
>>> This is
>>>> a serious regression because unreserved resources are the common case.
>>>>
>>>> The simple solution is to revert the earlier fix, deferring support for
>>>> reservations to another release. We are spending some time to find a
>>> fix
>>>> that works for all scenarios, but seems unlikely at this time. I am
>>>> reaching out to the original contributor to get their feedback.
>>>>
>>>> In the course of the investigation, a related flaw was discovered in
>>> Fenzo
>>>> that causes Flink to misinterpret offers that contain a mix of reserved
>>> and
>>>> unreserved resources. I believe that a small fix is possible purely
>>>> within Flink; an update to Fenzo does not appear necessary.
>>>>
>>>> Going forward, we will contribute an improved integration test suite
>>> with
>>>> which to test Flink under diverse Mesos conditions (e.g. reservations).
>>>>
>>>> Thanks,
>>>> Eron
>>>>
>>>> On Thu, Nov 30, 2017 at 9:47 PM, Tzu-Li (Gordon) Tai <
>>> tzuli...@apache.org>
>>>> wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> I’ve noticed a behavioral regression in the Kafka producer, that should
>>>>> also be considered a blocker: https://issues.apache.org/
>>>>> jira/browse/FLINK-8181
>>>>> There’s already a PR for the issue here: https://github.com/
>>>>> apache/flink/pull/5108
>>>>>
>>>>> Best,
>>>>> Gordon
>>>>>
>>>>> On 30 November 2017 at 5:27:22 PM, Fabian Hueske (fhue...@gmail.com)
>>>>> wrote:
>>>>>
>>>>> I've created a JIRA issue for the the Hadoop 2.9.0 build problem [1].
>>>>>
>>>>> Best, Fabian
>>>>>
>>>>> [1] https://issues.apache.org/jira/browse/FLINK-8177
>>>>>
>>>>> 2017-11-30 4:35 GMT+01:00 Eron Wright <eronwri...@gmail.com>:
>>>>>
>>>>>> Unfortunately we've identified a blocker bug for Flink on Mesos -
>>>>>> FLINK-8174. We'll have a patch ready on Thursday.
>>>>>>
>>>>>> Thanks,
>>>>>> Eron
>>>>>>
>>>>>> On Wed, Nov 29, 2017 at 3:40 PM, Eron Wright <eronwri...@gmail.com>
>>>>> wrote:
>>>>>>
>>>>>>> On Dell EMC side, we're testing the RC2 on DCOS 1.10.0. Seeing a
>>>>>>> potential issue with offer acceptance and we'll update the thread
>>> with
>>>>> a
>>>>>> +1
>>>>>>> or with a more concrete issue within 24 hours.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> Eron
>>>>>>>
>>>>>>> On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <
>>> ches...@apache.org>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> I don't think anyone has taken a look yet, nor was there a
>>> discussion
>>>>> as
>>>>>>>> to postponing it.
>>>>>>>>
>>>>>>>> It just slipped through the cracks i guess...
>>>>>>>>
>>>>>>>>
>>>>>>>> On 29.11.2017 15:47, Gyula Fóra wrote:
>>>>>>>>
>>>>>>>>> Hi guys,
>>>>>>>>> I ran into this again while playing with savepoint/restore
>>>>> parallelism:
>>>>>>>>>
>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-7595
>>>>>>>>> https://github.com/apache/flink/pull/4651
>>>>>>>>>
>>>>>>>>> Anyone has some idea about the status of this PR or were we
>>> planning
>>>>> to
>>>>>>>>> postpone this to 1.5?
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> Gyula
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov.
>>> 29.,
>>>>>>>>> Sze,
>>>>>>>>> 13:10):
>>>>>>>>>
>>>>>>>>> OK, the situation is the following:
>>>>>>>>>>
>>>>>>>>>> The test class (org.apache.flink.yarn.UtilsTest) implements a
>>>>> Hadoop
>>>>>>>>>> interface (Container) that was extended in Hadoop 2.9.0 by a
>>> getter
>>>>>> and
>>>>>>>>>> setter.
>>>>>>>>>> By adding the methods, we can compile Flink for Hadoop 2.9.0.
>>>>> However,
>>>>>>>>>> the
>>>>>>>>>> getter/setter add a dependency on a class that was also added in
>>>>>> Hadoop
>>>>>>>>>> 2.9.0.
>>>>>>>>>> Therefore, the implementation is not backwards compatible with
>>>>> Hadoop
>>>>>>>>>> versions < 2.9.0.
>>>>>>>>>>
>>>>>>>>>> Not sure how we can fix the problem. We would need two version of
>>>>> the
>>>>>>>>>> class
>>>>>>>>>> that are chosen based on the Hadoop version. Do we have something
>>>>> like
>>>>>>>>>> that
>>>>>>>>>> somewhere else?
>>>>>>>>>>
>>>>>>>>>> Since this is only a problem in a test class, Flink 1.4.0 might
>>>>> still
>>>>>>>>>> work
>>>>>>>>>> very well with Hadoop 2.9.0.
>>>>>>>>>> However, this has not been tested AFAIK.
>>>>>>>>>>
>>>>>>>>>> Cheers, Fabian
>>>>>>>>>>
>>>>>>>>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>:
>>>>>>>>>>
>>>>>>>>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0
>>>>>> (released
>>>>>>>>>>> a
>>>>>>>>>>> few days ago) and got a compilation failure in a test class.
>>>>>>>>>>>
>>>>>>>>>>> Right now, I'm assessing how much we need to fix to support
>>> Hadoop
>>>>>>>>>>> 2.9.0.
>>>>>>>>>>> I'll report later.
>>>>>>>>>>>
>>>>>>>>>>> Best, Fabian
>>>>>>>>>>>
>>>>>>>>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org
>>>> :
>>>>>>>>>>>
>>>>>>>>>>> Agreed, this is a regression compared to the previous
>>>>> functionality.
>>>>>> I
>>>>>>>>>>>> updated the issue to "Blocker".
>>>>>>>>>>>>
>>>>>>>>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com>
>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Hi all,
>>>>>>>>>>>>>
>>>>>>>>>>>>> I have found the following issue:
>>>>>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
>>>>>>>>>>>>>
>>>>>>>>>>>>> I would say this is a blocker (I personally pass the
>>>>> ParameterTool
>>>>>>>>>>>>> all
>>>>>>>>>>>>>
>>>>>>>>>>>> over
>>>>>>>>>>>>
>>>>>>>>>>>>> the place in my production apps), but a pretty trivial issue to
>>>>>> fix,
>>>>>>>>>>>>>
>>>>>>>>>>>> we
>>>>>>>>>>
>>>>>>>>>>> can
>>>>>>>>>>>>
>>>>>>>>>>>>> wait a little to find other potential problems.
>>>>>>>>>>>>>
>>>>>>>>>>>>> I can submit a fix in a little bit.
>>>>>>>>>>>>>
>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>
>>>>>>>>>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont:
>>>>> 2017.
>>>>>>>>>>>>>
>>>>>>>>>>>> nov.
>>>>>>>>>>
>>>>>>>>>>> 29., Sze, 9:23):
>>>>>>>>>>>>>
>>>>>>>>>>>>> +1
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>> - No missing release Maven artifacts
>>>>>>>>>>>>>> - Staged Apache source & binary convenience releases looks
>>> good
>>>>>>>>>>>>>> - NOTICE / LICENSE is correct, README is sane
>>>>>>>>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop
>>>>> 2.8)
>>>>>>>>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
>>>>>>>>>>>>>>
>>>>>>>>>>>>> configuration
>>>>>>>>>>
>>>>>>>>>>> details)
>>>>>>>>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency
>>>>> clashes
>>>>>> on
>>>>>>>>>>>>>> cluster execution, works locally in IDE)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks a lot for managing the release Aljoscha!
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>> Gordon
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
>>>>>>>>>>>>>> s.rich...@data-artisans.com) wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> +1 (non-binding)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I tested Flink in a cluster setup on Google Cloud,
>>> YARN-per-job,
>>>>>>>>>>>>>>
>>>>>>>>>>>>> checked
>>>>>>>>>>>>
>>>>>>>>>>>>> that for all backends that HA, recovery, at-least-once,
>>>>> end-to-end
>>>>>>>>>>>>>>
>>>>>>>>>>>>> exactly
>>>>>>>>>>>>
>>>>>>>>>>>>> once (with Kafka11 Producer), savepoints, externalized
>>>>> checkpoints,
>>>>>>>>>>>>>>
>>>>>>>>>>>>> and
>>>>>>>>>>
>>>>>>>>>>> rescaling work correctly.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> aljos...@apache.org
>>>>>>>>>>
>>>>>>>>>>> :
>>>>>>>>>>>>>
>>>>>>>>>>>>>> +1
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>>> - NOTICE and LICENSE are correct
>>>>>>>>>>>>>>> - source doesn't contain binaries
>>>>>>>>>>>>>>> - verified signatures
>>>>>>>>>>>>>>> - verified hashes
>>>>>>>>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>> release-testing doc)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <
>>>>> aljos...@apache.org
>>>>>>>
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Phew, thanks for the update!
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org>
>>>>>> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was
>>>>>> corrupted
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> (makes
>>>>>>>>>>>>
>>>>>>>>>>>>> total sense) :P
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017.
>>> nov.
>>>>>>>>>>>>>>>>> 28.,
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> K,
>>>>>>>>>>>>
>>>>>>>>>>>>> 11:13):
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi Aljoscha,
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> building
>>>>>>>>>>
>>>>>>>>>>> the rc,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I seem to get this error no matter what I do:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> (shade-hadoop) on
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> invalid
>>>>>>>>>>>>
>>>>>>>>>>>>> LOC
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> header (bad signature) -> [Help 1]
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> (Apache Maven 3.3.9)
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Any idea what I am missing?
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont:
>>>>>> 2017.
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> nov.
>>>>>>>>>>>>
>>>>>>>>>>>>> 27.,
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> H, 19:35):
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Hi everyone,
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Please review and vote on release candidate #2 for the
>>>>>> version
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> 1.4.0, as
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> follows:
>>>>>>>>>>>>>>>>>>> [ ] +1, Approve the release
>>>>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide
>>> specific
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> comments)
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> The complete staging area is available for your review,
>>>>> which
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> includes:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> * JIRA release notes [1],
>>>>>>>>>>>>>>>>>>> * the official Apache source release and binary
>>> convenience
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> releases
>>>>>>>>>>>>
>>>>>>>>>>>>> to
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> key
>>>>>>>>>>
>>>>>>>>>>> with
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
>>>>>>>>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central
>>>>>> Repository
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> [4],
>>>>>>>>>>>>
>>>>>>>>>>>>> * source code tag "release-1.4.0-rc1" [5],
>>>>>>>>>>>>>>>>>>> * website pull request listing the new release [6].
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Please have a careful look at the website PR because I
>>>>>> changed
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> some
>>>>>>>>>>>>
>>>>>>>>>>>>> wording and we're now also releasing a binary without Hadoop
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> dependencies.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Please use this document for coordinating testing efforts:
>>> [7]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> The only change between RC1 and this RC2 is that the
>>> source
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> release
>>>>>>>>>>>>
>>>>>>>>>>>>> package does not include the erroneously included binary Ruby
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> dependencies
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> of the documentation anymore. Because of this I would like to
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> propose a
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> shorter voting time and close the vote around the time that
>>> RC1
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> would have
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> closed. This would mean closing by end of Wednesday. Please
>>> let
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> me
>>>>>>>>>>
>>>>>>>>>>> know if
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> you disagree with this. The vote is adopted by majority
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> approval,
>>>>>>>>>>
>>>>>>>>>>> with at
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> least 3 PMC affirmative votes.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>> Your friendly Release Manager
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> [1]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?
>>>>> proje
>>>>>>>>>>>>>>
>>>>>>>>>>>>> ctId=12315522&version=12340533
>>>>>>>>>>>>
>>>>>>>>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
>>>>>>>>>>>>>>>>>>> [3] https://dist.apache.org/repos/
>>> dist/release/flink/KEYS
>>>>>>>>>>>>>>>>>>> [4]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>> flink-1140
>>>>>>>>>>
>>>>>>>>>>> [5]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://git-wip-us.apache.org/
>>>>> repos/asf?p=flink.git;a=tag;h=
>>>>>>>>>>>>>>
>>>>>>>>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
>>>>>>>>>>>>
>>>>>>>>>>>>> [6] https://github.com/apache/flink-web/pull/95
>>>>>>>>>>>>>>>>>>> [7]
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://docs.google.com/document/d/
>>>>> 1HqYyrNoMSXwo8zBpZj7s39Uz
>>>>>>>>>>>>>>
>>>>>>>>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
>>>>>>>>>>>>
>>>>>>>>>>>>> Pro-tip: you can create a settings.xml file with these
>>> contents:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> <settings>
>>>>>>>>>>>>>>>>>>> <activeProfiles>
>>>>>>>>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
>>>>>>>>>>>>>>>>>>> </activeProfiles>
>>>>>>>>>>>>>>>>>>> <profiles>
>>>>>>>>>>>>>>>>>>> <profile>
>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>> <repositories>
>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>>>>>>
>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>>
>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>> <id>archetype</id>
>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>>>>>>
>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>>
>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>> </repositories>
>>>>>>>>>>>>>>>>>>> </profile>
>>>>>>>>>>>>>>>>>>> </profiles>
>>>>>>>>>>>>>>>>>>> </settings>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> And reference that in you maven commands via --settings
>>>>>>>>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a
>>>>>> quickstart
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> based
>>>>>>>>>>>>
>>>>>>>>>>>>> on the
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> staged release and for building against the staged jars.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>
>>>
>>