Neat! I wil try and make sure that we get the fix in ASAP.

> On 3. Dec 2017, at 12:34, Eron Wright <eronwri...@gmail.com> wrote:
> 
> Update: PR for "FLINK-8174 - Mesos RM unable to accept offers for
> unreserved resources" is ready for review.  We were able to solve (3),
> supporting any combination of reserved and unreserved resources.
> 
> On Fri, Dec 1, 2017 at 2:24 PM, Eron Wright <eronwri...@gmail.com> wrote:
> 
>> There are three levels of support we could land on.
>> 1. Flink works with unreserved resources (revert FLINK-7294).
>> 2. Flink works with unreserved resources, and correctly ignores reserved
>> resources (revert FLINK-7294 and mitigate Fenzo bug).
>> 3. Flink works with unreserved resources and reserved resources.
>> 
>> 3 is a moon shot.  Striving for 2.  Fallback on 1.
>> 
>> 
>> 
>> On Fri, Dec 1, 2017 at 2:10 PM, Aljoscha Krettek <aljos...@apache.org>
>> wrote:
>> 
>>> Thanks for the update!
>>> 
>>> Just to be clear, you're proposing going forward with the "simple fix" of
>>> reverting FLINK-7294?
>>> 
>>>> On 1. Dec 2017, at 18:39, Eron Wright <eronwri...@gmail.com> wrote:
>>>> 
>>>> Update on reported Mesos issue (FLINK-8174):
>>>> 
>>>> TLDR; a PR will be ready within 24 hours that will undo reservation
>>> support.
>>>> 
>>>> A couple of months ago, a fix (FLINK-7294) was merged related to how
>>> Flink
>>>> accepts Mesos resource offers.  The intention was to allow Flink to make
>>>> use of so-called +reserved+ resources, a Mesos feature which makes it
>>>> possible to reserve hosts for use by a specific framework/role.  The fix
>>>> inadvertently regressed the ability to use +unreserved+ resources.
>>> This is
>>>> a serious regression because unreserved resources are the common case.
>>>> 
>>>> The simple solution is to revert the earlier fix, deferring support for
>>>> reservations to another release.   We are spending some time to find a
>>> fix
>>>> that works for all scenarios, but seems unlikely at this time.   I am
>>>> reaching out to the original contributor to get their feedback.
>>>> 
>>>> In the course of the investigation, a related flaw was discovered in
>>> Fenzo
>>>> that causes Flink to misinterpret offers that contain a mix of reserved
>>> and
>>>> unreserved resources.   I believe that a small fix is possible purely
>>>> within Flink; an update to Fenzo does not appear necessary.
>>>> 
>>>> Going forward, we will contribute an improved integration test suite
>>> with
>>>> which to test Flink under diverse Mesos conditions (e.g. reservations).
>>>> 
>>>> Thanks,
>>>> Eron
>>>> 
>>>> On Thu, Nov 30, 2017 at 9:47 PM, Tzu-Li (Gordon) Tai <
>>> tzuli...@apache.org>
>>>> wrote:
>>>> 
>>>>> Hi,
>>>>> 
>>>>> I’ve noticed a behavioral regression in the Kafka producer, that should
>>>>> also be considered a blocker: https://issues.apache.org/
>>>>> jira/browse/FLINK-8181
>>>>> There’s already a PR for the issue here: https://github.com/
>>>>> apache/flink/pull/5108
>>>>> 
>>>>> Best,
>>>>> Gordon
>>>>> 
>>>>> On 30 November 2017 at 5:27:22 PM, Fabian Hueske (fhue...@gmail.com)
>>>>> wrote:
>>>>> 
>>>>> I've created a JIRA issue for the the Hadoop 2.9.0 build problem [1].
>>>>> 
>>>>> Best, Fabian
>>>>> 
>>>>> [1] https://issues.apache.org/jira/browse/FLINK-8177
>>>>> 
>>>>> 2017-11-30 4:35 GMT+01:00 Eron Wright <eronwri...@gmail.com>:
>>>>> 
>>>>>> Unfortunately we've identified a blocker bug for Flink on Mesos -
>>>>>> FLINK-8174. We'll have a patch ready on Thursday.
>>>>>> 
>>>>>> Thanks,
>>>>>> Eron
>>>>>> 
>>>>>> On Wed, Nov 29, 2017 at 3:40 PM, Eron Wright <eronwri...@gmail.com>
>>>>> wrote:
>>>>>> 
>>>>>>> On Dell EMC side, we're testing the RC2 on DCOS 1.10.0. Seeing a
>>>>>>> potential issue with offer acceptance and we'll update the thread
>>> with
>>>>> a
>>>>>> +1
>>>>>>> or with a more concrete issue within 24 hours.
>>>>>>> 
>>>>>>> Thanks,
>>>>>>> Eron
>>>>>>> 
>>>>>>> On Wed, Nov 29, 2017 at 6:54 AM, Chesnay Schepler <
>>> ches...@apache.org>
>>>>>>> wrote:
>>>>>>> 
>>>>>>>> I don't think anyone has taken a look yet, nor was there a
>>> discussion
>>>>> as
>>>>>>>> to postponing it.
>>>>>>>> 
>>>>>>>> It just slipped through the cracks i guess...
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On 29.11.2017 15:47, Gyula Fóra wrote:
>>>>>>>> 
>>>>>>>>> Hi guys,
>>>>>>>>> I ran into this again while playing with savepoint/restore
>>>>> parallelism:
>>>>>>>>> 
>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-7595
>>>>>>>>> https://github.com/apache/flink/pull/4651
>>>>>>>>> 
>>>>>>>>> Anyone has some idea about the status of this PR or were we
>>> planning
>>>>> to
>>>>>>>>> postpone this to 1.5?
>>>>>>>>> 
>>>>>>>>> Thanks,
>>>>>>>>> Gyula
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>> Fabian Hueske <fhue...@gmail.com> ezt írta (időpont: 2017. nov.
>>> 29.,
>>>>>>>>> Sze,
>>>>>>>>> 13:10):
>>>>>>>>> 
>>>>>>>>> OK, the situation is the following:
>>>>>>>>>> 
>>>>>>>>>> The test class (org.apache.flink.yarn.UtilsTest) implements a
>>>>> Hadoop
>>>>>>>>>> interface (Container) that was extended in Hadoop 2.9.0 by a
>>> getter
>>>>>> and
>>>>>>>>>> setter.
>>>>>>>>>> By adding the methods, we can compile Flink for Hadoop 2.9.0.
>>>>> However,
>>>>>>>>>> the
>>>>>>>>>> getter/setter add a dependency on a class that was also added in
>>>>>> Hadoop
>>>>>>>>>> 2.9.0.
>>>>>>>>>> Therefore, the implementation is not backwards compatible with
>>>>> Hadoop
>>>>>>>>>> versions < 2.9.0.
>>>>>>>>>> 
>>>>>>>>>> Not sure how we can fix the problem. We would need two version of
>>>>> the
>>>>>>>>>> class
>>>>>>>>>> that are chosen based on the Hadoop version. Do we have something
>>>>> like
>>>>>>>>>> that
>>>>>>>>>> somewhere else?
>>>>>>>>>> 
>>>>>>>>>> Since this is only a problem in a test class, Flink 1.4.0 might
>>>>> still
>>>>>>>>>> work
>>>>>>>>>> very well with Hadoop 2.9.0.
>>>>>>>>>> However, this has not been tested AFAIK.
>>>>>>>>>> 
>>>>>>>>>> Cheers, Fabian
>>>>>>>>>> 
>>>>>>>>>> 2017-11-29 12:47 GMT+01:00 Fabian Hueske <fhue...@gmail.com>:
>>>>>>>>>> 
>>>>>>>>>> I just tried to build the release-1.4 branch for Hadoop 2.9.0
>>>>>> (released
>>>>>>>>>>> a
>>>>>>>>>>> few days ago) and got a compilation failure in a test class.
>>>>>>>>>>> 
>>>>>>>>>>> Right now, I'm assessing how much we need to fix to support
>>> Hadoop
>>>>>>>>>>> 2.9.0.
>>>>>>>>>>> I'll report later.
>>>>>>>>>>> 
>>>>>>>>>>> Best, Fabian
>>>>>>>>>>> 
>>>>>>>>>>> 2017-11-29 11:16 GMT+01:00 Aljoscha Krettek <aljos...@apache.org
>>>> :
>>>>>>>>>>> 
>>>>>>>>>>> Agreed, this is a regression compared to the previous
>>>>> functionality.
>>>>>> I
>>>>>>>>>>>> updated the issue to "Blocker".
>>>>>>>>>>>> 
>>>>>>>>>>>> On 29. Nov 2017, at 10:01, Gyula Fóra <gyula.f...@gmail.com>
>>>>> wrote:
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Hi all,
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I have found the following issue:
>>>>>>>>>>>>> https://issues.apache.org/jira/browse/FLINK-8165
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I would say this is a blocker (I personally pass the
>>>>> ParameterTool
>>>>>>>>>>>>> all
>>>>>>>>>>>>> 
>>>>>>>>>>>> over
>>>>>>>>>>>> 
>>>>>>>>>>>>> the place in my production apps), but a pretty trivial issue to
>>>>>> fix,
>>>>>>>>>>>>> 
>>>>>>>>>>>> we
>>>>>>>>>> 
>>>>>>>>>>> can
>>>>>>>>>>>> 
>>>>>>>>>>>>> wait a little to find other potential problems.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> I can submit a fix in a little bit.
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Tzu-Li (Gordon) Tai <tzuli...@apache.org> ezt írta (időpont:
>>>>> 2017.
>>>>>>>>>>>>> 
>>>>>>>>>>>> nov.
>>>>>>>>>> 
>>>>>>>>>>> 29., Sze, 9:23):
>>>>>>>>>>>>> 
>>>>>>>>>>>>> +1
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>> - No missing release Maven artifacts
>>>>>>>>>>>>>> - Staged Apache source & binary convenience releases looks
>>> good
>>>>>>>>>>>>>> - NOTICE / LICENSE is correct, README is sane
>>>>>>>>>>>>>> - Built from source (macOS, Scala 2.11, Hadoop-free & Hadoop
>>>>> 2.8)
>>>>>>>>>>>>>> - Cluster testing on AWS EMR (see release-testing-doc for
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> configuration
>>>>>>>>>> 
>>>>>>>>>>> details)
>>>>>>>>>>>>>> - Tested Kinesis / Elasticsearch connector (no dependency
>>>>> clashes
>>>>>> on
>>>>>>>>>>>>>> cluster execution, works locally in IDE)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Thanks a lot for managing the release Aljoscha!
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Cheers,
>>>>>>>>>>>>>> Gordon
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> On 28 November 2017 at 8:32:42 PM, Stefan Richter (
>>>>>>>>>>>>>> s.rich...@data-artisans.com) wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> +1 (non-binding)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> I tested Flink in a cluster setup on Google Cloud,
>>> YARN-per-job,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> checked
>>>>>>>>>>>> 
>>>>>>>>>>>>> that for all backends that HA, recovery, at-least-once,
>>>>> end-to-end
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> exactly
>>>>>>>>>>>> 
>>>>>>>>>>>>> once (with Kafka11 Producer), savepoints, externalized
>>>>> checkpoints,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> and
>>>>>>>>>> 
>>>>>>>>>>> rescaling work correctly.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> Am 28.11.2017 um 11:47 schrieb Aljoscha Krettek <
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> aljos...@apache.org
>>>>>>>>>> 
>>>>>>>>>>> :
>>>>>>>>>>>>> 
>>>>>>>>>>>>>> +1
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Verified:
>>>>>>>>>>>>>>> - NOTICE and LICENSE are correct
>>>>>>>>>>>>>>> - source doesn't contain binaries
>>>>>>>>>>>>>>> - verified signatures
>>>>>>>>>>>>>>> - verified hashes
>>>>>>>>>>>>>>> - cluster testing on AWS and Cloudera VM (with Kerberos) (see
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> release-testing doc)
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:20, Aljoscha Krettek <
>>>>> aljos...@apache.org
>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Phew, thanks for the update!
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> On 28. Nov 2017, at 11:19, Gyula Fóra <gyf...@apache.org>
>>>>>> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Ok seems like I had to remove the snappy jar as it was
>>>>>> corrupted
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> (makes
>>>>>>>>>>>> 
>>>>>>>>>>>>> total sense) :P
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Gyula Fóra <gyf...@apache.org> ezt írta (időpont: 2017.
>>> nov.
>>>>>>>>>>>>>>>>> 28.,
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> K,
>>>>>>>>>>>> 
>>>>>>>>>>>>> 11:13):
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Hi Aljoscha,
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Thanks for the release candidate. I am having a hard time
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> building
>>>>>>>>>> 
>>>>>>>>>>> the rc,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> I seem to get this error no matter what I do:
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal
>>>>>>>>>>>>>>>>>> org.apache.maven.plugins:maven-shade-plugin:2.4.1:shade
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> (shade-hadoop) on
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> project flink-shaded-hadoop2-uber: Error creating shaded jar:
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> invalid
>>>>>>>>>>>> 
>>>>>>>>>>>>> LOC
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> header (bad signature) -> [Help 1]
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> (Apache Maven 3.3.9)
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Any idea what I am missing?
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>> Gyula
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Aljoscha Krettek <aljos...@apache.org> ezt írta (időpont:
>>>>>> 2017.
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> nov.
>>>>>>>>>>>> 
>>>>>>>>>>>>> 27.,
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> H, 19:35):
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> Hi everyone,
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Please review and vote on release candidate #2 for the
>>>>>> version
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 1.4.0, as
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> follows:
>>>>>>>>>>>>>>>>>>> [ ] +1, Approve the release
>>>>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide
>>> specific
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> comments)
>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> The complete staging area is available for your review,
>>>>> which
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> includes:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> * JIRA release notes [1],
>>>>>>>>>>>>>>>>>>> * the official Apache source release and binary
>>> convenience
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> releases
>>>>>>>>>>>> 
>>>>>>>>>>>>> to
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> be deployed to dist.apache.org[2], which are signed with the
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> key
>>>>>>>>>> 
>>>>>>>>>>> with
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> fingerprint F2A67A8047499BBB3908D17AA8F4FD97121D7293 [3],
>>>>>>>>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central
>>>>>> Repository
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> [4],
>>>>>>>>>>>> 
>>>>>>>>>>>>> * source code tag "release-1.4.0-rc1" [5],
>>>>>>>>>>>>>>>>>>> * website pull request listing the new release [6].
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Please have a careful look at the website PR because I
>>>>>> changed
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> some
>>>>>>>>>>>> 
>>>>>>>>>>>>> wording and we're now also releasing a binary without Hadoop
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> dependencies.
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Please use this document for coordinating testing efforts:
>>> [7]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> The only change between RC1 and this RC2 is that the
>>> source
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> release
>>>>>>>>>>>> 
>>>>>>>>>>>>> package does not include the erroneously included binary Ruby
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> dependencies
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> of the documentation anymore. Because of this I would like to
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> propose a
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> shorter voting time and close the vote around the time that
>>> RC1
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> would have
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> closed. This would mean closing by end of Wednesday. Please
>>> let
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> me
>>>>>>>>>> 
>>>>>>>>>>> know if
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> you disagree with this. The vote is adopted by majority
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> approval,
>>>>>>>>>> 
>>>>>>>>>>> with at
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> least 3 PMC affirmative votes.
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>> Your friendly Release Manager
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> [1]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?
>>>>> proje
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> ctId=12315522&version=12340533
>>>>>>>>>>>> 
>>>>>>>>>>>>> [2] http://people.apache.org/~aljoscha/flink-1.4.0-rc2/
>>>>>>>>>>>>>>>>>>> [3] https://dist.apache.org/repos/
>>> dist/release/flink/KEYS
>>>>>>>>>>>>>>>>>>> [4]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>> flink-1140
>>>>>>>>>> 
>>>>>>>>>>> [5]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://git-wip-us.apache.org/
>>>>> repos/asf?p=flink.git;a=tag;h=
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> ea751b7b23b23446ed3fcdeed564bbe8bf4adf9c
>>>>>>>>>>>> 
>>>>>>>>>>>>> [6] https://github.com/apache/flink-web/pull/95
>>>>>>>>>>>>>>>>>>> [7]
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://docs.google.com/document/d/
>>>>> 1HqYyrNoMSXwo8zBpZj7s39Uz
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> UdlFcFO8TRpHNZ_cl44/edit?usp=sharing
>>>>>>>>>>>> 
>>>>>>>>>>>>> Pro-tip: you can create a settings.xml file with these
>>> contents:
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> <settings>
>>>>>>>>>>>>>>>>>>> <activeProfiles>
>>>>>>>>>>>>>>>>>>> <activeProfile>flink-1.4.0</activeProfile>
>>>>>>>>>>>>>>>>>>> </activeProfiles>
>>>>>>>>>>>>>>>>>>> <profiles>
>>>>>>>>>>>>>>>>>>> <profile>
>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>> <repositories>
>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>> <id>flink-1.4.0</id>
>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>> 
>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>> <repository>
>>>>>>>>>>>>>>>>>>> <id>archetype</id>
>>>>>>>>>>>>>>>>>>> <url>
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> https://repository.apache.org/
>>>>> content/repositories/orgapache
>>>>>>>>>>>>>> 
>>>>>>>>>>>>> flink-1140/
>>>>>>>>>>>> 
>>>>>>>>>>>>> </url>
>>>>>>>>>>>>>>>>>>> </repository>
>>>>>>>>>>>>>>>>>>> </repositories>
>>>>>>>>>>>>>>>>>>> </profile>
>>>>>>>>>>>>>>>>>>> </profiles>
>>>>>>>>>>>>>>>>>>> </settings>
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> And reference that in you maven commands via --settings
>>>>>>>>>>>>>>>>>>> path/to/settings.xml. This is useful for creating a
>>>>>> quickstart
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> based
>>>>>>>>>>>> 
>>>>>>>>>>>>> on the
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> staged release and for building against the staged jars.
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>>> 
>>> 
>>> 
>> 

Reply via email to