Hi Kurt, With the same argument as before, given that it is mentioned in the release announcement that it is a preview feature, I would not block this release because of it. Nevertheless, it would be important to mention this explicitly in the release notes [1].
Regards, Gordon [1] https://github.com/apache/flink/pull/9438 On Thu, Aug 15, 2019 at 11:29 AM Timo Walther <twal...@apache.org> wrote: > Hi Kurt, > > I agree that this is a serious bug. However, I would not block the > release because of this. As you said, there is a workaround and the > `execute()` works in the most common case of a single execution. We can > fix this in a minor release shortly after. > > What do others think? > > Regards, > Timo > > > Am 15.08.19 um 11:23 schrieb Kurt Young: > > HI, > > > > We just find a serious bug around blink planner: > > https://issues.apache.org/jira/browse/FLINK-13708 > > When user reused the table environment instance, and call `execute` > method > > multiple times for > > different sql, the later call will trigger the earlier ones to be > > re-executed. > > > > It's a serious bug but seems we also have a work around, which is never > > reuse the table environment > > object. I'm not sure if we should treat this one as blocker issue of > 1.9.0. > > > > What's your opinion? > > > > Best, > > Kurt > > > > > > On Thu, Aug 15, 2019 at 2:01 PM Gary Yao <g...@ververica.com> wrote: > > > >> +1 (non-binding) > >> > >> Jepsen test suite passed 10 times consecutively > >> > >> On Wed, Aug 14, 2019 at 5:31 PM Aljoscha Krettek <aljos...@apache.org> > >> wrote: > >> > >>> +1 > >>> > >>> I did some testing on a Google Cloud Dataproc cluster (it gives you a > >>> managed YARN and Google Cloud Storage (GCS)): > >>> - tried both YARN session mode and YARN per-job mode, also using > >>> bin/flink list/cancel/etc. against a YARN session cluster > >>> - ran examples that write to GCS, both with the native Hadoop > >> FileSystem > >>> and a custom “plugin” FileSystem > >>> - ran stateful streaming jobs that use GCS as a checkpoint backend > >>> - tried running SQL programs on YARN using the SQL Cli: this worked > for > >>> YARN session mode but not for YARN per-job mode. Looking at the code I > >>> don’t think per-job mode would work from seeing how it is implemented. > >> But > >>> I think it’s an OK restriction to have for now > >>> - in all the testing I had fine-grained recovery (region failover) > >>> enabled but I didn’t simulate any failures > >>> > >>>> On 14. Aug 2019, at 15:20, Kurt Young <ykt...@gmail.com> wrote: > >>>> > >>>> Hi, > >>>> > >>>> Thanks for preparing this release candidate. I have verified the > >>> following: > >>>> - verified the checksums and GPG files match the corresponding release > >>> files > >>>> - verified that the source archives do not contains any binaries > >>>> - build the source release with Scala 2.11 successfully. > >>>> - ran `mvn verify` locally, met 2 issuses [FLINK-13687] and > >>> [FLINK-13688], > >>>> but > >>>> both are not release blockers. Other than that, all tests are passed. > >>>> - ran all e2e tests which don't need download external packages (it's > >>> very > >>>> unstable > >>>> in China and almost impossible to download them), all passed. > >>>> - started local cluster, ran some examples. Met a small website > display > >>>> issue > >>>> [FLINK-13591], which is also not a release blocker. > >>>> > >>>> Although we have pushed some fixes around blink planner and hive > >>>> integration > >>>> after RC2, but consider these are both preview features, I'm lean to > be > >>> ok > >>>> to release > >>>> without these fixes. > >>>> > >>>> +1 from my side. (binding) > >>>> > >>>> Best, > >>>> Kurt > >>>> > >>>> > >>>> On Wed, Aug 14, 2019 at 5:13 PM Jark Wu <imj...@gmail.com> wrote: > >>>> > >>>>> Hi Gordon, > >>>>> > >>>>> I have verified the following things: > >>>>> > >>>>> - build the source release with Scala 2.12 and Scala 2.11 > successfully > >>>>> - checked/verified signatures and hashes > >>>>> - checked that all POM files point to the same version > >>>>> - ran some flink table related end-to-end tests locally and succeeded > >>>>> (except TPC-H e2e failed which is reported in FLINK-13704) > >>>>> - started cluster for both Scala 2.11 and 2.12, ran examples, > verified > >>> web > >>>>> ui and log output, nothing unexpected > >>>>> - started cluster, ran a SQL query to temporal join with kafka source > >>> and > >>>>> mysql jdbc table, and write results to kafka again. Using DDL to > >> create > >>> the > >>>>> source and sinks. looks good. > >>>>> - reviewed the release PR > >>>>> > >>>>> As FLINK-13704 is not recognized as blocker issue, so +1 from my side > >>>>> (non-binding). > >>>>> > >>>>> On Tue, 13 Aug 2019 at 17:07, Till Rohrmann <trohrm...@apache.org> > >>> wrote: > >>>>>> Hi Richard, > >>>>>> > >>>>>> although I can see that it would be handy for users who have PubSub > >> set > >>>>> up, > >>>>>> I would rather not include examples which require an external > >>> dependency > >>>>>> into the Flink distribution. I think examples should be > >> self-contained. > >>>>> My > >>>>>> concern is that we would bloat the distribution for many users at > the > >>>>>> benefit of a few. Instead, I think it would be better to make these > >>>>>> examples available differently, maybe through Flink's ecosystem > >> website > >>>>> or > >>>>>> maybe a new examples section in Flink's documentation. > >>>>>> > >>>>>> Cheers, > >>>>>> Till > >>>>>> > >>>>>> On Tue, Aug 13, 2019 at 9:43 AM Jark Wu <imj...@gmail.com> wrote: > >>>>>> > >>>>>>> Hi Till, > >>>>>>> > >>>>>>> After thinking about we can use VARCHAR as an alternative of > >>>>>>> timestamp/time/date. > >>>>>>> I'm fine with not recognize it as a blocker issue. > >>>>>>> We can fix it into 1.9.1. > >>>>>>> > >>>>>>> > >>>>>>> Thanks, > >>>>>>> Jark > >>>>>>> > >>>>>>> > >>>>>>> On Tue, 13 Aug 2019 at 15:10, Richard Deurwaarder <rich...@xeli.eu > > > >>>>>> wrote: > >>>>>>>> Hello all, > >>>>>>>> > >>>>>>>> I noticed the PubSub example jar is not included in the examples/ > >> dir > >>>>>> of > >>>>>>>> flink-dist. I've created > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13700 > >>>>>>>> + https://github.com/apache/flink/pull/9424/files to fix this. > >>>>>>>> > >>>>>>>> I will leave it up to you to decide if we want to add this to > >> 1.9.0. > >>>>>>>> Regards, > >>>>>>>> > >>>>>>>> Richard > >>>>>>>> > >>>>>>>> On Tue, Aug 13, 2019 at 9:04 AM Till Rohrmann < > >> trohrm...@apache.org> > >>>>>>>> wrote: > >>>>>>>> > >>>>>>>>> Hi Jark, > >>>>>>>>> > >>>>>>>>> thanks for reporting this issue. Could this be a documented > >>>>>> limitation > >>>>>>> of > >>>>>>>>> Blink's preview version? I think we have agreed that the Blink > SQL > >>>>>>>> planner > >>>>>>>>> will be rather a preview feature than production ready. Hence it > >>>>>> could > >>>>>>>>> still contain some bugs. My concern is that there might be still > >>>>>> other > >>>>>>>>> issues which we'll discover bit by bit and could postpone the > >>>>> release > >>>>>>>> even > >>>>>>>>> further if we say Blink bugs are blockers. > >>>>>>>>> > >>>>>>>>> Cheers, > >>>>>>>>> Till > >>>>>>>>> > >>>>>>>>> On Tue, Aug 13, 2019 at 7:42 AM Jark Wu <imj...@gmail.com> > wrote: > >>>>>>>>> > >>>>>>>>>> Hi all, > >>>>>>>>>> > >>>>>>>>>> I just find an issue when testing connector DDLs against blink > >>>>>>> planner > >>>>>>>>> for > >>>>>>>>>> rc2. > >>>>>>>>>> This issue lead to the DDL doesn't work when containing > >>>>>>>>> timestamp/date/time > >>>>>>>>>> type. > >>>>>>>>>> I have created an issue FLINK-13699[1] and a pull request for > >>>>> this. > >>>>>>>>>> IMO, this can be a blocker issue of 1.9 release. Because > >>>>>>>>>> timestamp/date/time are primitive types, and this will break the > >>>>>> DDL > >>>>>>>>>> feature. > >>>>>>>>>> However, I want to hear more thoughts from the community whether > >>>>> we > >>>>>>>>> should > >>>>>>>>>> recognize it as a blocker. > >>>>>>>>>> > >>>>>>>>>> Thanks, > >>>>>>>>>> Jark > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> [1]: https://issues.apache.org/jira/browse/FLINK-13699 > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> > >>>>>>>>>> On Mon, 12 Aug 2019 at 22:46, Becket Qin <becket....@gmail.com> > >>>>>>> wrote: > >>>>>>>>>>> Thanks Gordon, will do that. > >>>>>>>>>>> > >>>>>>>>>>> On Mon, Aug 12, 2019 at 4:42 PM Tzu-Li (Gordon) Tai < > >>>>>>>>> tzuli...@apache.org > >>>>>>>>>>> wrote: > >>>>>>>>>>> > >>>>>>>>>>>> Concerning FLINK-13231: > >>>>>>>>>>>> > >>>>>>>>>>>> Since this is a @PublicEvolving interface, technically it is > >>>>> ok > >>>>>>> to > >>>>>>>>>> break > >>>>>>>>>>>> it across releases (including across bugfix releases?). > >>>>>>>>>>>> So, @Becket if you do merge it now, please mark the fix > >>>>> version > >>>>>>> as > >>>>>>>>>> 1.9.1. > >>>>>>>>>>>> During the voting process, in the case a new RC is created, > >>>>> we > >>>>>>>>> usually > >>>>>>>>>>>> check the list of changes compared to the previous RC, and > >>>>>>> correct > >>>>>>>>> the > >>>>>>>>>>> "Fix > >>>>>>>>>>>> Version" of the corresponding JIRAs to be the right version > >>>>> (in > >>>>>>> the > >>>>>>>>>> case, > >>>>>>>>>>>> it would be corrected to 1.9.0 instead of 1.9.1). > >>>>>>>>>>>> > >>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:25 PM Till Rohrmann < > >>>>>>>> trohrm...@apache.org> > >>>>>>>>>>>> wrote: > >>>>>>>>>>>> > >>>>>>>>>>>>> I agree that it would be nicer. Not sure whether we should > >>>>>>> cancel > >>>>>>>>> the > >>>>>>>>>> RC > >>>>>>>>>>>>> for this issue given that it is open for quite some time and > >>>>>>>> hasn't > >>>>>>>>>> been > >>>>>>>>>>>>> addressed until very recently. Maybe we could include it on > >>>>>> the > >>>>>>>>>>> shortlist > >>>>>>>>>>>>> of nice-to-do things which we do in case that the RC gets > >>>>>>>> cancelled. > >>>>>>>>>>>>> Cheers, > >>>>>>>>>>>>> Till > >>>>>>>>>>>>> > >>>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:18 PM Becket Qin < > >>>>>>> becket....@gmail.com> > >>>>>>>>>>> wrote: > >>>>>>>>>>>>>> Hi Till, > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Yes, I think we have already documented in that way. So > >>>>>>>> technically > >>>>>>>>>>>>>> speaking it is fine to change it later. It is just better > >>>>> if > >>>>>> we > >>>>>>>>> could > >>>>>>>>>>>>>> avoid > >>>>>>>>>>>>>> doing that. > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Thanks, > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> Jiangjie (Becket) Qin > >>>>>>>>>>>>>> > >>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:09 PM Till Rohrmann < > >>>>>>>>> trohrm...@apache.org> > >>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>> > >>>>>>>>>>>>>>> Could we say that the PubSub connector is public evolving > >>>>>>>>> instead? > >>>>>>>>>>>>>>> Cheers, > >>>>>>>>>>>>>>> Till > >>>>>>>>>>>>>>> > >>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 3:18 PM Becket Qin < > >>>>>>>> becket....@gmail.com > >>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>> Hi all, > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> FLINK-13231(palindrome!) has a minor Google PubSub > >>>>>>> connector > >>>>>>>>> API > >>>>>>>>>>>>>> change > >>>>>>>>>>>>>>>> regarding how to config rate limiting. The GCP PubSub > >>>>>>>> connector > >>>>>>>>>> is > >>>>>>>>>>> a > >>>>>>>>>>>>>>> newly > >>>>>>>>>>>>>>>> introduced connector in 1.9, so it would be nice to > >>>>>> include > >>>>>>>>> this > >>>>>>>>>>>>>> change > >>>>>>>>>>>>>>>> into 1.9 rather than later to avoid a public API > >>>>> change. > >>>>>> I > >>>>>>> am > >>>>>>>>>>>>>> thinking of > >>>>>>>>>>>>>>>> making this as a blocker for 1.9. Want to check what do > >>>>>>>> others > >>>>>>>>>>> think. > >>>>>>>>>>>>>>>> Thanks, > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> Jiangjie (Becket) Qin > >>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 2:04 PM Zili Chen < > >>>>>>>>> wander4...@gmail.com> > >>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>> Hi Kurt, > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Thanks for your explanation. For [1] I think at least > >>>>>> we > >>>>>>>>> should > >>>>>>>>>>>>>> change > >>>>>>>>>>>>>>>>> the JIRA issue field, like unset the fixed version. > >>>>> For > >>>>>>>> [2] I > >>>>>>>>>> can > >>>>>>>>>>>>>> see > >>>>>>>>>>>>>>>>> the change is all in test scope but wonder if such a > >>>>>>> commit > >>>>>>>>>> still > >>>>>>>>>>>>>>> invalid > >>>>>>>>>>>>>>>>> the release candidate. IIRC previous RC VOTE threads > >>>>>>> would > >>>>>>>>>>> contain > >>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>> release manual/guide, I will try to look up it, too. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Best, > >>>>>>>>>>>>>>>>> tison. > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>> Kurt Young <ykt...@gmail.com> 于2019年8月12日周一 > >>>>> 下午5:42写道: > >>>>>>>>>>>>>>>>>> Hi Zili, > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Thanks for the heads up. The 2 issues you mentioned > >>>>>>> were > >>>>>>>>>> opened > >>>>>>>>>>>>>> by > >>>>>>>>>>>>>>> me. > >>>>>>>>>>>>>>>> We > >>>>>>>>>>>>>>>>>> have > >>>>>>>>>>>>>>>>>> found the reason of the second issue and a PR was > >>>>>>> opened > >>>>>>>>> for > >>>>>>>>>>> it. > >>>>>>>>>>>>>> As > >>>>>>>>>>>>>>>> said > >>>>>>>>>>>>>>>>> in > >>>>>>>>>>>>>>>>>> jira, the > >>>>>>>>>>>>>>>>>> issue was just a testing problem, should not be > >>>>>> blocker > >>>>>>>> of > >>>>>>>>>>> 1.9.0 > >>>>>>>>>>>>>>>> release. > >>>>>>>>>>>>>>>>>> However, > >>>>>>>>>>>>>>>>>> we will still merge it into 1.9 branch. > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> Best, > >>>>>>>>>>>>>>>>>> Kurt > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 5:38 PM Zili Chen < > >>>>>>>>>>> wander4...@gmail.com> > >>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>> Hi, > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> I just noticed that a few hours ago there were > >>>>> two > >>>>>>> new > >>>>>>>>>> issues > >>>>>>>>>>>>>>>>>>> filed and marked as blockers to 1.9.0[1][2]. > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> Now [1] is closed as duplication but still marked > >>>>>> as > >>>>>>>>>>>>>>>>>>> a blocker to 1.9.0, while [2] is downgrade to > >>>>>> "Major" > >>>>>>>>>>> priority > >>>>>>>>>>>>>>>>>>> but still target to be fixed in 1.9.0. > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> It would be worth to have attention of our > >>>>> release > >>>>>>>>> manager > >>>>>>>>>> at > >>>>>>>>>>>>>>> least. > >>>>>>>>>>>>>>>>>>> Best, > >>>>>>>>>>>>>>>>>>> tison. > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> [1] > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13687 > >>>>>>>>>>>>>>>>>>> [2] > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13688 > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>> Gyula Fóra <gyula.f...@gmail.com> 于2019年8月12日周一 > >>>>>>>>> 下午5:10写道: > >>>>>>>>>>>>>>>>>>>> Thanks Stephan :) > >>>>>>>>>>>>>>>>>>>> That looks easy enough, will try! > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> Gyula > >>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 11:00 AM Stephan Ewen < > >>>>>>>>>>>>>> se...@apache.org> > >>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>> Hi Gyula! > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Thanks for reporting this. > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Can you try to simply build Flink without > >>>>>> Hadoop > >>>>>>>> and > >>>>>>>>>> then > >>>>>>>>>>>>>>>> exporting > >>>>>>>>>>>>>>>>>>>>> HADOOP_CLASSPATH to your CloudEra libs? > >>>>>>>>>>>>>>>>>>>>> That is the recommended way these days. > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> Best, > >>>>>>>>>>>>>>>>>>>>> Stephan > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 10:48 AM Gyula Fóra < > >>>>>>>>>>>>>>>> gyula.f...@gmail.com> > >>>>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>>> Thanks Dawid, > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> In the meantime I also figured out that I > >>>>>> need > >>>>>>> to > >>>>>>>>>> build > >>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/flink-shaded > >>>>>> project > >>>>>>>>>> locally > >>>>>>>>>>>>>> with > >>>>>>>>>>>>>>>>>>>>>> -Dhadoop.version set to the specific hadoop > >>>>>>>> version > >>>>>>>>>> if > >>>>>>>>>>> I > >>>>>>>>>>>>>> want > >>>>>>>>>>>>>>>>>>> something > >>>>>>>>>>>>>>>>>>>>>> different. > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> Cheers, > >>>>>>>>>>>>>>>>>>>>>> Gyula > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 9:54 AM Dawid > >>>>>>> Wysakowicz > >>>>>>>> < > >>>>>>>>>>>>>>>>>>>> dwysakow...@apache.org > >>>>>>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> Hi Gyula, > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> As for the issues with mapr maven > >>>>>> repository, > >>>>>>>> you > >>>>>>>>>>> might > >>>>>>>>>>>>>>> have > >>>>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>>> look > >>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>>>>>>>>>>> this message: > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> > >> > https://lists.apache.org/thread.html/77f4db930216e6da0d6121065149cef43ff3ea33c9ffe9b1a3047210@%3Cdev.flink.apache.org%3E > >>>>>>>>>>>>>>>>>>>>>>> Try using the "unsafe-mapr-repo" profile. > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> Best, > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> Dawid > >>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>> On 11/08/2019 19:31, Gyula Fóra wrote: > >>>>>>>>>>>>>>>>>>>>>>>> Hi again, > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> How do I build the RC locally with the > >>>>>>> hadoop > >>>>>>>>>>> version > >>>>>>>>>>>>>>>>>> specified? > >>>>>>>>>>>>>>>>>>>>> Seems > >>>>>>>>>>>>>>>>>>>>>>> like > >>>>>>>>>>>>>>>>>>>>>>>> no matter what I do I run into > >>>>> dependency > >>>>>>>>>> problems > >>>>>>>>>>>>>> with > >>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>> shaded > >>>>>>>>>>>>>>>>>>>>>> hadoop > >>>>>>>>>>>>>>>>>>>>>>>> dependencies. > >>>>>>>>>>>>>>>>>>>>>>>> This seems to have worked in the past. > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> There might be some documentation > >>>>>> somewhere > >>>>>>>>> that > >>>>>>>>>> I > >>>>>>>>>>>>>>> couldnt > >>>>>>>>>>>>>>>>>> find, > >>>>>>>>>>>>>>>>>>>> so I > >>>>>>>>>>>>>>>>>>>>>>> would > >>>>>>>>>>>>>>>>>>>>>>>> appreciate any pointers :) > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> Thanks! > >>>>>>>>>>>>>>>>>>>>>>>> Gyula > >>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>> On Sun, Aug 11, 2019 at 6:57 PM Gyula > >>>>>> Fóra > >>>>>>> < > >>>>>>>>>>>>>>>>>> gyula.f...@gmail.com > >>>>>>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>>>>>> Hi! > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> I am trying to build 1.9.0-rc2 with > >>>>> the > >>>>>>>>>>>>>> -Pvendor-repos > >>>>>>>>>>>>>>>>> profile > >>>>>>>>>>>>>>>>>>>>>> enabled. > >>>>>>>>>>>>>>>>>>>>>>> I > >>>>>>>>>>>>>>>>>>>>>>>>> get the following error: > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> mvn clean install -DskipTests > >>>>>>> -Pvendor-repos > >>>>>>>>>>>>>>>>>>>> -Dhadoop.version=2.6.0 > >>>>>>>>>>>>>>>>>>>>>>>>> -Pinclude-hadoop (ignore that the > >>>>> hadoop > >>>>>>>>> version > >>>>>>>>>>> is > >>>>>>>>>>>>>> not > >>>>>>>>>>>>>>> a > >>>>>>>>>>>>>>>>>> vendor > >>>>>>>>>>>>>>>>>>>>>> hadoop > >>>>>>>>>>>>>>>>>>>>>>>>> version) > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal on > >>>>>> project > >>>>>>>>>>>>>>> flink-hadoop-fs: > >>>>>>>>>>>>>>>>>> Could > >>>>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>>>>> resolve dependencies for project > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>> org.apache.flink:flink-hadoop-fs:jar:1.9.0: > >>>>>>>>>> Failed > >>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>> collect > >>>>>>>>>>>>>>>>>>>>>>> dependencies > >>>>>>>>>>>>>>>>>>>>>>>>> at > >>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0: > >>>>>>>>>>>>>>>>>> Failed > >>>>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>>>>>> read > >>>>>>>>>>>>>>>>>>>>>>>>> artifact descriptor for > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0: > >>>>>>>>>>>>>>>> Could > >>>>>>>>>>>>>>>>>> not > >>>>>>>>>>>>>>>>>>>>>> transfer > >>>>>>>>>>>>>>>>>>>>>>>>> artifact > >>>>>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:pom:2.6.0-7.0 > >>>>>>>>>>>>>>>>>>>>> from/to > >>>>>>>>>>>>>>>>>>>>>>>>> mapr-releases ( > >>>>>>>>>> https://repository.mapr.com/maven/ > >>>>>>>>>>> ): > >>>>>> sun.security.validator.ValidatorException: > >>>>>>>>> PKIX > >>>>>>>>>>> path > >>>>>>>>>>>>>>>>> building > >>>>>>>>>>>>>>>>>>>>> failed: > >>>>>> sun.security.provider.certpath.SunCertPathBuilderException: > >>>>>>>>>>>>>>>>>>> unable > >>>>>>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>>>>>>> find > >>>>>>>>>>>>>>>>>>>>>>>>> valid certification path to requested > >>>>>>> target > >>>>>>>>> -> > >>>>>>>>>>>>>> [Help 1] > >>>>>>>>>>>>>>>>>>>>>>>>> This looks like a TLS error. Might not > >>>>>> be > >>>>>>>>>> related > >>>>>>>>>>>>>> to the > >>>>>>>>>>>>>>>>>> release > >>>>>>>>>>>>>>>>>>>> but > >>>>>>>>>>>>>>>>>>>>>> it > >>>>>>>>>>>>>>>>>>>>>>>>> could be good to know. > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> Cheers, > >>>>>>>>>>>>>>>>>>>>>>>>> Gyula > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Aug 9, 2019 at 6:26 PM Tzu-Li > >>>>>>>> (Gordon) > >>>>>>>>>>> Tai < > >>>>>>>>>>>>>>>>>>>>>> tzuli...@apache.org > >>>>>>>>>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> Please note that the unresolved > >>>>> issues > >>>>>>> that > >>>>>>>>> are > >>>>>>>>>>>>>> still > >>>>>>>>>>>>>>>>> tagged > >>>>>>>>>>>>>>>>>>>> with a > >>>>>>>>>>>>>>>>>>>>>> fix > >>>>>>>>>>>>>>>>>>>>>>>>>> version "1.9.0", as seen in the JIRA > >>>>>>>> release > >>>>>>>>>>> notes > >>>>>>>>>>>>>> [1], > >>>>>>>>>>>>>>>> are > >>>>>>>>>>>>>>>>>>>> issues > >>>>>>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>>>>>>>>>> update documents for new features. > >>>>>>>>>>>>>>>>>>>>>>>>>> I've left them still associated with > >>>>>>> 1.9.0 > >>>>>>>>>> since > >>>>>>>>>>>>>> these > >>>>>>>>>>>>>>>>> should > >>>>>>>>>>>>>>>>>>>> still > >>>>>>>>>>>>>>>>>>>>>> be > >>>>>>>>>>>>>>>>>>>>>>>>>> updated for 1.9.0 soon along with the > >>>>>>>>> official > >>>>>>>>>>>>>> release. > >>>>>>>>>>>>>>>>>>>>>>>>>> [1] > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>> > >> > https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601 > >>>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Aug 9, 2019 at 6:17 PM Tzu-Li > >>>>>>>>> (Gordon) > >>>>>>>>>>> Tai > >>>>>>>>>>>>>> < > >>>>>>>>>>>>>>>>>>>>>>> tzuli...@apache.org> > >>>>>>>>>>>>>>>>>>>>>>>>>> wrote: > >>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Hi all, > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Release candidate #2 for Apache > >>>>> Flink > >>>>>>>> 1.9.0 > >>>>>>>>> is > >>>>>>>>>>> now > >>>>>>>>>>>>>>> ready > >>>>>>>>>>>>>>>>> for > >>>>>>>>>>>>>>>>>>>> your > >>>>>>>>>>>>>>>>>>>>>>>>>> review. > >>>>>>>>>>>>>>>>>>>>>>>>>>> This is the first voting candidate > >>>>> for > >>>>>>>>> 1.9.0, > >>>>>>>>>>>>>>> following > >>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>> preview > >>>>>>>>>>>>>>>>>>>>>>>>>>> candidates RC0 and RC1. > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Please review and vote on release > >>>>>>>> candidate > >>>>>>>>> #2 > >>>>>>>>>>> for > >>>>>>>>>>>>>>>> version > >>>>>>>>>>>>>>>>>>>> 1.9.0, > >>>>>>>>>>>>>>>>>>>>> as > >>>>>>>>>>>>>>>>>>>>>>>>>>> follows: > >>>>>>>>>>>>>>>>>>>>>>>>>>> [ ] +1, Approve the release > >>>>>>>>>>>>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release > >>>>>>> (please > >>>>>>>>>>> provide > >>>>>>>>>>>>>>>>> specific > >>>>>>>>>>>>>>>>>>>>>> comments) > >>>>>>>>>>>>>>>>>>>>>>>>>>> The complete staging area is > >>>>> available > >>>>>>> for > >>>>>>>>>> your > >>>>>>>>>>>>>>> review, > >>>>>>>>>>>>>>>>>> which > >>>>>>>>>>>>>>>>>>>>>>> includes: > >>>>>>>>>>>>>>>>>>>>>>>>>>> * JIRA release notes [1], > >>>>>>>>>>>>>>>>>>>>>>>>>>> * the official Apache source release > >>>>>> and > >>>>>>>>>> binary > >>>>>>>>>>>>>>>>> convenience > >>>>>>>>>>>>>>>>>>>>> releases > >>>>>>>>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>>>>>>>>>> be > >>>>>>>>>>>>>>>>>>>>>>>>>>> deployed to dist.apache.org [2], > >>>>>> which > >>>>>>>> are > >>>>>>>>>>> signed > >>>>>>>>>>>>>>> with > >>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>> key > >>>>>>>>>>>>>>>>>>>>> with > >>>>>>>>>>>>>>>>>>>>>>>>>>> fingerprint > >>>>>>>>>>>>>> 1C1E2394D3194E1944613488F320986D35C33D6A > >>>>>>>>>>>>>>>> [3], > >>>>>>>>>>>>>>>>>>>>>>>>>>> * all artifacts to be deployed to > >>>>> the > >>>>>>>> Maven > >>>>>>>>>>>>>> Central > >>>>>>>>>>>>>>>>>> Repository > >>>>>>>>>>>>>>>>>>>>> [4], > >>>>>>>>>>>>>>>>>>>>>>>>>>> * source code tag > >>>>> “release-1.9.0-rc2” > >>>>>>> [5]. > >>>>>>>>>>>>>>>>>>>>>>>>>>> Robert is also preparing a pull > >>>>>> request > >>>>>>>> for > >>>>>>>>>> the > >>>>>>>>>>>>>>>>> announcement > >>>>>>>>>>>>>>>>>>>> blog > >>>>>>>>>>>>>>>>>>>>>> post > >>>>>>>>>>>>>>>>>>>>>>>>>> in > >>>>>>>>>>>>>>>>>>>>>>>>>>> the works, and will update this > >>>>> voting > >>>>>>>>> thread > >>>>>>>>>>>>>> with a > >>>>>>>>>>>>>>>> link > >>>>>>>>>>>>>>>>> to > >>>>>>>>>>>>>>>>>>> the > >>>>>>>>>>>>>>>>>>>>>> pull > >>>>>>>>>>>>>>>>>>>>>>>>>>> request shortly afterwards. > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >>>>>>>>>>>>>>>>>>>>>>>>>>> The vote will be open for *at least > >>>>> 72 > >>>>>>>>> hours*. > >>>>>>>>>>>>>>>>>>>>>>>>>>> Please cast your votes before *Aug. > >>>>>> 14th > >>>>>>>>>> (Wed.) > >>>>>>>>>>>>>> 2019, > >>>>>>>>>>>>>>>>> 17:00 > >>>>>>>>>>>>>>>>>> PM > >>>>>>>>>>>>>>>>>>>>>> CET*.It > >>>>>>>>>>>>>>>>>>>>>>>>>> is > >>>>>>>>>>>>>>>>>>>>>>>>>>> adopted by majority approval, with > >>>>> at > >>>>>>>> least > >>>>>>>>> 3 > >>>>>>>>>>> PMC > >>>>>>>>>>>>>>>>>> affirmative > >>>>>>>>>>>>>>>>>>>>> votes. > >>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks, > >>>>>>>>>>>>>>>>>>>>>>>>>>> Gordon[1] > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >> > https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601 > >>>>>>>>>>>>>>>>>>>>>>>>>>> [2] > >>>>>> https://dist.apache.org/repos/dist/dev/flink/flink-1.9.0-rc2/ > >>>>>>>>>>>>>>>>>>>>>>>>>>> [3] > >>>>>>>>>>>>>>>> https://dist.apache.org/repos/dist/release/flink/KEYS > >>>>>>>>>>>>>>>>>>>>>>>>>>> [4] > >>> https://repository.apache.org/content/repositories/orgapacheflink-1234 > >>>>>>>>>>>>>>>>>>>>>>>>>>> [5] > >>>>>>>>>>>>>>>>>>>>>>>>>>> > >> > https://gitbox.apache.org/repos/asf?p=flink.git;a=tag;h=refs/tags/release-1.9.0-rc2 > >>>>>>>>>>>>>>>>>>>>>>> > >>> > >