After going through the licenses, I found 2 suspicions but not sure if they are valid or not.
1. flink-state-processing-api is packaged in to flink-dist jar, but not included in NOTICE-binary file (the one under the root directory) like other modules. 2. flink-runtime-web distributed some JavaScript dependencies through source codes, the licenses and NOTICE file were only updated inside the module of flink-runtime-web, but not the NOTICE file and licenses directory which under the root directory. Another minor issue I just found is: FLINK-13558 tries to include table examples to flink-dist, but I cannot find it in the binary distribution of RC2. Best, Kurt On Thu, Aug 15, 2019 at 6:19 PM Kurt Young <ykt...@gmail.com> wrote: > Hi Gordon & Timo, > > Thanks for the feedback, and I agree with it. I will document this in the > release notes. > > Best, > Kurt > > > On Thu, Aug 15, 2019 at 6:14 PM Tzu-Li (Gordon) Tai <tzuli...@apache.org> > wrote: > >> Hi Kurt, >> >> With the same argument as before, given that it is mentioned in the >> release >> announcement that it is a preview feature, I would not block this release >> because of it. >> Nevertheless, it would be important to mention this explicitly in the >> release notes [1]. >> >> Regards, >> Gordon >> >> [1] https://github.com/apache/flink/pull/9438 >> >> On Thu, Aug 15, 2019 at 11:29 AM Timo Walther <twal...@apache.org> wrote: >> >> > Hi Kurt, >> > >> > I agree that this is a serious bug. However, I would not block the >> > release because of this. As you said, there is a workaround and the >> > `execute()` works in the most common case of a single execution. We can >> > fix this in a minor release shortly after. >> > >> > What do others think? >> > >> > Regards, >> > Timo >> > >> > >> > Am 15.08.19 um 11:23 schrieb Kurt Young: >> > > HI, >> > > >> > > We just find a serious bug around blink planner: >> > > https://issues.apache.org/jira/browse/FLINK-13708 >> > > When user reused the table environment instance, and call `execute` >> > method >> > > multiple times for >> > > different sql, the later call will trigger the earlier ones to be >> > > re-executed. >> > > >> > > It's a serious bug but seems we also have a work around, which is >> never >> > > reuse the table environment >> > > object. I'm not sure if we should treat this one as blocker issue of >> > 1.9.0. >> > > >> > > What's your opinion? >> > > >> > > Best, >> > > Kurt >> > > >> > > >> > > On Thu, Aug 15, 2019 at 2:01 PM Gary Yao <g...@ververica.com> wrote: >> > > >> > >> +1 (non-binding) >> > >> >> > >> Jepsen test suite passed 10 times consecutively >> > >> >> > >> On Wed, Aug 14, 2019 at 5:31 PM Aljoscha Krettek < >> aljos...@apache.org> >> > >> wrote: >> > >> >> > >>> +1 >> > >>> >> > >>> I did some testing on a Google Cloud Dataproc cluster (it gives you >> a >> > >>> managed YARN and Google Cloud Storage (GCS)): >> > >>> - tried both YARN session mode and YARN per-job mode, also using >> > >>> bin/flink list/cancel/etc. against a YARN session cluster >> > >>> - ran examples that write to GCS, both with the native Hadoop >> > >> FileSystem >> > >>> and a custom “plugin” FileSystem >> > >>> - ran stateful streaming jobs that use GCS as a checkpoint >> backend >> > >>> - tried running SQL programs on YARN using the SQL Cli: this >> worked >> > for >> > >>> YARN session mode but not for YARN per-job mode. Looking at the >> code I >> > >>> don’t think per-job mode would work from seeing how it is >> implemented. >> > >> But >> > >>> I think it’s an OK restriction to have for now >> > >>> - in all the testing I had fine-grained recovery (region >> failover) >> > >>> enabled but I didn’t simulate any failures >> > >>> >> > >>>> On 14. Aug 2019, at 15:20, Kurt Young <ykt...@gmail.com> wrote: >> > >>>> >> > >>>> Hi, >> > >>>> >> > >>>> Thanks for preparing this release candidate. I have verified the >> > >>> following: >> > >>>> - verified the checksums and GPG files match the corresponding >> release >> > >>> files >> > >>>> - verified that the source archives do not contains any binaries >> > >>>> - build the source release with Scala 2.11 successfully. >> > >>>> - ran `mvn verify` locally, met 2 issuses [FLINK-13687] and >> > >>> [FLINK-13688], >> > >>>> but >> > >>>> both are not release blockers. Other than that, all tests are >> passed. >> > >>>> - ran all e2e tests which don't need download external packages >> (it's >> > >>> very >> > >>>> unstable >> > >>>> in China and almost impossible to download them), all passed. >> > >>>> - started local cluster, ran some examples. Met a small website >> > display >> > >>>> issue >> > >>>> [FLINK-13591], which is also not a release blocker. >> > >>>> >> > >>>> Although we have pushed some fixes around blink planner and hive >> > >>>> integration >> > >>>> after RC2, but consider these are both preview features, I'm lean >> to >> > be >> > >>> ok >> > >>>> to release >> > >>>> without these fixes. >> > >>>> >> > >>>> +1 from my side. (binding) >> > >>>> >> > >>>> Best, >> > >>>> Kurt >> > >>>> >> > >>>> >> > >>>> On Wed, Aug 14, 2019 at 5:13 PM Jark Wu <imj...@gmail.com> wrote: >> > >>>> >> > >>>>> Hi Gordon, >> > >>>>> >> > >>>>> I have verified the following things: >> > >>>>> >> > >>>>> - build the source release with Scala 2.12 and Scala 2.11 >> > successfully >> > >>>>> - checked/verified signatures and hashes >> > >>>>> - checked that all POM files point to the same version >> > >>>>> - ran some flink table related end-to-end tests locally and >> succeeded >> > >>>>> (except TPC-H e2e failed which is reported in FLINK-13704) >> > >>>>> - started cluster for both Scala 2.11 and 2.12, ran examples, >> > verified >> > >>> web >> > >>>>> ui and log output, nothing unexpected >> > >>>>> - started cluster, ran a SQL query to temporal join with kafka >> source >> > >>> and >> > >>>>> mysql jdbc table, and write results to kafka again. Using DDL to >> > >> create >> > >>> the >> > >>>>> source and sinks. looks good. >> > >>>>> - reviewed the release PR >> > >>>>> >> > >>>>> As FLINK-13704 is not recognized as blocker issue, so +1 from my >> side >> > >>>>> (non-binding). >> > >>>>> >> > >>>>> On Tue, 13 Aug 2019 at 17:07, Till Rohrmann <trohrm...@apache.org >> > >> > >>> wrote: >> > >>>>>> Hi Richard, >> > >>>>>> >> > >>>>>> although I can see that it would be handy for users who have >> PubSub >> > >> set >> > >>>>> up, >> > >>>>>> I would rather not include examples which require an external >> > >>> dependency >> > >>>>>> into the Flink distribution. I think examples should be >> > >> self-contained. >> > >>>>> My >> > >>>>>> concern is that we would bloat the distribution for many users at >> > the >> > >>>>>> benefit of a few. Instead, I think it would be better to make >> these >> > >>>>>> examples available differently, maybe through Flink's ecosystem >> > >> website >> > >>>>> or >> > >>>>>> maybe a new examples section in Flink's documentation. >> > >>>>>> >> > >>>>>> Cheers, >> > >>>>>> Till >> > >>>>>> >> > >>>>>> On Tue, Aug 13, 2019 at 9:43 AM Jark Wu <imj...@gmail.com> >> wrote: >> > >>>>>> >> > >>>>>>> Hi Till, >> > >>>>>>> >> > >>>>>>> After thinking about we can use VARCHAR as an alternative of >> > >>>>>>> timestamp/time/date. >> > >>>>>>> I'm fine with not recognize it as a blocker issue. >> > >>>>>>> We can fix it into 1.9.1. >> > >>>>>>> >> > >>>>>>> >> > >>>>>>> Thanks, >> > >>>>>>> Jark >> > >>>>>>> >> > >>>>>>> >> > >>>>>>> On Tue, 13 Aug 2019 at 15:10, Richard Deurwaarder < >> rich...@xeli.eu >> > > >> > >>>>>> wrote: >> > >>>>>>>> Hello all, >> > >>>>>>>> >> > >>>>>>>> I noticed the PubSub example jar is not included in the >> examples/ >> > >> dir >> > >>>>>> of >> > >>>>>>>> flink-dist. I've created >> > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13700 >> > >>>>>>>> + https://github.com/apache/flink/pull/9424/files to fix this. >> > >>>>>>>> >> > >>>>>>>> I will leave it up to you to decide if we want to add this to >> > >> 1.9.0. >> > >>>>>>>> Regards, >> > >>>>>>>> >> > >>>>>>>> Richard >> > >>>>>>>> >> > >>>>>>>> On Tue, Aug 13, 2019 at 9:04 AM Till Rohrmann < >> > >> trohrm...@apache.org> >> > >>>>>>>> wrote: >> > >>>>>>>> >> > >>>>>>>>> Hi Jark, >> > >>>>>>>>> >> > >>>>>>>>> thanks for reporting this issue. Could this be a documented >> > >>>>>> limitation >> > >>>>>>> of >> > >>>>>>>>> Blink's preview version? I think we have agreed that the Blink >> > SQL >> > >>>>>>>> planner >> > >>>>>>>>> will be rather a preview feature than production ready. Hence >> it >> > >>>>>> could >> > >>>>>>>>> still contain some bugs. My concern is that there might be >> still >> > >>>>>> other >> > >>>>>>>>> issues which we'll discover bit by bit and could postpone the >> > >>>>> release >> > >>>>>>>> even >> > >>>>>>>>> further if we say Blink bugs are blockers. >> > >>>>>>>>> >> > >>>>>>>>> Cheers, >> > >>>>>>>>> Till >> > >>>>>>>>> >> > >>>>>>>>> On Tue, Aug 13, 2019 at 7:42 AM Jark Wu <imj...@gmail.com> >> > wrote: >> > >>>>>>>>> >> > >>>>>>>>>> Hi all, >> > >>>>>>>>>> >> > >>>>>>>>>> I just find an issue when testing connector DDLs against >> blink >> > >>>>>>> planner >> > >>>>>>>>> for >> > >>>>>>>>>> rc2. >> > >>>>>>>>>> This issue lead to the DDL doesn't work when containing >> > >>>>>>>>> timestamp/date/time >> > >>>>>>>>>> type. >> > >>>>>>>>>> I have created an issue FLINK-13699[1] and a pull request for >> > >>>>> this. >> > >>>>>>>>>> IMO, this can be a blocker issue of 1.9 release. Because >> > >>>>>>>>>> timestamp/date/time are primitive types, and this will break >> the >> > >>>>>> DDL >> > >>>>>>>>>> feature. >> > >>>>>>>>>> However, I want to hear more thoughts from the community >> whether >> > >>>>> we >> > >>>>>>>>> should >> > >>>>>>>>>> recognize it as a blocker. >> > >>>>>>>>>> >> > >>>>>>>>>> Thanks, >> > >>>>>>>>>> Jark >> > >>>>>>>>>> >> > >>>>>>>>>> >> > >>>>>>>>>> [1]: https://issues.apache.org/jira/browse/FLINK-13699 >> > >>>>>>>>>> >> > >>>>>>>>>> >> > >>>>>>>>>> >> > >>>>>>>>>> On Mon, 12 Aug 2019 at 22:46, Becket Qin < >> becket....@gmail.com> >> > >>>>>>> wrote: >> > >>>>>>>>>>> Thanks Gordon, will do that. >> > >>>>>>>>>>> >> > >>>>>>>>>>> On Mon, Aug 12, 2019 at 4:42 PM Tzu-Li (Gordon) Tai < >> > >>>>>>>>> tzuli...@apache.org >> > >>>>>>>>>>> wrote: >> > >>>>>>>>>>> >> > >>>>>>>>>>>> Concerning FLINK-13231: >> > >>>>>>>>>>>> >> > >>>>>>>>>>>> Since this is a @PublicEvolving interface, technically it >> is >> > >>>>> ok >> > >>>>>>> to >> > >>>>>>>>>> break >> > >>>>>>>>>>>> it across releases (including across bugfix releases?). >> > >>>>>>>>>>>> So, @Becket if you do merge it now, please mark the fix >> > >>>>> version >> > >>>>>>> as >> > >>>>>>>>>> 1.9.1. >> > >>>>>>>>>>>> During the voting process, in the case a new RC is created, >> > >>>>> we >> > >>>>>>>>> usually >> > >>>>>>>>>>>> check the list of changes compared to the previous RC, and >> > >>>>>>> correct >> > >>>>>>>>> the >> > >>>>>>>>>>> "Fix >> > >>>>>>>>>>>> Version" of the corresponding JIRAs to be the right version >> > >>>>> (in >> > >>>>>>> the >> > >>>>>>>>>> case, >> > >>>>>>>>>>>> it would be corrected to 1.9.0 instead of 1.9.1). >> > >>>>>>>>>>>> >> > >>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:25 PM Till Rohrmann < >> > >>>>>>>> trohrm...@apache.org> >> > >>>>>>>>>>>> wrote: >> > >>>>>>>>>>>> >> > >>>>>>>>>>>>> I agree that it would be nicer. Not sure whether we should >> > >>>>>>> cancel >> > >>>>>>>>> the >> > >>>>>>>>>> RC >> > >>>>>>>>>>>>> for this issue given that it is open for quite some time >> and >> > >>>>>>>> hasn't >> > >>>>>>>>>> been >> > >>>>>>>>>>>>> addressed until very recently. Maybe we could include it >> on >> > >>>>>> the >> > >>>>>>>>>>> shortlist >> > >>>>>>>>>>>>> of nice-to-do things which we do in case that the RC gets >> > >>>>>>>> cancelled. >> > >>>>>>>>>>>>> Cheers, >> > >>>>>>>>>>>>> Till >> > >>>>>>>>>>>>> >> > >>>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:18 PM Becket Qin < >> > >>>>>>> becket....@gmail.com> >> > >>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>> Hi Till, >> > >>>>>>>>>>>>>> >> > >>>>>>>>>>>>>> Yes, I think we have already documented in that way. So >> > >>>>>>>> technically >> > >>>>>>>>>>>>>> speaking it is fine to change it later. It is just better >> > >>>>> if >> > >>>>>> we >> > >>>>>>>>> could >> > >>>>>>>>>>>>>> avoid >> > >>>>>>>>>>>>>> doing that. >> > >>>>>>>>>>>>>> >> > >>>>>>>>>>>>>> Thanks, >> > >>>>>>>>>>>>>> >> > >>>>>>>>>>>>>> Jiangjie (Becket) Qin >> > >>>>>>>>>>>>>> >> > >>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 4:09 PM Till Rohrmann < >> > >>>>>>>>> trohrm...@apache.org> >> > >>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>> Could we say that the PubSub connector is public >> evolving >> > >>>>>>>>> instead? >> > >>>>>>>>>>>>>>> Cheers, >> > >>>>>>>>>>>>>>> Till >> > >>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 3:18 PM Becket Qin < >> > >>>>>>>> becket....@gmail.com >> > >>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>> Hi all, >> > >>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>> FLINK-13231(palindrome!) has a minor Google PubSub >> > >>>>>>> connector >> > >>>>>>>>> API >> > >>>>>>>>>>>>>> change >> > >>>>>>>>>>>>>>>> regarding how to config rate limiting. The GCP PubSub >> > >>>>>>>> connector >> > >>>>>>>>>> is >> > >>>>>>>>>>> a >> > >>>>>>>>>>>>>>> newly >> > >>>>>>>>>>>>>>>> introduced connector in 1.9, so it would be nice to >> > >>>>>> include >> > >>>>>>>>> this >> > >>>>>>>>>>>>>> change >> > >>>>>>>>>>>>>>>> into 1.9 rather than later to avoid a public API >> > >>>>> change. >> > >>>>>> I >> > >>>>>>> am >> > >>>>>>>>>>>>>> thinking of >> > >>>>>>>>>>>>>>>> making this as a blocker for 1.9. Want to check what do >> > >>>>>>>> others >> > >>>>>>>>>>> think. >> > >>>>>>>>>>>>>>>> Thanks, >> > >>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>> Jiangjie (Becket) Qin >> > >>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 2:04 PM Zili Chen < >> > >>>>>>>>> wander4...@gmail.com> >> > >>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>> Hi Kurt, >> > >>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>> Thanks for your explanation. For [1] I think at least >> > >>>>>> we >> > >>>>>>>>> should >> > >>>>>>>>>>>>>> change >> > >>>>>>>>>>>>>>>>> the JIRA issue field, like unset the fixed version. >> > >>>>> For >> > >>>>>>>> [2] I >> > >>>>>>>>>> can >> > >>>>>>>>>>>>>> see >> > >>>>>>>>>>>>>>>>> the change is all in test scope but wonder if such a >> > >>>>>>> commit >> > >>>>>>>>>> still >> > >>>>>>>>>>>>>>> invalid >> > >>>>>>>>>>>>>>>>> the release candidate. IIRC previous RC VOTE threads >> > >>>>>>> would >> > >>>>>>>>>>> contain >> > >>>>>>>>>>>>>> a >> > >>>>>>>>>>>>>>>>> release manual/guide, I will try to look up it, too. >> > >>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>> Best, >> > >>>>>>>>>>>>>>>>> tison. >> > >>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>> Kurt Young <ykt...@gmail.com> 于2019年8月12日周一 >> > >>>>> 下午5:42写道: >> > >>>>>>>>>>>>>>>>>> Hi Zili, >> > >>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>> Thanks for the heads up. The 2 issues you mentioned >> > >>>>>>> were >> > >>>>>>>>>> opened >> > >>>>>>>>>>>>>> by >> > >>>>>>>>>>>>>>> me. >> > >>>>>>>>>>>>>>>> We >> > >>>>>>>>>>>>>>>>>> have >> > >>>>>>>>>>>>>>>>>> found the reason of the second issue and a PR was >> > >>>>>>> opened >> > >>>>>>>>> for >> > >>>>>>>>>>> it. >> > >>>>>>>>>>>>>> As >> > >>>>>>>>>>>>>>>> said >> > >>>>>>>>>>>>>>>>> in >> > >>>>>>>>>>>>>>>>>> jira, the >> > >>>>>>>>>>>>>>>>>> issue was just a testing problem, should not be >> > >>>>>> blocker >> > >>>>>>>> of >> > >>>>>>>>>>> 1.9.0 >> > >>>>>>>>>>>>>>>> release. >> > >>>>>>>>>>>>>>>>>> However, >> > >>>>>>>>>>>>>>>>>> we will still merge it into 1.9 branch. >> > >>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>> Best, >> > >>>>>>>>>>>>>>>>>> Kurt >> > >>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 5:38 PM Zili Chen < >> > >>>>>>>>>>> wander4...@gmail.com> >> > >>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>> Hi, >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> I just noticed that a few hours ago there were >> > >>>>> two >> > >>>>>>> new >> > >>>>>>>>>> issues >> > >>>>>>>>>>>>>>>>>>> filed and marked as blockers to 1.9.0[1][2]. >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> Now [1] is closed as duplication but still marked >> > >>>>>> as >> > >>>>>>>>>>>>>>>>>>> a blocker to 1.9.0, while [2] is downgrade to >> > >>>>>> "Major" >> > >>>>>>>>>>> priority >> > >>>>>>>>>>>>>>>>>>> but still target to be fixed in 1.9.0. >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> It would be worth to have attention of our >> > >>>>> release >> > >>>>>>>>> manager >> > >>>>>>>>>> at >> > >>>>>>>>>>>>>>> least. >> > >>>>>>>>>>>>>>>>>>> Best, >> > >>>>>>>>>>>>>>>>>>> tison. >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> [1] >> > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13687 >> > >>>>>>>>>>>>>>>>>>> [2] >> > >>>>>>> https://issues.apache.org/jira/browse/FLINK-13688 >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>> Gyula Fóra <gyula.f...@gmail.com> 于2019年8月12日周一 >> > >>>>>>>>> 下午5:10写道: >> > >>>>>>>>>>>>>>>>>>>> Thanks Stephan :) >> > >>>>>>>>>>>>>>>>>>>> That looks easy enough, will try! >> > >>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>> Gyula >> > >>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 11:00 AM Stephan Ewen < >> > >>>>>>>>>>>>>> se...@apache.org> >> > >>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>> Hi Gyula! >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> Thanks for reporting this. >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> Can you try to simply build Flink without >> > >>>>>> Hadoop >> > >>>>>>>> and >> > >>>>>>>>>> then >> > >>>>>>>>>>>>>>>> exporting >> > >>>>>>>>>>>>>>>>>>>>> HADOOP_CLASSPATH to your CloudEra libs? >> > >>>>>>>>>>>>>>>>>>>>> That is the recommended way these days. >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> Best, >> > >>>>>>>>>>>>>>>>>>>>> Stephan >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 10:48 AM Gyula Fóra < >> > >>>>>>>>>>>>>>>> gyula.f...@gmail.com> >> > >>>>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>>> Thanks Dawid, >> > >>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>> In the meantime I also figured out that I >> > >>>>>> need >> > >>>>>>> to >> > >>>>>>>>>> build >> > >>>>>>>>>>>>>> the >> > >>>>>>>>>>>>>>>>>>>>>> https://github.com/apache/flink-shaded >> > >>>>>> project >> > >>>>>>>>>> locally >> > >>>>>>>>>>>>>> with >> > >>>>>>>>>>>>>>>>>>>>>> -Dhadoop.version set to the specific hadoop >> > >>>>>>>> version >> > >>>>>>>>>> if >> > >>>>>>>>>>> I >> > >>>>>>>>>>>>>> want >> > >>>>>>>>>>>>>>>>>>> something >> > >>>>>>>>>>>>>>>>>>>>>> different. >> > >>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>> Cheers, >> > >>>>>>>>>>>>>>>>>>>>>> Gyula >> > >>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>> On Mon, Aug 12, 2019 at 9:54 AM Dawid >> > >>>>>>> Wysakowicz >> > >>>>>>>> < >> > >>>>>>>>>>>>>>>>>>>> dwysakow...@apache.org >> > >>>>>>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> Hi Gyula, >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> As for the issues with mapr maven >> > >>>>>> repository, >> > >>>>>>>> you >> > >>>>>>>>>>> might >> > >>>>>>>>>>>>>>> have >> > >>>>>>>>>>>>>>>> a >> > >>>>>>>>>>>>>>>>>> look >> > >>>>>>>>>>>>>>>>>>>> at >> > >>>>>>>>>>>>>>>>>>>>>>> this message: >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >> >> > >> https://lists.apache.org/thread.html/77f4db930216e6da0d6121065149cef43ff3ea33c9ffe9b1a3047210@%3Cdev.flink.apache.org%3E >> > >>>>>>>>>>>>>>>>>>>>>>> Try using the "unsafe-mapr-repo" profile. >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> Best, >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> Dawid >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>> On 11/08/2019 19:31, Gyula Fóra wrote: >> > >>>>>>>>>>>>>>>>>>>>>>>> Hi again, >> > >>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>> How do I build the RC locally with the >> > >>>>>>> hadoop >> > >>>>>>>>>>> version >> > >>>>>>>>>>>>>>>>>> specified? >> > >>>>>>>>>>>>>>>>>>>>> Seems >> > >>>>>>>>>>>>>>>>>>>>>>> like >> > >>>>>>>>>>>>>>>>>>>>>>>> no matter what I do I run into >> > >>>>> dependency >> > >>>>>>>>>> problems >> > >>>>>>>>>>>>>> with >> > >>>>>>>>>>>>>>> the >> > >>>>>>>>>>>>>>>>>>> shaded >> > >>>>>>>>>>>>>>>>>>>>>> hadoop >> > >>>>>>>>>>>>>>>>>>>>>>>> dependencies. >> > >>>>>>>>>>>>>>>>>>>>>>>> This seems to have worked in the past. >> > >>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>> There might be some documentation >> > >>>>>> somewhere >> > >>>>>>>>> that >> > >>>>>>>>>> I >> > >>>>>>>>>>>>>>> couldnt >> > >>>>>>>>>>>>>>>>>> find, >> > >>>>>>>>>>>>>>>>>>>> so I >> > >>>>>>>>>>>>>>>>>>>>>>> would >> > >>>>>>>>>>>>>>>>>>>>>>>> appreciate any pointers :) >> > >>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>> Thanks! >> > >>>>>>>>>>>>>>>>>>>>>>>> Gyula >> > >>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>> On Sun, Aug 11, 2019 at 6:57 PM Gyula >> > >>>>>> Fóra >> > >>>>>>> < >> > >>>>>>>>>>>>>>>>>> gyula.f...@gmail.com >> > >>>>>>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>>>>>> Hi! >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>> I am trying to build 1.9.0-rc2 with >> > >>>>> the >> > >>>>>>>>>>>>>> -Pvendor-repos >> > >>>>>>>>>>>>>>>>> profile >> > >>>>>>>>>>>>>>>>>>>>>> enabled. >> > >>>>>>>>>>>>>>>>>>>>>>> I >> > >>>>>>>>>>>>>>>>>>>>>>>>> get the following error: >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>> mvn clean install -DskipTests >> > >>>>>>> -Pvendor-repos >> > >>>>>>>>>>>>>>>>>>>> -Dhadoop.version=2.6.0 >> > >>>>>>>>>>>>>>>>>>>>>>>>> -Pinclude-hadoop (ignore that the >> > >>>>> hadoop >> > >>>>>>>>> version >> > >>>>>>>>>>> is >> > >>>>>>>>>>>>>> not >> > >>>>>>>>>>>>>>> a >> > >>>>>>>>>>>>>>>>>> vendor >> > >>>>>>>>>>>>>>>>>>>>>> hadoop >> > >>>>>>>>>>>>>>>>>>>>>>>>> version) >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>> [ERROR] Failed to execute goal on >> > >>>>>> project >> > >>>>>>>>>>>>>>> flink-hadoop-fs: >> > >>>>>>>>>>>>>>>>>> Could >> > >>>>>>>>>>>>>>>>>>>> not >> > >>>>>>>>>>>>>>>>>>>>>>>>> resolve dependencies for project >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>> org.apache.flink:flink-hadoop-fs:jar:1.9.0: >> > >>>>>>>>>> Failed >> > >>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>> collect >> > >>>>>>>>>>>>>>>>>>>>>>> dependencies >> > >>>>>>>>>>>>>>>>>>>>>>>>> at >> > >>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0: >> > >>>>>>>>>>>>>>>>>> Failed >> > >>>>>>>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>>>>>>> read >> > >>>>>>>>>>>>>>>>>>>>>>>>> artifact descriptor for >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:jar:2.6.0-7.0: >> > >>>>>>>>>>>>>>>> Could >> > >>>>>>>>>>>>>>>>>> not >> > >>>>>>>>>>>>>>>>>>>>>> transfer >> > >>>>>>>>>>>>>>>>>>>>>>>>> artifact >> > >>>>>>>>>>>>>>>>> org.apache.flink:flink-shaded-hadoop-2:pom:2.6.0-7.0 >> > >>>>>>>>>>>>>>>>>>>>> from/to >> > >>>>>>>>>>>>>>>>>>>>>>>>> mapr-releases ( >> > >>>>>>>>>> https://repository.mapr.com/maven/ >> > >>>>>>>>>>> ): >> > >>>>>> sun.security.validator.ValidatorException: >> > >>>>>>>>> PKIX >> > >>>>>>>>>>> path >> > >>>>>>>>>>>>>>>>> building >> > >>>>>>>>>>>>>>>>>>>>> failed: >> > >>>>>> sun.security.provider.certpath.SunCertPathBuilderException: >> > >>>>>>>>>>>>>>>>>>> unable >> > >>>>>>>>>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>>>>>>>> find >> > >>>>>>>>>>>>>>>>>>>>>>>>> valid certification path to requested >> > >>>>>>> target >> > >>>>>>>>> -> >> > >>>>>>>>>>>>>> [Help 1] >> > >>>>>>>>>>>>>>>>>>>>>>>>> This looks like a TLS error. Might not >> > >>>>>> be >> > >>>>>>>>>> related >> > >>>>>>>>>>>>>> to the >> > >>>>>>>>>>>>>>>>>> release >> > >>>>>>>>>>>>>>>>>>>> but >> > >>>>>>>>>>>>>>>>>>>>>> it >> > >>>>>>>>>>>>>>>>>>>>>>>>> could be good to know. >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>> Cheers, >> > >>>>>>>>>>>>>>>>>>>>>>>>> Gyula >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Aug 9, 2019 at 6:26 PM Tzu-Li >> > >>>>>>>> (Gordon) >> > >>>>>>>>>>> Tai < >> > >>>>>>>>>>>>>>>>>>>>>> tzuli...@apache.org >> > >>>>>>>>>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>> Please note that the unresolved >> > >>>>> issues >> > >>>>>>> that >> > >>>>>>>>> are >> > >>>>>>>>>>>>>> still >> > >>>>>>>>>>>>>>>>> tagged >> > >>>>>>>>>>>>>>>>>>>> with a >> > >>>>>>>>>>>>>>>>>>>>>> fix >> > >>>>>>>>>>>>>>>>>>>>>>>>>> version "1.9.0", as seen in the JIRA >> > >>>>>>>> release >> > >>>>>>>>>>> notes >> > >>>>>>>>>>>>>> [1], >> > >>>>>>>>>>>>>>>> are >> > >>>>>>>>>>>>>>>>>>>> issues >> > >>>>>>>>>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>>>>>>>>>>> update documents for new features. >> > >>>>>>>>>>>>>>>>>>>>>>>>>> I've left them still associated with >> > >>>>>>> 1.9.0 >> > >>>>>>>>>> since >> > >>>>>>>>>>>>>> these >> > >>>>>>>>>>>>>>>>> should >> > >>>>>>>>>>>>>>>>>>>> still >> > >>>>>>>>>>>>>>>>>>>>>> be >> > >>>>>>>>>>>>>>>>>>>>>>>>>> updated for 1.9.0 soon along with the >> > >>>>>>>>> official >> > >>>>>>>>>>>>>> release. >> > >>>>>>>>>>>>>>>>>>>>>>>>>> [1] >> > >>>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>> >> > >> >> > >> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601 >> > >>>>>>>>>>>>>>>>>>>>>>>>>> On Fri, Aug 9, 2019 at 6:17 PM Tzu-Li >> > >>>>>>>>> (Gordon) >> > >>>>>>>>>>> Tai >> > >>>>>>>>>>>>>> < >> > >>>>>>>>>>>>>>>>>>>>>>> tzuli...@apache.org> >> > >>>>>>>>>>>>>>>>>>>>>>>>>> wrote: >> > >>>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Hi all, >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Release candidate #2 for Apache >> > >>>>> Flink >> > >>>>>>>> 1.9.0 >> > >>>>>>>>> is >> > >>>>>>>>>>> now >> > >>>>>>>>>>>>>>> ready >> > >>>>>>>>>>>>>>>>> for >> > >>>>>>>>>>>>>>>>>>>> your >> > >>>>>>>>>>>>>>>>>>>>>>>>>> review. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> This is the first voting candidate >> > >>>>> for >> > >>>>>>>>> 1.9.0, >> > >>>>>>>>>>>>>>> following >> > >>>>>>>>>>>>>>>>> the >> > >>>>>>>>>>>>>>>>>>>>> preview >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> candidates RC0 and RC1. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Please review and vote on release >> > >>>>>>>> candidate >> > >>>>>>>>> #2 >> > >>>>>>>>>>> for >> > >>>>>>>>>>>>>>>> version >> > >>>>>>>>>>>>>>>>>>>> 1.9.0, >> > >>>>>>>>>>>>>>>>>>>>> as >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> follows: >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [ ] +1, Approve the release >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [ ] -1, Do not approve the release >> > >>>>>>> (please >> > >>>>>>>>>>> provide >> > >>>>>>>>>>>>>>>>> specific >> > >>>>>>>>>>>>>>>>>>>>>> comments) >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> The complete staging area is >> > >>>>> available >> > >>>>>>> for >> > >>>>>>>>>> your >> > >>>>>>>>>>>>>>> review, >> > >>>>>>>>>>>>>>>>>> which >> > >>>>>>>>>>>>>>>>>>>>>>> includes: >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> * JIRA release notes [1], >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> * the official Apache source release >> > >>>>>> and >> > >>>>>>>>>> binary >> > >>>>>>>>>>>>>>>>> convenience >> > >>>>>>>>>>>>>>>>>>>>> releases >> > >>>>>>>>>>>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>>>>>>>>>>> be >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> deployed to dist.apache.org [2], >> > >>>>>> which >> > >>>>>>>> are >> > >>>>>>>>>>> signed >> > >>>>>>>>>>>>>>> with >> > >>>>>>>>>>>>>>>>> the >> > >>>>>>>>>>>>>>>>>>> key >> > >>>>>>>>>>>>>>>>>>>>> with >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> fingerprint >> > >>>>>>>>>>>>>> 1C1E2394D3194E1944613488F320986D35C33D6A >> > >>>>>>>>>>>>>>>> [3], >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> * all artifacts to be deployed to >> > >>>>> the >> > >>>>>>>> Maven >> > >>>>>>>>>>>>>> Central >> > >>>>>>>>>>>>>>>>>> Repository >> > >>>>>>>>>>>>>>>>>>>>> [4], >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> * source code tag >> > >>>>> “release-1.9.0-rc2” >> > >>>>>>> [5]. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Robert is also preparing a pull >> > >>>>>> request >> > >>>>>>>> for >> > >>>>>>>>>> the >> > >>>>>>>>>>>>>>>>> announcement >> > >>>>>>>>>>>>>>>>>>>> blog >> > >>>>>>>>>>>>>>>>>>>>>> post >> > >>>>>>>>>>>>>>>>>>>>>>>>>> in >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> the works, and will update this >> > >>>>> voting >> > >>>>>>>>> thread >> > >>>>>>>>>>>>>> with a >> > >>>>>>>>>>>>>>>> link >> > >>>>>>>>>>>>>>>>> to >> > >>>>>>>>>>>>>>>>>>> the >> > >>>>>>>>>>>>>>>>>>>>>> pull >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> request shortly afterwards. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> The vote will be open for *at least >> > >>>>> 72 >> > >>>>>>>>> hours*. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Please cast your votes before *Aug. >> > >>>>>> 14th >> > >>>>>>>>>> (Wed.) >> > >>>>>>>>>>>>>> 2019, >> > >>>>>>>>>>>>>>>>> 17:00 >> > >>>>>>>>>>>>>>>>>> PM >> > >>>>>>>>>>>>>>>>>>>>>> CET*.It >> > >>>>>>>>>>>>>>>>>>>>>>>>>> is >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> adopted by majority approval, with >> > >>>>> at >> > >>>>>>>> least >> > >>>>>>>>> 3 >> > >>>>>>>>>>> PMC >> > >>>>>>>>>>>>>>>>>> affirmative >> > >>>>>>>>>>>>>>>>>>>>> votes. >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Thanks, >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> Gordon[1] >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> >> > >> >> > >> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12344601 >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [2] >> > >>>>>> https://dist.apache.org/repos/dist/dev/flink/flink-1.9.0-rc2/ >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [3] >> > >>>>>>>>>>>>>>>> https://dist.apache.org/repos/dist/release/flink/KEYS >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [4] >> > >>> >> https://repository.apache.org/content/repositories/orgapacheflink-1234 >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> [5] >> > >>>>>>>>>>>>>>>>>>>>>>>>>>> >> > >> >> > >> https://gitbox.apache.org/repos/asf?p=flink.git;a=tag;h=refs/tags/release-1.9.0-rc2 >> > >>>>>>>>>>>>>>>>>>>>>>> >> > >>> >> > >> > >> >