Thank you all for the votes. Regarding to the license issue reported by Dawid, Chesnay's point sounds reasonable for me. It should not be a blocker issue.
Since the voting time has passed, I will conclude the vote result in a separate thread. > 在 2020年12月7日,下午9:45,Chesnay Schepler <ches...@apache.org> 写道: > > I've filed https://issues.apache.org/jira/browse/FLINK-20519 for the hbase > issue. > > Since we still ship the protobuf license I don't think this is a blocker. > > On 12/7/2020 10:52 AM, Dawid Wysakowicz wrote: >> >> +1 (binding) >> >> * Verified the checksums >> * Verified that the source archives do not contain any binaries >> * Built Flink from sources >> * Run a streaming WorldCount example in BATCH and STREAM mode >> * Run a slightly heavier WorldCount version in BATCH and STREAM mode >> * Verified licensing of Hbase connectors >> *** I found one issue that we do not list protobuf-java-util in >> flink-sql-connector-hbase-2.2 NOTICE file, which is pulled in through >> hbase-shaded-miscellaneous (there is a couple more dependencies we do not >> list from it, but they are Apache licensed)* >> * Run some Table examples (I found two examples are not runnable[1]), but I >> would not consider it a blocker >> >> Let me know what you think about the licensing issue. >> >> Best, >> Dawid >> >> [1]https://issues.apache.org/jira/browse/FLINK-20464 >> On 07/12/2020 08:07, Leonard Xu wrote: >>> +1 (non-binding) >>> >>> - checked/verified signatures and hashes >>> - built from source code with scala 2.11 succeeded >>> - checked that there are no missing artifacts >>> - started a cluster, WebUI was accessible, submitted a wordcount job and >>> ran succeeded, no suspicious log output >>> - tested using SQL Client to submit job and the query result is expected >>> - tested read/write from/to sql kafka/upsert-kafka connector in SQL Client >>> - tested read/write/join hvie table in SQL Client >>> >>> Best, >>> Leonard Xu >>> >>> >>>> 在 2020年12月7日,11:17,Guowei Ma<guowei....@gmail.com> 写道: >>>> >>>> +1(non-binding) >>>> - build from source >>>> - build a docker image >>>> - start a session from local k8s cluster >>>> - submit a wordcount job in streaming mode. >>>> - submit a wordcount job in batch mode. >>>> Best, >>>> Guowei >>>> >>>> >>>> On Sat, Dec 5, 2020 at 3:13 PM Zhu Zhu<reed...@gmail.com> wrote: >>>> >>>>> +1 (binding) >>>>> >>>>> - verified signature and checksum >>>>> - built from source >>>>> - run testing jobs on yarn with manually triggered failures. checked logs >>>>> and WebUI of those jobs >>>>> * DataStream job (paralelism=1000) with multiple disjoint pipelined >>>>> regions >>>>> * DataSet job (paralelism=1000) with all edges blocking >>>>> >>>>> Thanks, >>>>> Zhu >>>>> >>>>> Till Rohrmann<trohrm...@apache.org> 于2020年12月4日周五 下午11:45写道: >>>>> >>>>>> +1 (binding) >>>>>> >>>>>> * Verified the checksums >>>>>> * Ran RC on Minikube cluster >>>>>> ** Session mode >>>>>> ** Application mode >>>>>> * Built Flink from sources >>>>>> >>>>>> Cheers, >>>>>> Till >>>>>> >>>>>> On Fri, Dec 4, 2020 at 2:15 PM Wei Zhong<weizhong0...@gmail.com> wrote: >>>>>> >>>>>>> +1 (non-binding) >>>>>>> >>>>>>> - verified checksums and signatures >>>>>>> - build Flink with Scala 2.11 >>>>>>> - pip install pyflink on Windows python 3.7 >>>>>>> - run a python job with udfs on Windows >>>>>>> - pyflink shell works well on local mode and remote mode >>>>>>> >>>>>>> Best, >>>>>>> Wei >>>>>>> >>>>>>>> 在 2020年12月4日,17:21,Yang Wang<danrtsey...@gmail.com> 写道: >>>>>>>> >>>>>>>> +1 (non-binding) >>>>>>>> >>>>>>>> * Build from source >>>>>>>> * Deploy Flink cluster in following deployments with HA >>>>>> enabled(ZooKeeper >>>>>>>> and K8s), including kill JobManager and check failover >>>>>>>> * Native K8s Session >>>>>>>> * Native K8s Application >>>>>>>> * Yarn Session >>>>>>>> * Yarn Per-Job >>>>>>>> * Yarn Application >>>>>>>> * Check webUI and logs in different deployments especially via >>>>> `kubectl >>>>>>>> logs` in K8s >>>>>>>> >>>>>>>> Best, >>>>>>>> Yang >>>>>>>> >>>>>>>> Xintong Song<tonysong...@gmail.com> 于2020年12月4日周五 下午3:00写道: >>>>>>>> >>>>>>>>> +1 (non-binding) >>>>>>>>> >>>>>>>>> - Verified checksums and signatures >>>>>>>>> - No binaries found in source archive >>>>>>>>> - Build from source >>>>>>>>> - Tried a couple of example jobs in various deployment mode >>>>>>>>> - Local >>>>>>>>> - Standalone >>>>>>>>> - Native Kubernetes Application >>>>>>>>> - Native Kubernetes Session >>>>>>>>> - Yarn Job >>>>>>>>> - Yarn Session >>>>>>>>> - Changing memory configurations, things work as expected >>>>>>>>> - UI looks good >>>>>>>>> - Logs look good >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Thank you~ >>>>>>>>> >>>>>>>>> Xintong Song >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Thu, Dec 3, 2020 at 9:18 PM Rui Li<lirui.fu...@gmail.com> >>>>> wrote: >>>>>>>>>> +1 (non-binding) >>>>>>>>>> >>>>>>>>>> Built from source and verified hive connector tests for different >>>>>> hive >>>>>>>>>> versions. >>>>>>>>>> Setup a cluster to connect to a real hive warehouse and run some >>>>>>> queries >>>>>>>>>> successfully. >>>>>>>>>> >>>>>>>>>> On Thu, Dec 3, 2020 at 8:44 PM Xingbo Huang<hxbks...@gmail.com> >>>>>>> wrote: >>>>>>>>>>> +1 (non-binding) >>>>>>>>>>> >>>>>>>>>>> Checks: >>>>>>>>>>> 1. verified checksums and signatures >>>>>>>>>>> 2. build Flink with Scala 2.11 >>>>>>>>>>> 3. pip install pyflink in MacOS/CentOS under py35,py36,py37,py38 >>>>>>>>>>> 4. test Pandas UDAF/General UDAF/Python DataStream MapFunction >>>>>>>>>>> 5. start standalone cluster and submit a python udf job. >>>>>>>>>>> 6. verified NOTICE/LICENSE files of some regular modules >>>>>>>>>>> >>>>>>>>>>> I observed that the NOTICE file of flink-sql-connector-hbase-2.2 >>>>>> lists >>>>>>>>> 3 >>>>>>>>>>> dependencies that are not bundled in: >>>>>>>>>>> commons-lang:commons-lang:2.6 >>>>>>>>>>> org.apache.hbase:hbase-hadoop-compat:2.2.3 >>>>>>>>>>> org.apache.hbase:hbase-hadoop2-compat:2.2.3 >>>>>>>>>>> >>>>>>>>>>> I guess listing more than dependencies with apache licensed >>>>>> shouldn't >>>>>>>>> be >>>>>>>>>> a >>>>>>>>>>> blocker issue. I have opened a PR[1] to fix it. >>>>>>>>>>> >>>>>>>>>>> [1]https://github.com/apache/flink/pull/14299 >>>>>>>>>>> >>>>>>>>>>> Best, >>>>>>>>>>> Xingbo >>>>>>>>>>> >>>>>>>>>>> Robert Metzger<rmetz...@apache.org> 于2020年12月3日周四 下午5:36写道: >>>>>>>>>>> >>>>>>>>>>>> There's now a pull request for the announcement blog post, please >>>>>>>>> help >>>>>>>>>>>> checking it:https://github.com/apache/flink-web/pull/397 >>>>>>>>>>>> >>>>>>>>>>>> On Thu, Dec 3, 2020 at 9:03 AM Robert Metzger < >>>>> rmetz...@apache.org >>>>>>>>>>> wrote: >>>>>>>>>>>>> +1 (binding) >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> Checks: >>>>>>>>>>>>> - checksums seem correct >>>>>>>>>>>>> - source archive code compiles >>>>>>>>>>>>> - Compiled a test job against the staging repository >>>>>>>>>>>>> - launched a standalone cluster, ran some test jobs against it >>>>>>>>>>>>> - quickstart contains correct version >>>>>>>>>>>>> - regular jars contain correct NOTICE file >>>>>>>>>>>>> - Looked a bit over the output of >>>>>>>>>>>>> git diff release-1.11.2...release-1.12 -- "**/pom.xml" >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> I noticed that at least one more jar file contains an invalid >>>>>>>>> LICENSE >>>>>>>>>>>> file >>>>>>>>>>>>> in it's root. This has already been the case with Flink 1.11, >>>>> and >>>>>>>>>> from >>>>>>>>>>>> the >>>>>>>>>>>>> context (apache flink jar, all the other license and notice >>>>> files >>>>>>>>>> talk >>>>>>>>>>>>> about this being an Apache project) it should be clear that the >>>>>>>>>> license >>>>>>>>>>>>> file is not meant for the whole jar file contents. >>>>>>>>>>>>> I will still extend the automated LicenseChecker to resolve >>>>> this, >>>>>>>>>> but I >>>>>>>>>>>>> don't want to cancel the release because of this. >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> >>>>>>>>>>>>> On Wed, Dec 2, 2020 at 11:19 AM Robert Metzger < >>>>>>>>> rmetz...@apache.org> >>>>>>>>>>>>> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Hi everyone, >>>>>>>>>>>>>> >>>>>>>>>>>>>> We have resolved the licensing issue Chesnay found. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Please review and vote on the release candidate #3 for the >>>>>> version >>>>>>>>>>>>>> 1.12.0, as follows: >>>>>>>>>>>>>> >>>>>>>>>>>>>> [ ] +1, Approve the release >>>>>>>>>>>>>> [ ] -1, Do not approve the release (please provide specific >>>>>>>>>> comments) >>>>>>>>>>>>>> The complete staging area is available for your review, which >>>>>>>>>>> includes: >>>>>>>>>>>>>> * JIRA release notes [1a], and website release notes [1b] >>>>>>>>>>>>>> * the official Apache source release and binary convenience >>>>>>>>> releases >>>>>>>>>>> to >>>>>>>>>>>>>> be deployed to dist.apache.org [2], which are signed with the >>>>>> key >>>>>>>>>>> with >>>>>>>>>>>>>> fingerprint D9839159 [3], >>>>>>>>>>>>>> * all artifacts to be deployed to the Maven Central Repository >>>>>>>>> [4], >>>>>>>>>>>>>> * source code tag "release-1.12.0-rc3" [5] >>>>>>>>>>>>>> >>>>>>>>>>>>>> We will soon publish the PR for the release announcement blog >>>>>>>>> post! >>>>>>>>>>>>>> The vote will be open for at least 72 hours. It is adopted by >>>>>>>>>> majority >>>>>>>>>>>>>> approval, with at least 3 PMC affirmative votes. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thanks, >>>>>>>>>>>>>> Dian & Robert >>>>>>>>>>>>>> >>>>>>>>>>>>>> [1a] >>>>>>>>>>>>>> >>>>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12348263 >>>>>>>>>>>>>> [1b]https://github.com/apache/flink/pull/14195 >>>>>>>>>>>>>> [2] >>>>>>>>> https://dist.apache.org/repos/dist/dev/flink/flink-1.12.0-rc3/ >>>>>>>>>>>>>> [3]https://dist.apache.org/repos/dist/release/flink/KEYS >>>>>>>>>>>>>> [4] >>>>>>>>>>>>>> >>>>>>> https://repository.apache.org/content/repositories/orgapacheflink-1404 >>>>>>>>>>>>>> [5] >>>>>>>>> https://github.com/apache/flink/releases/tag/release-1.12.0-rc3 >>>>>>>>>> -- >>>>>>>>>> Best regards! >>>>>>>>>> Rui Li >>>>>>>>>> >