+1  (non-biding)

checklist
- checksum & gpg  ok
- build from source, ok
- all pom version point to 1.12.0
- run some program in local env, ok, and no strange log showed

Best,
Congxian


Dawid Wysakowicz <dwysakow...@apache.org> 于2020年12月7日周一 下午5:52写道:

> +1 (binding)
>
> * Verified the checksums
> * Verified that the source archives do not contain any binaries
> * Built Flink from sources
> * Run a streaming WorldCount example in BATCH and STREAM mode
> * Run a slightly heavier WorldCount version in BATCH and STREAM mode
> * Verified licensing of Hbase connectors*** I found one issue that we do not 
> list protobuf-java-util in flink-sql-connector-hbase-2.2 NOTICE file, which 
> is pulled in through hbase-shaded-miscellaneous (there is a couple more 
> dependencies we do not list from it, but they are Apache licensed)*
> * Run some Table examples (I found two examples are not runnable[1]), but I 
> would not consider it a blocker
>
> Let me know what you think about the licensing issue.
>
> Best,
> Dawid
>
> [1]https://issues.apache.org/jira/browse/FLINK-20464
>
> On 07/12/2020 08:07, Leonard Xu wrote:
>
> +1 (non-binding)
>
> - checked/verified signatures and hashes
> - built from source code with scala 2.11 succeeded
> - checked that there are no missing artifacts
> - started a cluster, WebUI was accessible, submitted a wordcount job and ran 
> succeeded, no suspicious log output
> - tested using SQL Client to submit job and the query result is expected
> - tested read/write from/to sql kafka/upsert-kafka connector in SQL Client
> - tested read/write/join hvie table in SQL Client
>
> Best,
> Leonard Xu
>
>
>
> 在 2020年12月7日,11:17,Guowei Ma <guowei....@gmail.com> <guowei....@gmail.com> 写道:
>
> +1(non-binding)
> - build from source
> - build a docker image
> - start a session from local k8s cluster
> - submit a wordcount job in streaming mode.
> - submit a wordcount job in batch mode.
> Best,
> Guowei
>
>
> On Sat, Dec 5, 2020 at 3:13 PM Zhu Zhu <reed...@gmail.com> 
> <reed...@gmail.com> wrote:
>
>
> +1 (binding)
>
> - verified signature and checksum
> - built from source
> - run testing jobs on yarn with manually triggered failures. checked logs
> and WebUI of those jobs
>  * DataStream job (paralelism=1000) with multiple disjoint pipelined
> regions
>  * DataSet job (paralelism=1000) with all edges blocking
>
> Thanks,
> Zhu
>
> Till Rohrmann <trohrm...@apache.org> <trohrm...@apache.org> 于2020年12月4日周五 
> 下午11:45写道:
>
>
> +1 (binding)
>
> * Verified the checksums
> * Ran RC on Minikube cluster
> ** Session mode
> ** Application mode
> * Built Flink from sources
>
> Cheers,
> Till
>
> On Fri, Dec 4, 2020 at 2:15 PM Wei Zhong <weizhong0...@gmail.com> 
> <weizhong0...@gmail.com> wrote:
>
>
> +1 (non-binding)
>
> - verified checksums and signatures
> - build Flink with Scala 2.11
> - pip install pyflink on Windows python 3.7
> - run a python job with udfs on Windows
> - pyflink shell works well on local mode and remote mode
>
> Best,
> Wei
>
>
> 在 2020年12月4日,17:21,Yang Wang <danrtsey...@gmail.com> <danrtsey...@gmail.com> 
> 写道:
>
> +1 (non-binding)
>
> * Build from source
> * Deploy Flink cluster in following deployments with HA
>
> enabled(ZooKeeper
>
> and K8s), including kill JobManager and check failover
> * Native K8s Session
> * Native K8s Application
> * Yarn Session
> * Yarn Per-Job
> * Yarn Application
> * Check webUI and logs in different deployments especially via
>
> `kubectl
>
> logs` in K8s
>
> Best,
> Yang
>
> Xintong Song <tonysong...@gmail.com> <tonysong...@gmail.com> 于2020年12月4日周五 
> 下午3:00写道:
>
>
> +1 (non-binding)
>
>  - Verified checksums and signatures
>  - No binaries found in source archive
>  - Build from source
>  - Tried a couple of example jobs in various deployment mode
>     - Local
>     - Standalone
>     - Native Kubernetes Application
>     - Native Kubernetes Session
>     - Yarn Job
>     - Yarn Session
>  - Changing memory configurations, things work as expected
>  - UI looks good
>  - Logs look good
>
>
>
> Thank you~
>
> Xintong Song
>
>
>
> On Thu, Dec 3, 2020 at 9:18 PM Rui Li <lirui.fu...@gmail.com> 
> <lirui.fu...@gmail.com>
>
> wrote:
>
> +1 (non-binding)
>
> Built from source and verified hive connector tests for different
>
> hive
>
> versions.
> Setup a cluster to connect to a real hive warehouse and run some
>
> queries
>
> successfully.
>
> On Thu, Dec 3, 2020 at 8:44 PM Xingbo Huang <hxbks...@gmail.com> 
> <hxbks...@gmail.com>
>
> wrote:
>
> +1 (non-binding)
>
> Checks:
> 1. verified checksums and signatures
> 2. build Flink with Scala 2.11
> 3. pip install pyflink in MacOS/CentOS under py35,py36,py37,py38
> 4. test Pandas UDAF/General UDAF/Python DataStream MapFunction
> 5. start standalone cluster and submit a python udf job.
> 6. verified NOTICE/LICENSE files of some regular modules
>
> I observed that the NOTICE file of flink-sql-connector-hbase-2.2
>
> lists
>
> 3
>
> dependencies that are not bundled in:
> commons-lang:commons-lang:2.6
> org.apache.hbase:hbase-hadoop-compat:2.2.3
> org.apache.hbase:hbase-hadoop2-compat:2.2.3
>
> I guess listing more than dependencies with apache licensed
>
> shouldn't
>
> be
>
> a
>
> blocker issue. I have opened a PR[1] to fix it.
>
> [1] https://github.com/apache/flink/pull/14299
>
> Best,
> Xingbo
>
> Robert Metzger <rmetz...@apache.org> <rmetz...@apache.org> 于2020年12月3日周四 
> 下午5:36写道:
>
>
> There's now a pull request for the announcement blog post, please
>
> help
>
> checking it: https://github.com/apache/flink-web/pull/397
>
> On Thu, Dec 3, 2020 at 9:03 AM Robert Metzger <
>
> rmetz...@apache.org
>
> wrote:
>
> +1 (binding)
>
>
> Checks:
> - checksums seem correct
> - source archive code compiles
> - Compiled a test job against the staging repository
> - launched a standalone cluster, ran some test jobs against it
> - quickstart contains correct version
> - regular jars contain correct NOTICE file
> - Looked a bit over the output of
>    git diff release-1.11.2...release-1.12 --  "**/pom.xml"
>
>
>
> I noticed that at least one more jar file contains an invalid
>
> LICENSE
>
> file
>
> in it's root. This has already been the case with Flink 1.11,
>
> and
>
> from
>
> the
>
> context (apache flink jar, all the other license and notice
>
> files
>
> talk
>
> about this being an Apache project) it should be clear that the
>
> license
>
> file is not meant for the whole jar file contents.
> I will still extend the automated LicenseChecker to resolve
>
> this,
>
> but I
>
> don't want to cancel the release because of this.
>
>
>
> On Wed, Dec 2, 2020 at 11:19 AM Robert Metzger <
>
> rmetz...@apache.org>
>
> wrote:
>
>
> Hi everyone,
>
> We have resolved the licensing issue Chesnay found.
>
> Please review and vote on the release candidate #3 for the
>
> version
>
> 1.12.0, as follows:
>
> [ ] +1, Approve the release
> [ ] -1, Do not approve the release (please provide specific
>
> comments)
>
> The complete staging area is available for your review, which
>
> includes:
>
> * JIRA release notes [1a], and website release notes [1b]
> * the official Apache source release and binary convenience
>
> releases
>
> to
>
> be deployed to dist.apache.org [2], which are signed with the
>
> key
>
> with
>
> fingerprint D9839159 [3],
> * all artifacts to be deployed to the Maven Central Repository
>
> [4],
>
> * source code tag "release-1.12.0-rc3" [5]
>
> We will soon publish the PR for the release announcement blog
>
> post!
>
> The vote will be open for at least 72 hours. It is adopted by
>
> majority
>
> approval, with at least 3 PMC affirmative votes.
>
> Thanks,
> Dian & Robert
>
> [1a]
>
>
> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12348263
>
> [1b] https://github.com/apache/flink/pull/14195
> [2]
>
> https://dist.apache.org/repos/dist/dev/flink/flink-1.12.0-rc3/
>
> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
> [4]
>
>
> https://repository.apache.org/content/repositories/orgapacheflink-1404
>
> [5]
>
> https://github.com/apache/flink/releases/tag/release-1.12.0-rc3
>
> --
> Best regards!
> Rui Li
>
>
>

Reply via email to