I hereby close the vote. The result will be announced in a separate thread.

On Thu, Aug 9, 2018 at 11:47 AM Till Rohrmann <trohrm...@apache.org> wrote:

> +1
>
> - Checked checksums and signatures
> - Verified that no unwanted binaries are contained in source release
> - Checked LICENSE and NOTICE file
> - Checked that all newly added dependencies have a compatible license
> - Checked that a local cluster can be started and stopped without
> exceptions in the log
> - Verified that SBT quickstarts are up to date
> - Verified that Java quickstarts work with IntelliJ
> - Verified that all Jepsen tests pass
> - Verified that e2e tests modulo test_sql_client.sh (see
> https://issues.apache.org/jira/browse/FLINK-10107) pass
>
> Cheers,
> Till
>
> On Thu, Aug 9, 2018 at 8:18 AM vino yang <yanghua1...@gmail.com> wrote:
>
>> +1
>>
>> - checkout 1.6 source code and successfully run `mvn clean package
>> -DskipTests`
>> - searched '1.5' and '1.5.2' in all modules pom file and successfully
>> verified flink version was changed
>> - successfully run table and sql test locally
>>
>> Thanks, vino.
>>
>>
>> Timo Walther <twal...@apache.org> 于2018年8月9日周四 上午3:24写道:
>>
>> > +1
>> >
>> > - successfully run `mvn clean verify` locally
>> > - successfully run end-to-end tests locally (except for SQL Client
>> > end-to-end test)
>> >
>> > Found a bug in the class loading of SQL JAR files. This is not a blocker
>> > but a bug that we should fix soon. As an easy workaround user should not
>> > use different Kafka versions as SQL Client dependencies.
>> >
>> > Regards,
>> > Timo
>> >
>> > Am 08.08.18 um 18:10 schrieb Dawid Wysakowicz:
>> > > +1
>> > >
>> > > - verified compilation, tests
>> > > - verified checksum and gpg files
>> > > - verified sbt templates (g8, quickstart) - run assemblies on local
>> > cluster
>> > >
>> > > - I could not execute the nightly-tests.sh though. The tests that were
>> > > failing most often are:
>> > >      - test_streaming_file_sink.sh
>> > >      - test_streaming_elasticsearch.sh
>> > >
>> > > Those are connectors though and it might be only tests flakiness so I
>> > > think it should not block the release.
>> > >
>> > > On 08/08/18 16:36, Chesnay Schepler wrote:
>> > >> I did not use the tools/list_deps.py script as I wasn't aware that it
>> > >> existed.
>> > >>
>> > >> Even if I were I wouldn't have used it and in fact would advocate for
>> > >> removing it.
>> > >> It manually parses and constructs dependency information which is
>> > >> utterly unnecessary as maven already provides this functionality,
>> with
>> > >> the added bonus of also accounting for dependencyManagement and
>> > >> transitive dependencies which we obviously have to take into account.
>> > >>
>> > >> I used this one-liner instead:
>> > >> |mvn dependency:list | ||grep| |":.*:.*:.*"| || ||grep| |-||v| |-e
>> > >> ||"Finished at"| |-e ||"Some problems"| || ||cut| |-d] -f2- | ||sed|
>> > >> |'s/:[a-z]*$//g'| || ||sort| |-u
>> > >>
>> > >> |which I have documented here:
>> > >> https://cwiki.apache.org/confluence/display/FLINK/Dependencies
>> > >>
>> > >> On 08.08.2018 15:06, Aljoscha Krettek wrote:
>> > >>> +1
>> > >>>
>> > >>> - verified checksum and gpg files
>> > >>> - verified LICENSE and NOTICE: NOTICE didn't change from 1.5,
>> LICENSE
>> > >>> had one unnecessary part removed
>> > >>>
>> > >>> Side comment: I'm not sure whether the "Verify that the LICENSE and
>> > >>> NOTICE file is correct for the binary and source releases" part is
>> > >>> valid anymore because we only have one LICENSE and NOTICE file. also
>> > >>> "The LICENSE and NOTICE files in flink-dist/src/main/flink-bin refer
>> > >>> to the binary distribution and mention all of Flink's Maven
>> > >>> dependencies as well" can be dropped because we don't have them
>> > anymore.
>> > >>>
>> > >>> I came to the same conclusion on dependencies. I used
>> > >>> tools/list_deps.py and diff'ed the output for 1.5 and 1.6, that's
>> > >>> probably what Chesnay also did ... :-)
>> > >>>
>> > >>>> On 8. Aug 2018, at 14:43, Chesnay Schepler <ches...@apache.org>
>> > wrote:
>> > >>>>
>> > >>>> +1
>> > >>>>
>> > >>>> - verified source release contains no binaries
>> > >>>> - verified correct versions in source release
>> > >>>> - verified compilation, tests and E2E-tests pass (on travis)
>> > >>>> - verified checksum and gpg files
>> > >>>>
>> > >>>> New dependencies (excluding dependencies where we simply depend on
>> a
>> > >>>> different version now):
>> > >>>>      Apache licensed:
>> > >>>>          io.confluent:common-utils:jar:3.3.1
>> > >>>>          io.confluent:kafka-schema-registry-client:jar:3.3.1
>> > >>>>          io.prometheus:simpleclient_pushgateway:jar:0.3.0
>> > >>>>          various Apache Nifi dependencies
>> > >>>>          various Apache Parquet dependencies
>> > >>>>          various ElasticSearch dependencies
>> > >>>>      CDDL:
>> > >>>>          javax.ws.rs:javax.ws.rs-api:jar:2.1
>> > >>>>      Bouncycastle (MIT-like):
>> > >>>>          org.bouncycastle:bcpkix-jdk15on:jar:1.59
>> > >>>>          org.bouncycastle:bcprov-jdk15on:jar:1.59
>> > >>>>      MIT:
>> > >>>>          org.projectlombok:lombok:jar:1.16.20
>> > >>>>
>> > >>>> On 08.08.2018 13:28, Till Rohrmann wrote:
>> > >>>>> Thanks for reporting these problems Chesnay. The usage string in
>> > >>>>> `standalone-job.sh` is out dated and should be updated. The same
>> > >>>>> applies to
>> > >>>>> the typo.
>> > >>>>>
>> > >>>>> When calling `standalone-job.sh start --job-classname foobar.Job`
>> > >>>>> please
>> > >>>>> make sure that the user code jar is contained in the classpath
>> (e.g.
>> > >>>>> putting the jar in the lib directory). Documenting this behaviour
>> > >>>>> is part
>> > >>>>> of the pending issue FLINK-10001.
>> > >>>>>
>> > >>>>> We should fix all of these issues. They are, however, no release
>> > >>>>> blockers.
>> > >>>>>
>> > >>>>> Cheers,
>> > >>>>> Till
>> > >>>>>
>> > >>>>> On Wed, Aug 8, 2018 at 11:31 AM Chesnay Schepler
>> > >>>>> <ches...@apache.org> wrote:
>> > >>>>>
>> > >>>>>> I found some issues with the standalone-job.sh script.
>> > >>>>>>
>> > >>>>>> I ran "./bin/standalone-job.sh start" as described by the usage
>> > >>>>>> string.
>> > >>>>>>
>> > >>>>>>       2018-08-08 09:22:34,385 ERROR
>> > >>>>>>       org.apache.flink.runtime.entrypoint.ClusterEntrypoint
>> > -
>> > >>>>>>       Could not parse command line arguments [--configDir,
>> > >>>>>>       /home/zento/svn/flink-1.6.0/flink-1.6.0/conf].
>> > >>>>>>       org.apache.flink.runtime.entrypoint.FlinkParseException:
>> > >>>>>> Failed to
>> > >>>>>>       parse the command line arguments.
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:52)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:143)
>> > >>>>>>
>> > >>>>>>       Caused by: org.apache.commons.cli.MissingOptionException:
>> > >>>>>> Missing
>> > >>>>>>       required option: j
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.commons.cli.DefaultParser.checkRequiredOptions(DefaultParser.java:199)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:130)
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>> org.apache.commons.cli.DefaultParser.parse(DefaultParser.java:81)
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.parser.CommandLineParser.parse(CommandLineParser.java:50)
>> > >>>>>>
>> > >>>>>>               ... 1 more
>> > >>>>>>
>> > >>>>>> The script should fail earlier if no jar is provided, with a
>> better
>> > >>>>>> error message.
>> > >>>>>> It is also undocumented, and the usage instructions don't appear
>> > >>>>>> correct.
>> > >>>>>>
>> > >>>>>> Passing a jar with the -j option leads to a
>> ClassNotFoundException:
>> > >>>>>> "./bin/standalone-job.sh start -j
>> examples/streaming/WordCount.jar"
>> > >>>>>>
>> > >>>>>>       2018-08-08 09:26:30,562 ERROR
>> > >>>>>>       org.apache.flink.runtime.entrypoint.ClusterEntrypoint
>> > -
>> > >>>>>>       Cluster initialization failed.
>> > >>>>>>       java.lang.reflect.UndeclaredThrowableException
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1854)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.security.HadoopSecurityContext.runSecured(HadoopSecurityContext.java:41)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startCluster(ClusterEntrypoint.java:189)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.main(StandaloneJobClusterEntryPoint.java:158)
>> > >>>>>>
>> > >>>>>>       Caused by: org.apache.flink.util.FlinkException: Could not
>> > >>>>>> load the
>> > >>>>>>       provied entrypoint class.
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:92)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.retrieveJobGraph(StandaloneJobClusterEntryPoint.java:75)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.JobClusterEntrypoint.createDispatcher(JobClusterEntrypoint.java:107)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.startClusterComponents(ClusterEntrypoint.java:353)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.runCluster(ClusterEntrypoint.java:232)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.runtime.entrypoint.ClusterEntrypoint.lambda$startCluster$0(ClusterEntrypoint.java:190)
>> > >>>>>>
>> > >>>>>>                at
>> > >>>>>> java.security.AccessController.doPrivileged(Native Method)
>> > >>>>>>                at
>> javax.security.auth.Subject.doAs(Subject.java:422)
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1836)
>> > >>>>>>
>> > >>>>>>                ... 3 more
>> > >>>>>>       Caused by: java.lang.ClassNotFoundException:
>> > >>>>>>       examples/streaming/WordCount.jar
>> > >>>>>>                at
>> > >>>>>> java.net.URLClassLoader.findClass(URLClassLoader.java:381)
>> > >>>>>>                at
>> > >>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:424)
>> > >>>>>>                at
>> > >>>>>>
>>  sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:338)
>> > >>>>>>                at
>> > >>>>>> java.lang.ClassLoader.loadClass(ClassLoader.java:357)
>> > >>>>>>                at
>> > >>>>>>
>> > >>>>>>
>> >
>> org.apache.flink.container.entrypoint.StandaloneJobClusterEntryPoint.createPackagedProgram(StandaloneJobClusterEntryPoint.java:89)
>> > >>>>>>
>> > >>>>>>                ... 11 more
>> > >>>>>>
>> > >>>>>> So this seems to not work at all, but maybe I'm using it wrong?
>> > >>>>>>
>> > >>>>>> (There's also typo in "Could not load the provied entrypoint
>> class")
>> > >>>>>>
>> > >>>>>> On 08.08.2018 10:33, Piotr Nowojski wrote:
>> > >>>>>>> +1 from my side
>> > >>>>>>>
>> > >>>>>>> I’ve spent some time playing around with various examples
>> > (batching,
>> > >>>>>> streaming and SQL) on EMR 6 nodes cluster with yarn deployment,
>> with
>> > >>>>>> different configuration options (number of task
>> > >>>>>> managers/memory/Flip6/credit base flow control/metrics) and
>> > >>>>>> everything
>> > >>>>>> looks now fine (after fixing
>> > >>>>>> https://issues.apache.org/jira/browse/FLINK-9969 <
>> > >>>>>> https://issues.apache.org/jira/browse/FLINK-9969> ).
>> > >>>>>>> Piotrek
>> > >>>>>>>
>> > >>>>>>>> On 7 Aug 2018, at 17:17, Till Rohrmann <trohrm...@apache.org>
>> > >>>>>>>> wrote:
>> > >>>>>>>>
>> > >>>>>>>> Hi everyone,
>> > >>>>>>>> Please review and vote on the release candidate #4 for the
>> version
>> > >>>>>> 1.6.0,
>> > >>>>>>>> as follows:
>> > >>>>>>>> [ ] +1, Approve the release
>> > >>>>>>>> [ ] -1, Do not approve the release (please provide specific
>> > >>>>>>>> comments)
>> > >>>>>>>>
>> > >>>>>>>>
>> > >>>>>>>> The complete staging area is available for your review, which
>> > >>>>>>>> includes:
>> > >>>>>>>> * JIRA release notes [1],
>> > >>>>>>>> * the official Apache source release and binary convenience
>> > >>>>>>>> releases to
>> > >>>>>> be
>> > >>>>>>>> deployed to dist.apache.org [2], which are signed with the key
>> > with
>> > >>>>>>>> fingerprint 1F302569A96CFFD5 [3],
>> > >>>>>>>> * all artifacts to be deployed to the Maven Central Repository
>> > [4],
>> > >>>>>>>> * source code tag "release-1.6.0-rc4" [5],
>> > >>>>>>>> * website pull request listing the new release and adding
>> > >>>>>>>> announcement
>> > >>>>>> blog
>> > >>>>>>>> post [6].
>> > >>>>>>>>
>> > >>>>>>>> Please use this document for coordinating testing efforts: [7]
>> > >>>>>>>>
>> > >>>>>>>> The vote will be shortened since we only adde a minor fix on
>> top
>> > >>>>>>>> of the
>> > >>>>>> RC
>> > >>>>>>>> 3. It will close on Wednesday 6:30pm CET. It is adopted by
>> > majority
>> > >>>>>>>> approval, with at least 3 PMC affirmative votes.
>> > >>>>>>>>
>> > >>>>>>>> Thanks,
>> > >>>>>>>> Your friendly Release Manager
>> > >>>>>>>>
>> > >>>>>>>> [1]
>> > >>>>>>>>
>> > >>>>>>
>> >
>> https://issues.apache.org/jira/secure/ReleaseNote.jspa?projectId=12315522&version=12342760
>> > >>>>>>
>> > >>>>>>>> [2] https://dist.apache.org/repos/dist/dev/flink/flink-1.6.0/
>> > >>>>>>>> [3] https://dist.apache.org/repos/dist/release/flink/KEYS
>> > >>>>>>>> [4]
>> > >>>>>>
>> > https://repository.apache.org/content/repositories/orgapacheflink-1178
>> > >>>>>>
>> > >>>>>>>> [5] https://github.com/apache/flink/tree/release-1.6.0-rc4
>> > >>>>>>>> [6] https://github.com/apache/flink-web/pull/117
>> > >>>>>>>> [7]
>> > >>>>>>>>
>> > >>>>>>
>> >
>> https://docs.google.com/document/d/1upBFZQ7tbaSkYvDiLqfUFXKg8Xxs-lVheEfb66e4jpo/edit?usp=sharing
>> > >>>>>>
>> > >>>>>>>> Pro-tip: you can create a settings.xml file with these
>> contents:
>> > >>>>>>>>
>> > >>>>>>>> <settings>
>> > >>>>>>>> <activeProfiles>
>> > >>>>>>>>     <activeProfile>flink-1.6.0</activeProfile>
>> > >>>>>>>> </activeProfiles>
>> > >>>>>>>> <profiles>
>> > >>>>>>>>     <profile>
>> > >>>>>>>>       <id>flink-1.6.0</id>
>> > >>>>>>>>       <repositories>
>> > >>>>>>>>         <repository>
>> > >>>>>>>>           <id>flink-1.6.0</id>
>> > >>>>>>>>           <url>
>> > >>>>>>>>
>> > >>>>>>>>
>> > https://repository.apache.org/content/repositories/orgapacheflink-1178/
>> > >>>>>>>>
>> > >>>>>>>>           </url>
>> > >>>>>>>>         </repository>
>> > >>>>>>>>         <repository>
>> > >>>>>>>>           <id>archetype</id>
>> > >>>>>>>>           <url>
>> > >>>>>>>>
>> > >>>>>>>>
>> > https://repository.apache.org/content/repositories/orgapacheflink-1178/
>> > >>>>>>>>
>> > >>>>>>>>           </url>
>> > >>>>>>>>         </repository>
>> > >>>>>>>>       </repositories>
>> > >>>>>>>>     </profile>
>> > >>>>>>>> </profiles>
>> > >>>>>>>> </settings>
>> > >>>>>>>>
>> > >>>>>>>> And reference that in you maven commands via --settings
>> > >>>>>>>> path/to/settings.xml. This is useful for creating a quickstart
>> > >>>>>>>> based on
>> > >>>>>> the
>> > >>>>>>>> staged release and for building against the staged jars.
>> > >>
>> >
>> >
>>
>

Reply via email to