Sorry, the missing link for [3] is this: https://docs.google.com/document/d/1cOkycJwEKVjG_onnpl3bQNTq7uebh48zDtIJxceyU2E/edit <https://docs.google.com/document/d/1cOkycJwEKVjG_onnpl3bQNTq7uebh48zDtIJxceyU2E/edit>
Regarding build time, I recently merged this change: https://issues.apache.org/jira/browse/FLINK-8911 <https://issues.apache.org/jira/browse/FLINK-8911>. It introduces a separation between "pre-commit-tests" and "nightly-tests". The idea being the the former are executed for each pull request, as they are now. The latter are executed nightly or manually if you want to verify a release. This way, the nightly tests can get quite involved without blowing up build time. What do you think? Best, Aljoscha > On 10. Mar 2018, at 17:18, Bowen Li <bowenl...@gmail.com> wrote: > > This would be nice. BTW, the [3] link is missing > > I have a few questions, given concerns that 1) the full build time > will take longer and longer time and 2) build queue on Travis will be more > congested > > > - How much time will be added to full build time according to estimation? > - Do we have plans to further parallelize full build? > - Are new integration tests gonna run for every full build on both user > machine and Travis by default? > - Shall we add ways to skip it? > - if manually, add some flag or mvn parameter to disable it? > - if automatically, when there's new code changes, enable it to > recognize if the code change is trivial (like editing docs) and thus not > run those heavy, time-consuming integration tests? > > Thanks, > Bowen > > > On Fri, Mar 9, 2018 at 4:40 AM, Renjie Liu <liurenjie2...@gmail.com> wrote: > >> +1 >> On Fri, 9 Mar 2018 at 5:47 PM Aljoscha Krettek <aljos...@apache.org> >> wrote: >> >>> Hi All, >>> >>> Stephan pointed this out the other day to me, so here goes: as some of >>> you might now, there are end-to-end tests in flink-end-to-end tests that >>> run a proper Flink cluster (on the local machine) and execute some tests. >>> This catches bugs that you only catch when using Flink as a user because >>> they exercise the whole system. We should add tests there that verify >>> integration with other systems. For example, there's a bunch of >>> Docker-compose configurations for starting complete Hadoop clusters [1] >> or >>> Mesos [2] and there is other files for starting ZooKeeper, Kafka, ... We >>> can use this to spin up a testing cluster and run Flink on YARN and Mesos >>> and have a reproducible environment. >>> >>> As a next step, we could perform the sort of tests we do for a release, >> as >>> described here: [3]. For example, the test where we run a job and mess >> with >>> processes and see that Flink correctly recovers and that the HA setup >> works >>> as intended. >>> >>> What do you think? >>> >>> By the way, I'm also mostly writing to see if anyone has some experience >>> with Docker/Docker compose and would be interested in getting started on >>> this. I would do it myself because having more automated tests would help >>> be sleep better at night but I'm currently too busy with other things. 😉 >>> >>> [1] >>> https://github.com/big-data-europe/docker-hadoop/blob/ >> master/docker-compose.yml >>> [2] https://github.com/bobrik/mesos-compose >>> [3] >> >> -- >> Liu, Renjie >> Software Engineer, MVAD >>