I fully agree with Aljoscha and Chesnay (although my recent PR
experience was still close to what Stanislav describes).

@Robert: Do we have standard labels that we apply to tickets that
report a flaky test? I think this would be helpful to make sure that
we have a good overview of the state of flaky tests.

Best,

Ufuk

On Wed, Feb 27, 2019 at 3:04 PM Aljoscha Krettek <aljos...@apache.org> wrote:
>
> I agree with Chesnay, and I would like to add that the most important step 
> towards fixing flakiness is awareness and willingness. As soon as you accept 
> flakiness and start working around it (as you mentioned) more flakiness will 
> creep in, making it harder to get rid of it in the future.
>
> Aljoscha
>
> > On 27. Feb 2019, at 12:04, Chesnay Schepler <ches...@apache.org> wrote:
> >
> > We've been in the same position a while back with the same effects. We 
> > solved it by creating JIRAs for every failing test and cracking down hard 
> > on them; I don't think there's any other way to address this.
> > However to truly solve this one must look at the original cause to prevent 
> > new flaky tests from being added.
> > From what I remember, many of our tests were flaky because they relied on 
> > timings (e.g. lets Thread.sleep for X and assume Y has happened) or had 
> > similar race-conditions, and committers nowadays are rather observant for 
> > these issues.
> >
> > By now the majority of our builds succeed.
> > We don't to anything like running the builds multiple times before a merge. 
> > I know some committers always run a PR at least once against master, but 
> > this certainly doesn't apply to everyone.
> > There are still tests that fail from time-to-time, but my impressions is 
> > that people still check which tests are failing to ensure they are 
> > unrelated, and track them regardless.
> >
> > On 26.02.2019 17:28, Stanislav Kozlovski wrote:
> >> Hey there Flink community,
> >>
> >> I work on a fellow open-source project - Apache Kafka - and there we have 
> >> been fighting flaky tests a lot. We run Java 8 and Java 11 builds on every 
> >> Pull Request and due to test flakiness, almost all of them turn out red 
> >> with 1 or 2 tests (completely unrelated to the change in the PR) failing. 
> >> This has resulted in committers either ignoring them and merging the 
> >> changes or in the worst case rerunning the hour-long build until it 
> >> becomes green.
> >> This test flakiness has also slowed down our releases significantly.
> >>
> >> In general, I was just curious to understand if this is a problem that 
> >> your project faces as well. Does your project have a lot of intermittently 
> >> failing tests, do you have any active process of addressing such tests 
> >> (during the initial review, after realizing it is flaky, etc). Any 
> >> pointers will be greatly appreciated!
> >>
> >> Thanks,
> >> Stanislav
> >>
> >>
> >
>

Reply via email to