I think the problem is less with bugs being introduced by new commits but
rather bugs which are already in the code base.

2015-06-04 11:52 GMT+02:00 Matthias J. Sax <mj...@informatik.hu-berlin.de>:

> I have another idea: the problem is, that some commit might de-stabilize
> a former stable test. This in not detected, because the build was
> ("accidentally") green and the code in merged.
>
> We could reduce the probability that this happens, if a pull request
> must pass the test-run multiple times (maybe 5x). Of course, it takes
> much time to run all test on Travis such often and increases the time
> until something can be merged. But it might be worth the effort.
>
> Options on that?
>
> On 06/04/2015 11:35 AM, Ufuk Celebi wrote:
> > Thanks for the feedback and the suggestions.
> >
> > As Stephan said, the "we have to fix it asap" usually does not work
> well. I think blocking master is not an option, exactly for the reasons
> that Fabian and Till outlined.
> >
> > From the comments so far, I don't feel like we are eager to adapt a
> disable policy.
> >
> > I still think it is a better policy. I think we actually don't decrease
> test coverage by disabling a flakey test, but increase it. For example the
> KafkaITCase is in one of the modules, which is tested in the middle of a
> build. If it fails (as it does sometimes), a lot of later tests don't run.
> I'm not sure if we have the time (or discipline) to trigger a 1hr build
> again when a known-to-fail test is failing and 4 of the other builds are
> succeeding.
> >
> > – Ufuk
> >
> > On 04 Jun 2015, at 09:25, Till Rohrmann <till.rohrm...@gmail.com> wrote:
> >
> >> I'm also in favour of quickly fixing the failing test cases but I think
> >> that blocking the master is a kind of drastic measure. IMO this creates
> a
> >> culture of blaming someone whereas I would prefer a more proactive
> >> approach. When you see a failing test case and know that someone
> recently
> >> worked on it, then ping him because maybe he can quickly fix it or knows
> >> about it. If he's not available, e.g. holidays, busy with other stuff,
> >> etc., then maybe one can investigate the problem oneself and fix it.
> >>
> >> But this is basically our current approach and I don't know how to
> enforce
> >> this policy by some means. Maybe it's making people more aware of it and
> >> motivating people to have a stable master.
> >>
> >> Cheers,
> >> Till
> >>
> >> On Thu, Jun 4, 2015 at 9:06 AM, Matthias J. Sax <
> >> mj...@informatik.hu-berlin.de> wrote:
> >>
> >>> I think, people should be forced to fixed failing tests asap. One way
> to
> >>> go, could be to lock the master branch until the test is fixed. If
> >>> nobody can push to the master, pressure is very high for the
> responsible
> >>> developer to get it done asap. Not sure if this is Apache compatible.
> >>>
> >>> Just a thought (from industry experience).
> >>>
> >>>
> >>> On 06/04/2015 08:10 AM, Aljoscha Krettek wrote:
> >>>> I tend to agree with Ufuk, although it would be nice to fix them very
> >>> quickly.
> >>>>
> >>>> On Thu, Jun 4, 2015 at 1:26 AM, Stephan Ewen <se...@apache.org>
> wrote:
> >>>>> @matthias: That is the implicit policy right now. Seems not to
> work...
> >>>>>
> >>>>> On Thu, Jun 4, 2015 at 12:40 AM, Matthias J. Sax <
> >>>>> mj...@informatik.hu-berlin.de> wrote:
> >>>>>
> >>>>>> I basically agree that the current policy on not optimal. However, I
> >>>>>> would rather give failing tests "top priority" to get fixed (if
> >>> possible
> >>>>>> within one/a-few days) and not disable them.
> >>>>>>
> >>>>>> -Matthias
> >>>>>>
> >>>>>> On 06/04/2015 12:32 AM, Ufuk Celebi wrote:
> >>>>>>> Hey all,
> >>>>>>>
> >>>>>>> we have certain test cases, which are failing regularly on Travis.
> In
> >>> all
> >>>>>>> cases I can think of we just keep the test activated.
> >>>>>>>
> >>>>>>> I think this makes it very hard for regular contributors to take
> these
> >>>>>>> failures seriously. I think the following situation is not
> unrealistic
> >>>>>> with
> >>>>>>> the current policy: I know that test X is failing. I don't know
> that
> >>>>>> person
> >>>>>>> Y fixed this test. I see test X failing (again for a different
> reason)
> >>>>>> and
> >>>>>>> think that it is a "known issue".
> >>>>>>>
> >>>>>>> I think a better policy is to just disable the test, assign
> someone to
> >>>>>> fix
> >>>>>>> it, and then only enable it again after someone has fixed it.
> >>>>>>>
> >>>>>>> Is this reasonable? Or do we have good reasons to keep such tests
> >>> (there
> >>>>>>> are currently one or two) activated?
> >>>>>>>
> >>>>>>> – Ufuk
> >>>>>>>
> >>>>>>
> >>>>>>
> >>>>
> >>>
> >>>
> >
> >
>
>

Reply via email to