So what next? Shall we schedule nexmark runs and add a Bigquery sink to nexmark
output?
Le lundi 12 mars 2018 à 10:30 +0100, Etienne Chauchot a écrit :
> Thanks everyone for your comments and support.
>
> Le vendredi 09 mars 2018 à 21:28 +0000, Alan Myrvold a écrit :
> > Great ideas. I want to see a daily signal for anything that could prevent a
> > release from happening, and precommits
> > that are fast and reliable for areas that are commonly broken by code
> > changes.
> >
> > We are now running the java quickstarts daily on a cron schedule, using
> > direct, dataflow, and local spark and flink
> > in the beam_PostRelease_NightlySnapshot job, see
> > https://github.com/apache/beam/blob/master/release/build.gradle
> > This should provide a good signal for the examples integration tests
> > against these runners.
> >
> > As Kenn noted, the java_maveninstall also runs lots of tests. It would be
> > good to be more clear and intentional
> > about which tests run when, and to consider implementing additional "always
> > up" environments for use by the tests.
> >
> > Having the nexmark smoke tests run regularly and stored in a database would
> > really enhance our efforts, perhaps
> > starting with directrunner for the performance tests.
> Yes
>
> >
> > What area would have the most immediate impact? Nexmark smoke tests?
> Yes IMHO I think that Nexmark smoke tests would have a great return on
> investment. By just scheduling some of them (at
> first), we enable deep confidence in the runners on real user pipelines. In
> the past Nexmark has allowed to discover
> regressions in performance before a release and also to discover some bugs in
> some runners. But, please note that, for
> this last ability, Nexmark is limited currently: it only detects failures if
> an exception is thrown, there is no check
> of the correctness of the output PCollection because the aim was performance
> tests and there is no point adding a slow
> test for correctness. Nevertheless, if we store the output size (as I
> suggested in this thread), we can get a hint on
> a failure if the output size is different from the last stored output sizes.
>
> Etienne
>
> >
> >
> >
> >
> > On Fri, Mar 9, 2018 at 12:57 PM Kenneth Knowles <k...@google.com> wrote:
> > > On Fri, Mar 9, 2018 at 3:08 AM Etienne Chauchot <echauc...@apache.org>
> > > wrote:
> > > > Hi guys,
> > > >
> > > > I was looking at the various jenkins jobs and I wanted to submit a
> > > > proposition:
> > > >
> > > > - Validates runner tests: currently run at PostCommit for all the
> > > > runners. I think it is the quickest way to see
> > > > regressions. So keep it that way
> > > We've also toyed with precommit for runners where it is fast.
> > >
> > > > - Integration tests: AFAIK we only run the ones in examples module and
> > > > only on demand. What about running all
> > > > the IT (in
> > > > particular IO IT) as a cron job on a daily basis with direct runner?
> > > > Please note that it will require some
> > > > always up
> > > > backend infrastructure.
> > > I like this idea. We actually run more, but in postcommit. You can see
> > > the goal here: https://github.com/apache/be
> > > am/blob/master/.test-infra/jenkins/job_beam_PostCommit_Java_MavenInstall.groovy#L47
> > >
> > > There's no infrastructure set up that I see. It is only DirectRunner and
> > > DataflowRunner currently, as they are
> > > "always up". But so could be local Flink and Spark. Do the ITs spin up
> > > local versions of what they are connecting
> > > to?
> > >
> > > If we have adequate resources, I also think ValidatesRunner on a real
> > > cluster would add value, once we have the
> > > cluster set up / tear down or "always up".
> > >
> > > > - Performance tests: what about running Nexmark SMOKE test suite in
> > > > batch and streaming modes with all the
> > > > runners on a
> > > > daily basis and store the running times in a RRD database (to see
> > > > performance regressions)?
> > > I like this idea, too. I think we could do DirectRunner (and probably
> > > local Flink) as postcommit without being too
> > > expensive.
> > >
> > > Kenn
> > >
> > >
> > > > Please note that not all the
> > > > queries run in all the runners in all the modes right now. Also, we
> > > > have some streaming pipelines termination
> > > > issues
> > > > (see https://issues.apache.org/jira/browse/BEAM-2847)
> > > >
> > > > I know that Stephen Sisk use to work on these topics. I also talked to
> > > > guys from Polidea. But As I understood,
> > > > they
> > > > launch mainly integration tests on Dataflow runner.
> > > >
> > > > WDYT?
> > > >
> > > > Etienne
> > > >
> > > >
> > > >