Hi Piotr, Thanks for the comments!
bq. How are you planning to execute the end-to-end benchmarks and integrate them with our build process? Great question! We plan to execute the end-to-end benchmark in a small cluster (like 3 vm nodes) to better reflect network cost, triggering it through our Jenkins service for micro benchmark and show the result on code-speed center. Will add these into FLIP document if no objections. bq. Are you planning to monitor the throughput and latency at the same time? Good question. And you're right, we will stress the cluster to back-pressure and watch the throughput, latency doesn't mean much in the first test suites. Let me refine the document. Thanks. Best Regards, Yu On Wed, 30 Oct 2019 at 19:07, Piotr Nowojski <pi...@ververica.com> wrote: > Hi Yu, > > Thanks for bringing this up. > > +1 for the idea and the proposal from my side. > > I think that the proposed Test Job List might be a bit > redundant/excessive, but: > - we can always adjust this later, once we have the infrastructure in place > - as long as we have the computing resources and ability to quickly > interpret the results/catch regressions, it doesn’t hurt to have more > benchmarks/tests then strictly necessary. > > Which brings me to a question. How are you planning to execute the > end-to-end benchmarks and integrate them with our build process? > > Another smaller question: > > > In this initial stage we will only monitor and display job throughput > and latency. > > Are you planning to monitor the throughput and latency at the same time? > It might be a bit problematic, as when measuring the throughput you want to > saturate the system and hit some bottleneck, which will cause a > back-pressure (measuring latency at the same time when system is back > pressured doesn’t make much sense). > > Piotrek > > > On 30 Oct 2019, at 11:54, Yu Li <car...@gmail.com> wrote: > > > > Hi everyone, > > > > We would like to propose FLIP-83 that adds an end-to-end performance > > testing framework for Flink. We discovered some potential problems > through > > such an internal end-to-end performance testing framework before the > > release of 1.9.0 [1], so we'd like to contribute it to Flink community > as a > > supplement to the existing daily run micro performance benchmark [2] and > > nightly run end-to-end stability test [3]. > > > > The FLIP document could be found here: > > > https://cwiki.apache.org/confluence/display/FLINK/FLIP-83%3A+Flink+End-to-end+Performance+Testing+Framework > > > > Please kindly review the FLIP document and let us know if you have any > > comments/suggestions, thanks! > > > > [1] https://s.apache.org/m8kcq > > [2] https://github.com/dataArtisans/flink-benchmarks > > [3] https://github.com/apache/flink/tree/master/flink-end-to-end-tests > > > > Best Regards, > > Yu > >