This message intrigued me:

<https://lists.gnu.org/archive/html/lilypond-devel/2022-11/msg00222.html>

In it, Eric Benson reported a setup that allows testing new versions of LilyPond on a sizable body of work in a somewhat automated fashion.

Now, could automation like that also make use of the infrastructure for LilyPond's regression tests?

<http://lilypond.org/doc/v2.23/Documentation/contributor/regtest-comparison>

What effort/value would there be in making an enhanced convert-ly tool that tests a new version of LilyPond on a user's entire collection of work, reporting differences between old and new versions in performance and output?

Enabling something like this:

* New release of LilyPond comes out. Please test.

* Advanced users with large collections of LilyPond files do the equivalent of "make test-baseline," but for their collection instead of LilyPond's regtests. Elapsed time is recorded, also CPU and RAM info as seems good.

* New LilyPond gets installed

* Upgrade script runs convert-ly on the collection, first offering backup via convert-ly options or tarball-style.

* Equivalent of "make check" runs

* A report generates, optionally as email to lilypond-devel, with summary of regression test differences and old-vs-new elapsed time.

Ideally, this could quickly produce lots of good testing info for development versions of LilyPond, in a way encouraging user participation.
--
Karlin High
Missouri, USA

Reply via email to