+1 to more helpful guide on "how to usefully participate in RC validation" but also big +1 to Robert, Jack, Johanna.
TL;DR the RC validation is an opportunity for downstream testing. Robert alluded to the origin of the spreadsheet: I created it long ago to validate that the human language on our web page actually works. Maybe someone should automate that with an LLM now. Robert also alluded to clean environment: our gradle scripts and GHA scripts and CI environment are heavily enough engineered that they don't represent what a user will experience. We could potentially use our starter repos for an adequate smoke test here. Those are both ways that *we* can pretend to be users. But actual users checking the RC to make sure they'll have a smooth upgrade is by far the most impactful validation. This thread honestly makes me want to delete the spreadsheet but maybe come up with a guide for downstream projects to validate against an RC. Maybe that's an extreme reaction... Kenn On Wed, Oct 18, 2023 at 2:32 PM Robert Bradshaw via dev <dev@beam.apache.org> wrote: > +1 That's a great idea. They have incentive to make sure the issue was > resolved for them, plus we get to ensure there were no other regressions. > > On Wed, Oct 18, 2023 at 11:30 AM Johanna Öjeling via dev < > dev@beam.apache.org> wrote: > >> When I have contributed to Apache Airflow, they have tagged all >> contributors concerned in a GitHub issue when the RC is available and asked >> us to validate it. Example: #29424 >> <https://github.com/apache/airflow/issues/29424>. >> >> I found that to be an effective way to notify contributors of the RC and >> nudge them to help out. In the issue description there is a reference to >> the guidelines on how to test the RC and a note that people are encouraged >> to vote on the mailing list (which could admittedly be more highlighted >> because I did not pay attention to it until now and was unaware that >> contributors had a vote). >> >> It might be an idea to consider something similar here to increase the >> participation? >> >> On Tue, Oct 17, 2023 at 7:01 PM Jack McCluskey via dev < >> dev@beam.apache.org> wrote: >> >>> I'm +1 on helping explain what we mean by "validate the RC" since we're >>> really just asking users to see if their existing use cases work along with >>> our typical slate of tests. I don't know if offloading that work to our >>> active validators is the right approach though, documentation/screen share >>> of their specific workflow is definitely less useful than having a more >>> general outline of how to install the RC and things to look out for when >>> testing. >>> >>> On Tue, Oct 17, 2023 at 12:55 PM Austin Bennett <aus...@apache.org> >>> wrote: >>> >>>> Great effort. I'm also interested in streamlining releases -- so if >>>> there are alot of manual tests that could be automated, would be great >>>> to discover and then look to address. >>>> >>>> On Tue, Oct 17, 2023 at 8:47 AM Robert Bradshaw via dev < >>>> dev@beam.apache.org> wrote: >>>> >>>>> +1 >>>>> >>>>> I would also strongly suggest that people try out the release against >>>>> their own codebases. This has the benefit of ensuring the release won't >>>>> break your own code when they go out, and stress-tests the new code >>>>> against >>>>> real-world pipelines. (Ideally our own tests are all passing, and this >>>>> validation is automated as much as possible (though ensuring it matches >>>>> our >>>>> documentation and works in a clean environment still has value), but >>>>> there's a lot of code and uses out there that we don't have access to >>>>> during normal Beam development.) >>>>> >>>>> On Tue, Oct 17, 2023 at 8:21 AM Svetak Sundhar via dev < >>>>> dev@beam.apache.org> wrote: >>>>> >>>>>> Hi all, >>>>>> >>>>>> I’ve participated in RC testing for a few releases and have observed >>>>>> a bit of a knowledge gap in how releases can be tested. Given that Beam >>>>>> encourages contributors to vote on RC’s regardless of tenure, and that >>>>>> voting on an RC is a relatively low-effort, high leverage way to >>>>>> influence >>>>>> the release of the library, I propose the following: >>>>>> >>>>>> During the vote for the next release, voters can document the process >>>>>> they followed on a separate document, and add the link on column G >>>>>> here >>>>>> <https://docs.google.com/spreadsheets/d/1qk-N5vjXvbcEk68GjbkSZTR8AGqyNUM-oLFo_ZXBpJw/edit#gid=437054928>. >>>>>> One step further, could be a screencast of running the test, and >>>>>> attaching >>>>>> a link of that. >>>>>> >>>>>> We can keep repeating this through releases until we have >>>>>> documentation for many of the different tests. We can then add these docs >>>>>> into the repo. >>>>>> >>>>>> I’m proposing this because I’ve gathered the following feedback from >>>>>> colleagues that are tangentially involved with Beam: They are interested >>>>>> in >>>>>> participating in release validation, but don’t know how to get started. >>>>>> Happy to hear other suggestions too, if there are any to address the >>>>>> above. >>>>>> >>>>>> Thanks, >>>>>> >>>>>> >>>>>> Svetak Sundhar >>>>>> >>>>>> Data Engineer >>>>>> s <nellywil...@google.com>vetaksund...@google.com >>>>>> >>>>>>