Hi Etienne, I think it all depends on the actual maintainers of the connector to make a decision on that: if their unreleased version of the connector should be compatible with a new Flink version, then they should test against it. For example, that's already done at Elasticsearch [1] and JDBC [2].
Choosing which versions to support is a decision by the maintainers in the community, and it always requires an action by a maintainer to update the CI config to set the correct versions whenever a new Flink version is released. Best regards, Martijn [1] https://github.com/apache/flink-connector-elasticsearch/blob/main/.github/workflows/push_pr.yml [2] https://github.com/apache/flink-connector-jdbc/blob/main/.github/workflows/push_pr.yml On Wed, Jun 28, 2023 at 6:09 PM Etienne Chauchot <echauc...@apache.org> wrote: > > Hi all, > > Connectors are external to flink. As such, they need to be tested > against stable (released) versions of Flink. > > But I was wondering if it would make sense to test connectors in PRs > also against latest flink master snapshot to allow to discover failures > before merging the PRs, ** while the author is still available **, > rather than discovering them in nightly tests (that test against > snapshot) after the merge. That would allow the author to anticipate > potential failures and provide more future proof code (even if master is > subject to change before the connector release). > > Of course, if a breaking change was introduced in master, such tests > will fail. But they should be considered as a preview of how the code > will behave against the current snapshot of the next flink version. > > WDYT ? > > > Best > > Etienne