Hello all, I have been working with Péter and Marton on externalizing python connectors [1] from the main repo to the connector repositories. We have the code moved and the CI running tests for Kafka and AWS Connectors. I am now looking into the release process.
When we undertake a Flink release we perform the following steps [2], regarding Python: 1/ run python build on CI, 2/ download Wheels artifacts, 3/ upload artifacts to the dist and 4/ deploy to pypi. The plan is to follow the same steps for connectors, using Github actions instead of Azure pipeline. Today we have a single pypi project for pyflink that contains all the Flink libs, apache-flink [3]. I propose we create a new pypi project per connector using the existing connector version, and following naming convention: apache-<connector-name>, for example: apache-flink-connector-aws, apache-flink-connector-kafka. Therefore to use a DataStream API connector in python, users would need to first install the lib, for example "python -m pip install apache-flink-connector-aws". Once we have consensus I will update the release process and perform a release of the flink-connector-aws project to test it end-to-end. I look forward to any feedback. Thanks, Danny [1] https://issues.apache.org/jira/browse/FLINK-33528 [2] https://cwiki.apache.org/confluence/display/FLINK/Creating+a+Flink+Release [3] https://pypi.org/project/apache-flink/