Sounds like a good idea to me.

+1

Fabian

2015-12-10 15:31 GMT+01:00 Maximilian Michels <m...@apache.org>:

> Hi squirrels,
>
> By this time, we have numerous connectors which let you insert data
> into Flink or output data from Flink.
>
> On the streaming side we have
>
> - RollingSink
> - Flume
> - Kafka
> - Nifi
> - RabbitMQ
> - Twitter
>
> On the batch side we have
>
> - Avro
> - Hadoop compatibility
> - HBase
> - HCatalog
> - JDBC
>
>
> Many times we would have liked to release updates to the connectors or
> even create new ones in between Flink releases. This is currently not
> possible because the connectors are part of the main repository.
>
> Therefore, I have created a new repository at
> https://git-wip-us.apache.org/repos/asf/flink-connectors.git. The idea
> is to externalize the connectors to this repository. We can then
> update and release them independently of the main Flink repository. I
> think this will give us more flexibility in the development process.
>
> What do you think about this idea?
>
> Cheers,
> Max
>

Reply via email to