I had small chat with Till about how to help manage Flink ML Libraries
contributions, which use Flink ML as dependencies.

I suppose if this approached is the way to go for Flink connectors,
could we do the same for Flink ML libraries?


- Henry

On Fri, Dec 11, 2015 at 1:33 AM, Maximilian Michels <m...@apache.org> wrote:
> We should have release branches which are in sync with the release
> branches in the main repository. Connectors should be compatible
> across minor releases. The versioning could be of the form
> "flinkversion-connectorversion", e.g. 0.10-connector1.
>
>>The pluggable architecture is great! (why Don't we call it Flink plugins? my 
>>2 cents)
>
> We can still change the name. IMHO "Plugins" is a bit broad since this
> is currently only targeted at the connectors included in Flink.
>
>>Would we loose test coverage by putting the connectors into a separate 
>>repository/maven project?
>
> Not necessarily. Two possibilities:
>
> 1) Run a connectors test jar during the normal Travis tests in the
> main repository
> 2) Trigger a Travis test run at the connectors repository upon a
> commit into the main repository
>
> Option 1 seems like the better alternative because we would
> immediately see if a change breaks the connectors. Of course, if
> changes are made in the connectors repository, we would also run tests
> with the main repository.
>
> On Thu, Dec 10, 2015 at 11:00 PM, jun aoki <ja...@apache.org> wrote:
>> The pluggable architecture is great! (why Don't we call it Flink plugins?
>> my 2 cents)
>> It will be nice if we come up with an idea of what directory structure
>> should look like before start dumping connectors (plugins).
>> Also wonder what to do with versioning.
>> At some point, for example, Twitter v1 connector could be compatible with
>> flink 0.10 but Flume v2 connector could be compatible with trunk, etc. It
>> should be taken consideration either in the directory structure or
>> branching strategy.
>>
>> On Thu, Dec 10, 2015 at 7:12 AM, Aljoscha Krettek <aljos...@apache.org>
>> wrote:
>>
>>> We would need to have a stable interface between the connectors and flink
>>> and have very good checks that ensure that we don’t inadvertently break
>>> things.
>>>
>>> > On 10 Dec 2015, at 15:45, Fabian Hueske <fhue...@gmail.com> wrote:
>>> >
>>> > Sounds like a good idea to me.
>>> >
>>> > +1
>>> >
>>> > Fabian
>>> >
>>> > 2015-12-10 15:31 GMT+01:00 Maximilian Michels <m...@apache.org>:
>>> >
>>> >> Hi squirrels,
>>> >>
>>> >> By this time, we have numerous connectors which let you insert data
>>> >> into Flink or output data from Flink.
>>> >>
>>> >> On the streaming side we have
>>> >>
>>> >> - RollingSink
>>> >> - Flume
>>> >> - Kafka
>>> >> - Nifi
>>> >> - RabbitMQ
>>> >> - Twitter
>>> >>
>>> >> On the batch side we have
>>> >>
>>> >> - Avro
>>> >> - Hadoop compatibility
>>> >> - HBase
>>> >> - HCatalog
>>> >> - JDBC
>>> >>
>>> >>
>>> >> Many times we would have liked to release updates to the connectors or
>>> >> even create new ones in between Flink releases. This is currently not
>>> >> possible because the connectors are part of the main repository.
>>> >>
>>> >> Therefore, I have created a new repository at
>>> >> https://git-wip-us.apache.org/repos/asf/flink-connectors.git. The idea
>>> >> is to externalize the connectors to this repository. We can then
>>> >> update and release them independently of the main Flink repository. I
>>> >> think this will give us more flexibility in the development process.
>>> >>
>>> >> What do you think about this idea?
>>> >>
>>> >> Cheers,
>>> >> Max
>>> >>
>>>
>>>
>>
>>
>> --
>> -jun

Reply via email to