Yes, absolutely. Setting up another repository for Flink ML would be no problem.

On Sat, Dec 12, 2015 at 1:52 AM, Henry Saputra <henry.sapu...@gmail.com> wrote:
> I had small chat with Till about how to help manage Flink ML Libraries
> contributions, which use Flink ML as dependencies.
>
> I suppose if this approached is the way to go for Flink connectors,
> could we do the same for Flink ML libraries?
>
>
> - Henry
>
> On Fri, Dec 11, 2015 at 1:33 AM, Maximilian Michels <m...@apache.org> wrote:
>> We should have release branches which are in sync with the release
>> branches in the main repository. Connectors should be compatible
>> across minor releases. The versioning could be of the form
>> "flinkversion-connectorversion", e.g. 0.10-connector1.
>>
>>>The pluggable architecture is great! (why Don't we call it Flink plugins? my 
>>>2 cents)
>>
>> We can still change the name. IMHO "Plugins" is a bit broad since this
>> is currently only targeted at the connectors included in Flink.
>>
>>>Would we loose test coverage by putting the connectors into a separate 
>>>repository/maven project?
>>
>> Not necessarily. Two possibilities:
>>
>> 1) Run a connectors test jar during the normal Travis tests in the
>> main repository
>> 2) Trigger a Travis test run at the connectors repository upon a
>> commit into the main repository
>>
>> Option 1 seems like the better alternative because we would
>> immediately see if a change breaks the connectors. Of course, if
>> changes are made in the connectors repository, we would also run tests
>> with the main repository.
>>
>> On Thu, Dec 10, 2015 at 11:00 PM, jun aoki <ja...@apache.org> wrote:
>>> The pluggable architecture is great! (why Don't we call it Flink plugins?
>>> my 2 cents)
>>> It will be nice if we come up with an idea of what directory structure
>>> should look like before start dumping connectors (plugins).
>>> Also wonder what to do with versioning.
>>> At some point, for example, Twitter v1 connector could be compatible with
>>> flink 0.10 but Flume v2 connector could be compatible with trunk, etc. It
>>> should be taken consideration either in the directory structure or
>>> branching strategy.
>>>
>>> On Thu, Dec 10, 2015 at 7:12 AM, Aljoscha Krettek <aljos...@apache.org>
>>> wrote:
>>>
>>>> We would need to have a stable interface between the connectors and flink
>>>> and have very good checks that ensure that we don’t inadvertently break
>>>> things.
>>>>
>>>> > On 10 Dec 2015, at 15:45, Fabian Hueske <fhue...@gmail.com> wrote:
>>>> >
>>>> > Sounds like a good idea to me.
>>>> >
>>>> > +1
>>>> >
>>>> > Fabian
>>>> >
>>>> > 2015-12-10 15:31 GMT+01:00 Maximilian Michels <m...@apache.org>:
>>>> >
>>>> >> Hi squirrels,
>>>> >>
>>>> >> By this time, we have numerous connectors which let you insert data
>>>> >> into Flink or output data from Flink.
>>>> >>
>>>> >> On the streaming side we have
>>>> >>
>>>> >> - RollingSink
>>>> >> - Flume
>>>> >> - Kafka
>>>> >> - Nifi
>>>> >> - RabbitMQ
>>>> >> - Twitter
>>>> >>
>>>> >> On the batch side we have
>>>> >>
>>>> >> - Avro
>>>> >> - Hadoop compatibility
>>>> >> - HBase
>>>> >> - HCatalog
>>>> >> - JDBC
>>>> >>
>>>> >>
>>>> >> Many times we would have liked to release updates to the connectors or
>>>> >> even create new ones in between Flink releases. This is currently not
>>>> >> possible because the connectors are part of the main repository.
>>>> >>
>>>> >> Therefore, I have created a new repository at
>>>> >> https://git-wip-us.apache.org/repos/asf/flink-connectors.git. The idea
>>>> >> is to externalize the connectors to this repository. We can then
>>>> >> update and release them independently of the main Flink repository. I
>>>> >> think this will give us more flexibility in the development process.
>>>> >>
>>>> >> What do you think about this idea?
>>>> >>
>>>> >> Cheers,
>>>> >> Max
>>>> >>
>>>>
>>>>
>>>
>>>
>>> --
>>> -jun

Reply via email to