Yes, that would be the way to go.

We could follow Cask CDAP hydrator plugin repository [1] that support
different plugins to run in their main CDAP hydrator [2]  product

- Henry

[1] https://github.com/caskdata/hydrator-plugins
[2] https://github.com/caskdata/cdap

On Mon, Dec 14, 2015 at 1:49 AM, Maximilian Michels <m...@apache.org> wrote:
>>
>> Regarding Max suggestion to have version compatible connectors: I'm not
>> sure if we are able to maintain all connectors across different releases.
>>
>
> That was not my suggestion. Whenever we release, existing connectors should
> be compatible with that release. Otherwise, they should be removed from the
> release branch. This doesn't imply every connector version should be
> compatible across all releases.
>
> On Mon, Dec 14, 2015 at 10:39 AM, Robert Metzger <rmetz...@apache.org>
> wrote:
>> Regarding Max suggestion to have version compatible connectors: I'm not
>> sure if we are able to maintain all connectors across different releases.
> I
>> think its okay to have a document describing the minimum required Flink
>> version for each connector.
>>
>> With the interface stability guarantees from 1.0 on, the number of
> breaking
>> changes will go down.
>>
>> I'm against the name "plugins" because everything (documentation, code,
>> code comments, ...) is called "connectors" and it would be a pretty
>> breaking change. I also think that "connector" describes much better what
>> the whole thing is about.
>>
>>
>>
>> On Mon, Dec 14, 2015 at 10:20 AM, Maximilian Michels <m...@apache.org>
> wrote:
>>
>>> Yes, absolutely. Setting up another repository for Flink ML would be no
>>> problem.
>>>
>>> On Sat, Dec 12, 2015 at 1:52 AM, Henry Saputra <henry.sapu...@gmail.com>
>>> wrote:
>>> > I had small chat with Till about how to help manage Flink ML Libraries
>>> > contributions, which use Flink ML as dependencies.
>>> >
>>> > I suppose if this approached is the way to go for Flink connectors,
>>> > could we do the same for Flink ML libraries?
>>> >
>>> >
>>> > - Henry
>>> >
>>> > On Fri, Dec 11, 2015 at 1:33 AM, Maximilian Michels <m...@apache.org>
>>> wrote:
>>> >> We should have release branches which are in sync with the release
>>> >> branches in the main repository. Connectors should be compatible
>>> >> across minor releases. The versioning could be of the form
>>> >> "flinkversion-connectorversion", e.g. 0.10-connector1.
>>> >>
>>> >>>The pluggable architecture is great! (why Don't we call it Flink
>>> plugins? my 2 cents)
>>> >>
>>> >> We can still change the name. IMHO "Plugins" is a bit broad since this
>>> >> is currently only targeted at the connectors included in Flink.
>>> >>
>>> >>>Would we loose test coverage by putting the connectors into a separate
>>> repository/maven project?
>>> >>
>>> >> Not necessarily. Two possibilities:
>>> >>
>>> >> 1) Run a connectors test jar during the normal Travis tests in the
>>> >> main repository
>>> >> 2) Trigger a Travis test run at the connectors repository upon a
>>> >> commit into the main repository
>>> >>
>>> >> Option 1 seems like the better alternative because we would
>>> >> immediately see if a change breaks the connectors. Of course, if
>>> >> changes are made in the connectors repository, we would also run tests
>>> >> with the main repository.
>>> >>
>>> >> On Thu, Dec 10, 2015 at 11:00 PM, jun aoki <ja...@apache.org> wrote:
>>> >>> The pluggable architecture is great! (why Don't we call it Flink
>>> plugins?
>>> >>> my 2 cents)
>>> >>> It will be nice if we come up with an idea of what directory
> structure
>>> >>> should look like before start dumping connectors (plugins).
>>> >>> Also wonder what to do with versioning.
>>> >>> At some point, for example, Twitter v1 connector could be compatible
>>> with
>>> >>> flink 0.10 but Flume v2 connector could be compatible with trunk,
> etc.
>>> It
>>> >>> should be taken consideration either in the directory structure or
>>> >>> branching strategy.
>>> >>>
>>> >>> On Thu, Dec 10, 2015 at 7:12 AM, Aljoscha Krettek <
> aljos...@apache.org
>>> >
>>> >>> wrote:
>>> >>>
>>> >>>> We would need to have a stable interface between the connectors and
>>> flink
>>> >>>> and have very good checks that ensure that we don’t inadvertently
>>> break
>>> >>>> things.
>>> >>>>
>>> >>>> > On 10 Dec 2015, at 15:45, Fabian Hueske <fhue...@gmail.com> wrote:
>>> >>>> >
>>> >>>> > Sounds like a good idea to me.
>>> >>>> >
>>> >>>> > +1
>>> >>>> >
>>> >>>> > Fabian
>>> >>>> >
>>> >>>> > 2015-12-10 15:31 GMT+01:00 Maximilian Michels <m...@apache.org>:
>>> >>>> >
>>> >>>> >> Hi squirrels,
>>> >>>> >>
>>> >>>> >> By this time, we have numerous connectors which let you insert
> data
>>> >>>> >> into Flink or output data from Flink.
>>> >>>> >>
>>> >>>> >> On the streaming side we have
>>> >>>> >>
>>> >>>> >> - RollingSink
>>> >>>> >> - Flume
>>> >>>> >> - Kafka
>>> >>>> >> - Nifi
>>> >>>> >> - RabbitMQ
>>> >>>> >> - Twitter
>>> >>>> >>
>>> >>>> >> On the batch side we have
>>> >>>> >>
>>> >>>> >> - Avro
>>> >>>> >> - Hadoop compatibility
>>> >>>> >> - HBase
>>> >>>> >> - HCatalog
>>> >>>> >> - JDBC
>>> >>>> >>
>>> >>>> >>
>>> >>>> >> Many times we would have liked to release updates to the
>>> connectors or
>>> >>>> >> even create new ones in between Flink releases. This is currently
>>> not
>>> >>>> >> possible because the connectors are part of the main repository.
>>> >>>> >>
>>> >>>> >> Therefore, I have created a new repository at
>>> >>>> >> https://git-wip-us.apache.org/repos/asf/flink-connectors.git. The
>>> idea
>>> >>>> >> is to externalize the connectors to this repository. We can then
>>> >>>> >> update and release them independently of the main Flink
>>> repository. I
>>> >>>> >> think this will give us more flexibility in the development
>>> process.
>>> >>>> >>
>>> >>>> >> What do you think about this idea?
>>> >>>> >>
>>> >>>> >> Cheers,
>>> >>>> >> Max
>>> >>>> >>
>>> >>>>
>>> >>>>
>>> >>>
>>> >>>
>>> >>> --
>>> >>> -jun
>>>

Reply via email to