+1 to remove these two connectors. It's a lot of baggage in comparison to the workarounds that still make it possible to use older Kafka clusters.
On Fri, Aug 28, 2020 at 12:06 PM Aljoscha Krettek <aljos...@apache.org> wrote: > Yes, that should be the process. But I'd try it in a testing environment > before doing it in the production environment. > > Aljoscha > > On 27.08.20 11:42, Paul Lam wrote: > > Hi, > > > > I think it’s okay, given that we can either migrate to the universal > connector or still use the compatible 0.10/0.11 connector of 1.11 release > as Chesnay mentioned when upgrading to 1.12. > > > > IIUC, the migration process to the universal connector would be (please > correct me if I’m wrong): > > 1. Stop the job with a savepoint, committing the offset to Kafka brokers. > > 2. Modify user code, migrate to he universal connector, and change the > source operator id to discard the old connector states. > > 3. Start the job with the savepoint, and read Kafka from group offsets. > > > > Best, > > Paul Lam > > > >> 2020年8月27日 16:27,Aljoscha Krettek <aljos...@apache.org> 写道: > >> > >> @Konstantin: Yes, I'm talking about dropping those modules. We don't > have any special code for supporting Kafka 0.10/0.11 in the "modern" > connector, that comes from the Kafka Consumer/Producer code we're using. > >> > >> @Paul: The modern Kafka connector works with Kafka brokers as far back > as 0.10, would that be enough or do you still think we should have the > actual Kafka 0.10 Consumer code in Flink as well. > >> > >> Best, > >> Aljoscha > >> > >> On 25.08.20 23:15, Chesnay Schepler wrote: > >>> +1 to remove both the 1.10 and 1.11 connectors. > >>> The connectors have not been actively developed for some time. They > are basically just sitting around causing noise by causing test > instabilities and eating CI time. > >>> It would also allow us to really simplify the module structure of the > Kafka connectors. > >>> Users may continue to use the 1.11 version of the connectors with > future Flink versions, and we may even provide critical bug fixes in a 1.11 > bugfix release (albeit unlikely). > >>> While ultimately this is a separate topic I would also be in favor of > removing any migration paths we have from 0.11 to the universal connector; > >>> as these are already present in 1.11 users may migrate to the > universal connector before jumping to Flink 1.12+. > >>> On 25/08/2020 18:49, Konstantin Knauf wrote: > >>>> Hi Aljoscha, > >>>> > >>>> I am assuming you're asking about dropping the > flink-connector-kafka-0.10/0.11 modules, right? Or are you talking about > removing support for Kafka 0.10/0.11 from the universal connector? > >>>> > >>>> I am in favor of removing flink-connector-kafka-0.10/0.11 in the next > release. These modules would still be available in Flink 1.11- as a > reference, and could be used with Flink 1.12+ with small or no > modifications. To my knowledge, you also use the universal Kafka connector > with 0.10 brokers, but there might be a performance penalty if I remember > correctly. In general, I find it important to continuously reduce baggage > that accumulates over time and this seems like a good opportunity. > >>>> > >>>> Best, > >>>> > >>>> Konstantin > >>>> > >>>> On Tue, Aug 25, 2020 at 4:59 AM Paul Lam <paullin3...@gmail.com > <mailto:paullin3...@gmail.com> <mailto:paullin3...@gmail.com <mailto: > paullin3...@gmail.com>>> wrote: > >>>> > >>>> Hi Aljoscha, > >>>> > >>>> I'm lightly leaning towards keeping the 0.10 connector, for Kafka > >>>> 0.10 still has a steady user base in my observation. > >>>> > >>>> But if we drop 0.10 connector, can we ensure the users would be > >>>> able to smoothly migrate to 0.11 connector/universal connector? > >>>> > >>>> If I remember correctly, the universal connector is compatible > >>>> with 0.10 brokers, but I want to double check that. > >>>> > >>>> Best, > >>>> Paul Lam > >>>> > >>>>> 2020年8月24日 22:46,Aljoscha Krettek <aljos...@apache.org <mailto: > aljos...@apache.org> > >>>>> <mailto:aljos...@apache.org <mailto:aljos...@apache.org>>> 写道: > >>>>> > >>>>> Hi all, > >>>>> > >>>>> this thought came up on FLINK-17260 [1] but I think it would be > a > >>>>> good idea in general. The issue reminded us that Kafka didn't > >>>>> have an idempotent/fault-tolerant Producer before Kafka 0.11.0. > >>>>> By now we have had the "modern" Kafka connector that roughly > >>>>> follows new Kafka releases for a while and this one supports > >>>>> Kafka cluster versions as far back as 0.10.2.0 (I believe). > >>>>> > >>>>> What are your thoughts on removing support for older Kafka > >>>>> versions? And yes, I know that we had multiple discussions like > >>>>> this in the past but I'm trying to gauge the current sentiment. > >>>>> > >>>>> I'm cross-posting to the user-ml since this is important for > both > >>>>> users and developers. > >>>>> > >>>>> Best, > >>>>> Aljoscha > >>>>> > >>>>> [1] https://issues.apache.org/jira/browse/FLINK-17260 > >>>> > >>>> > >>>> > >>>> -- > >>>> > >>>> Konstantin Knauf > >>>> > >>>> https://twitter.com/snntrable > >>>> > >>>> https://github.com/knaufk > > > > > > -- Arvid Heise | Senior Java Developer <https://www.ververica.com/> Follow us @VervericaData -- Join Flink Forward <https://flink-forward.org/> - The Apache Flink Conference Stream Processing | Event Driven | Real Time -- Ververica GmbH | Invalidenstrasse 115, 10115 Berlin, Germany -- Ververica GmbH Registered at Amtsgericht Charlottenburg: HRB 158244 B Managing Directors: Timothy Alexander Steinert, Yip Park Tung Jason, Ji (Toni) Cheng