Hi Shivam,
a Kafka sink for the SQL Client will be part of Flink 1.6. For this we
need to do provide basic interfaces that sinks can extends as Rong
mentioned (FLINK-8866). In order to support all formats that also
sources support we also working on separating the connector from the
formats [1]. PR for these features are ready and I'm working on
integrating them right now. Once this is done and we have support for
INSERT INTO in SQL Client a Kafka sink implementation is straightforward.
Regards,
Timo
[1] https://issues.apache.org/jira/browse/FLINK-8558
Am 12.07.18 um 02:45 schrieb Rong Rong:
Hi Shivam,
Thank you for interested in contributing to Kafka Sink for SQL client.
Could you share your plan for implementation. I have some questions as
there might have been some overlap with current implementations.
On a higher level,
1. Are you using some type of metadata store to host topic schemas
(Kafka can essentially be schema-less), it might be great to take a
look at the TableSource/SinkFactory [1][2]
2. There's already a KafkaTableSource and KafkaTableSink available, I
am assuming you are trying to contribute to the configuration in SQL
Client to make it easier to interact with a Kafka table?
Thanks,
Rong
[1]: https://issues.apache.org/jira/browse/FLINK-8839
[2]: https://issues.apache.org/jira/browse/FLINK-8866
On Tue, Jul 10, 2018 at 3:28 AM Shivam Sharma
<28shivamsha...@gmail.com <mailto:28shivamsha...@gmail.com>> wrote:
Hi All,
We want to write Kafka Sink functionality for Flink(1.5) SQL
Client. We have read the code and chalk out a rough plan for
implementation.
Any guidance for this implementation will be very helpful.
Thanks
--
Shivam Sharma
Data Engineer @ Goibibo
Indian Institute Of Information Technology, Design and
Manufacturing Jabalpur
Mobile No- (+91) 8882114744
Email:- 28shivamsha...@gmail.com <mailto:28shivamsha...@gmail.com>
LinkedIn:-_https://www.linkedin.com/in/28shivamsharma_