Hello Shipeng, I am not an expert in Flink, just want to share some of my thoughts. Maybe others can give you better ideas. I think there is no directly available Protobuf support for Flink SQL. However, you can write a user-defined format to support it [1]. If you use DataStream API, you can leverage Kryo Serializer to serialize and deserialize with Protobuf format. [2]. There is an out-of-box integration for Protobuf here. You will need to convert it to Flink SQL after data ingestion.
[1] https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/sourceSinks.html#user-defined-sources-sinks [2] https://ci.apache.org/projects/flink/flink-docs-stable/dev/custom_serializers.html Best, Fuyao From: Shipeng Xie <xship...@vmware.com> Date: Friday, April 30, 2021 at 14:58 To: user@flink.apache.org <user@flink.apache.org> Subject: [External] : Protobuf support with Flink SQL and Kafka Connector Hi, In https://ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/<https://urldefense.com/v3/__https:/ci.apache.org/projects/flink/flink-docs-release-1.12/dev/table/connectors/formats/__;!!GqivPVa7Brio!P98PmXGjSsCo9MVOyHDZg39FQfaYhAKXwkO6xMz5STZfl95lJFWQDT7IOH2pa9w$>, it does not mention protobuf format. Does Flink SQL support protobuf format? If not, is there any plan to support it in the near future? Thanks!