Hi folks,
I am trying to consume avro data from Kafka in Flink. The data is produced by
Kafka connect using AvroConverter. I have created a
AvroDeserializationSchema.java
<https://gist.github.com/datafibers/ae9d624b6db44865ae14defe8a838123> used by
Flink consumer. Then, I use following code to read it.
public static void main(String[] args) throws Exception {
StreamExecutionEnvironment env =
StreamExecutionEnvironment.getExecutionEnvironment();
Properties properties = new Properties();
properties.setProperty("bootstrap.servers", “localhost:9092");
properties.setProperty("zookeeper.connect", “localhost:2181”);
Schema schema = new Parser().parse("{" + "\"name\": \"test\", "
+ "\"type\": \"record\", "
+ "\"fields\": "
+" [ "
+ " { \"name\": \"name\", \"type\":
\"string\" },"
+ " { \"name\": \"symbol\",
\"type\": \"string\" },"
+ " { \"name\": \"exchange\",
\"type\": \"string\"}"
+ "] "
+"}");
AvroDeserializationSchema avroSchema = new
AvroDeserializationSchema<>(schema);
FlinkKafkaConsumer09<GenericRecord> kafkaConsumer =
new FlinkKafkaConsumer09<>("myavrotopic",avroSchema,
properties);
DataStream<GenericRecord> messageStream =
env.addSource(kafkaConsumer);
messageStream.rebalance().print();
env.execute("Flink AVRO KAFKA Test");
}
Once, I run the code, I am able to get the schema information only as follows.
{"name":"", "symbol":"", "exchange":""}
{"name":"", "symbol":"", "exchange":""}
{"name":"", "symbol":"", "exchange":""}
{"name":"", "symbol":"", "exchange":”"}
Could anyone help to find out the issues why I cannot decode it?
Further troubleshooting, I found out if I use a kafka producer here
<https://gist.github.com/datafibers/d063b255b50fa34515c0ac9e24d4485c> to send
the avro data especially using kafka.serializer.DefaultEncoder. Above code can
get correct result. Does any body know how to either set DefaultEncoder in
Kafka Connect or set it when writing customized kafka connect? Or in the other
way, how should I modify the AvroDeserializationSchema.java for instead?
Thanks, I’ll post this to the Kafka user group as well.
Will