Hi,

Flink works very well with Kafka if you wish to stream data. Following  is how 
I am streaming data with Kafka and Flink.

FlinkKafkaConsumer08<Event> kafkaConsumer = new 
FlinkKafkaConsumer08<>(KAFKA_AVRO_TOPIC, avroSchema, properties);
DataStream<Event> messageStream = env.addSource(kafkaConsumer);

Is there a way to do a micro batch operation on the data coming from Kafka? 
What I want to do is to reduce or aggregate the events coming from Kafka. For 
instance I am getting 40000 events per second from Kafka and what I want is to 
group 2000 events into one and send it to my microservice for further 
processing. Can I use the Flink DataSet API for this or should I go with Spark 
or some other framework?

Thanks & Regards
Zeeshan Alam

Reply via email to