Hi,

I am running a Streaming pipeline(written in Beam Framework) with Flink.
*Operator sequence* is -> Reading the JSON data, Parse JSON String to the
Object and  Group the object based on common key.  I noticed that
GroupByKey operator throws away some data in between and hence I don't get
all the keys as output.

In the below screenshot, 1001 records are read from kafka Topic , each
record has unique ID .  After grouping it returns only 857 unique IDs.
Ideally it should return 1001 records from GroupByKey operator.


[image: Inline image 3]

Any idea, what can be the issue? Thanks in advance!

-- 
Thanks & Regards,
Nishu Tayal

Reply via email to