Hi, I am running a Streaming pipeline(written in Beam Framework) with Flink. *Operator sequence* is -> Reading the JSON data, Parse JSON String to the Object and Group the object based on common key. I noticed that GroupByKey operator throws away some data in between and hence I don't get all the keys as output.
In the below screenshot, 1001 records are read from kafka Topic , each record has unique ID . After grouping it returns only 857 unique IDs. Ideally it should return 1001 records from GroupByKey operator. [image: Inline image 3] Any idea, what can be the issue? Thanks in advance! -- Thanks & Regards, Nishu Tayal