Hi Prashant,

I would like to do it so that I can monitor the consumer group along with
my other consumer groups.

Thanks,
Megh

On Mon, May 19, 2025, 18:21 Prashant Sharma <scrapco...@gmail.com> wrote:

> Spark does not rely on Kafka's commit, in fact it tracks the stream
> progress itself and reads via offsets (e.g. from last read points). Why do
> you want to commit?
>
> On Mon, May 19, 2025 at 5:58 PM megh vidani <vidanimeg...@gmail.com>
> wrote:
>
>> Hello Spark Dev Community,
>>
>> Reaching out for the below problem statement.
>>
>> Thanks,
>> Megh
>>
>> On Mon, May 19, 2025, 13:16 megh vidani <vidanimeg...@gmail.com> wrote:
>>
>>> Hello Spark Community, I have a structured streaming job in which I'm
>>> consuming a topic with the same name in two different kafka clusters and
>>> then creating a union of these two streams. I've developed a custom
>>> query listener to commit the offsets back to the kafka clusters once every
>>> batch is completed in the onQueryProgress method. Now here's my
>>> problem: To commit these offsets back to Kafka, I need to iterate over the
>>> event.progress.sources list to fetch the individual source (2 in my case)
>>> and call the endOffsets method to fetch the latest offsets that have been
>>> consumed by the Query. As per the current structure of the source object,
>>> kafka bootstrap server info is not stored and it only stores the topic
>>> name. Since the topic name is same in both the streams, I can't determine
>>> which source belongs to cluster1 and which belongs to cluster2. This is
>>> basically leading to incorrect offsets getting committed to incorrect
>>> cluster. Has anyone encountered this issue before and have any solution
>>> to this? I tried to find a way to pass some custom metadata to the
>>> source object but there doesn't seem to be any. If anyone has any
>>> inputs please let me know. Thanks in advance. Megh
>>>
>>

Reply via email to