Hi Md Aamir,

Thank you for providing the details of your streaming application setup.

To assist you better in identifying and resolving the backpressure issue, I
have a few follow-up questions:



   1. Which version of Apache Flink are you using?
   2. Is the ProcessFunction stateful? If yes, what kind of state is
   managed?
   3. What are the configurations for the Kafka producer and consumer
   (e.g., fetch.size, batch.size, linger.ms, acks, compression.type etc)?
   4. Have you examined the backpressure monitoring graphs in the Flink Web
   UI? Which operator (source, process, or sink) is experiencing the
   bottleneck?
   5. Have you tried explicitly increasing the parallelism of the sink
   operator to match or exceed the source parallelism?
   6. Although I’m not deeply familiar with Kerberos integration, are there
   any errors or warnings in the logs related to Kerberos authentication?
   7. Have you verified if Kerberos ticket renewal contributes to any
   performance overhead for the sink operator?


Bests,
Samrat

On Wed, Dec 18, 2024 at 3:25 PM Mohammad Aamir Iqubal <aamir.y...@gmail.com>
wrote:

> Hi Team,
>
> I am running a streaming application in performance environment. The source
> is Kafka and sink is also Kafka but the sink Kafka is secured by kerberos.
> Message size is 102kb and source parallelism is 16
> process parallelism is 80 and sink parallelism is 12.
> I am using process function to remove few fields from json and mask some
> fields. When we send the data with around 100 TPS the application goes into
> backpressure. Please let me know how to fix the issue.
>
>
> Regards,
> Md Aamir Iqubal
> +918409561989
>

Reply via email to