Hi Md Aamir,
Thank you for providing the details of your streaming application setup. To assist you better in identifying and resolving the backpressure issue, I have a few follow-up questions: 1. Which version of Apache Flink are you using? 2. Is the ProcessFunction stateful? If yes, what kind of state is managed? 3. What are the configurations for the Kafka producer and consumer (e.g., fetch.size, batch.size, linger.ms, acks, compression.type etc)? 4. Have you examined the backpressure monitoring graphs in the Flink Web UI? Which operator (source, process, or sink) is experiencing the bottleneck? 5. Have you tried explicitly increasing the parallelism of the sink operator to match or exceed the source parallelism? 6. Although I’m not deeply familiar with Kerberos integration, are there any errors or warnings in the logs related to Kerberos authentication? 7. Have you verified if Kerberos ticket renewal contributes to any performance overhead for the sink operator? Bests, Samrat On Wed, Dec 18, 2024 at 3:25 PM Mohammad Aamir Iqubal <aamir.y...@gmail.com> wrote: > Hi Team, > > I am running a streaming application in performance environment. The source > is Kafka and sink is also Kafka but the sink Kafka is secured by kerberos. > Message size is 102kb and source parallelism is 16 > process parallelism is 80 and sink parallelism is 12. > I am using process function to remove few fields from json and mask some > fields. When we send the data with around 100 TPS the application goes into > backpressure. Please let me know how to fix the issue. > > > Regards, > Md Aamir Iqubal > +918409561989 >