Isn't there a way to do it with kafka connector instead of kafka client? Isn't there any way to throttle kafka connector? Seems like a common problem.
Regards, Abhishek Singla On Mon, Feb 24, 2025 at 7:24 PM daniel williams <daniel.willi...@gmail.com> wrote: > I think you should be using a foreachPartition and a broadcast to build > your producer. From there you will have full control of all options and > serialization needed via direct access to the KafkaProducer, as well as all > options therein associated (e.g. callbacks, interceptors, etc). > > -dan > > > On Mon, Feb 24, 2025 at 6:26 AM Abhishek Singla < > abhisheksingla...@gmail.com> wrote: > >> Hi Team, >> >> I am using spark to read from S3 and write to Kafka. >> >> Spark Version: 3.1.2 >> Scala Version: 2.12 >> Spark Kafka connector: spark-sql-kafka-0-10_2.12 >> >> I want to throttle kafka producer. I tried using *linger.ms >> <http://linger.ms>* and *batch.size* config but I can see in *ProducerConfig: >> ProducerConfig values* at runtime that they are not being set. Is there >> something I am missing? Is there any other way to throttle kafka writes? >> >> *dataset.write().format("kafka").options(options).save();* >> >> Regards, >> Abhishek Singla >> >> >> >> >>