Hi,

I think the issue is that you are overriding the kafka-clients that comes
with  <artifactId>spark-sql-kafka-0-10_2.12</artifactId>


I'd try removing the kafka-clients and see if it works


On Sun, 6 Dec 2020 at 08:01, Amit Joshi <mailtojoshia...@gmail.com> wrote:

> Hi All,
>
> I am running the Spark Structured Streaming along with Kafka.
> Below is the pom.xml
>
> <properties>
>     <maven.compiler.source>1.8</maven.compiler.source>
>     <maven.compiler.target>1.8</maven.compiler.target>
>     <encoding>UTF-8</encoding>
>     <!-- Put the Scala version of the cluster -->
>     <scalaVersion>2.12.10</scalaVersion>
>     <sparkVersion>3.0.1</sparkVersion>
> </properties>
>
> <dependency>
>     <groupId>org.apache.kafka</groupId>
>     <artifactId>kafka-clients</artifactId>
>     <version>2.1.0</version>
> </dependency>
>
> <dependency>
>     <groupId>org.apache.spark</groupId>
>     <artifactId>spark-core_2.12</artifactId>
>     <version>${sparkVersion}</version>
>     <scope>provided</scope>
> </dependency>
> <!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql -->
> <dependency>
>     <groupId>org.apache.spark</groupId>
>     <artifactId>spark-sql_2.12</artifactId>
>     <version>${sparkVersion}</version>
>     <scope>provided</scope>
> </dependency>
> <!-- https://mvnrepository.com/artifact/org.apache.spark/spark-sql-kafka-0-10 
> -->
> <dependency>
>     <groupId>org.apache.spark</groupId>
>     <artifactId>spark-sql-kafka-0-10_2.12</artifactId>
>     <version>${sparkVersion}</version>
> </dependency>
>
> Building the fat jar with shade plugin. The jar is running as expected in my 
> local setup with the command
>
> *spark-submit --master local[*] --class com.stream.Main --num-executors 3 
> --driver-memory 2g --executor-cores 2 --executor-memory 3g 
> prism-event-synch-rta.jar*
>
> But when I am trying to run same jar in spark cluster using yarn with command:
>
> *spark-submit --master yarn --deploy-mode cluster --class com.stream.Main 
> --num-executors 4 --driver-memory 2g --executor-cores 1 --executor-memory 4g  
> gs://jars/prism-event-synch-rta.jar*
>
> Getting the this exception:
>
>       
>
>
> *at org.apache.spark.sql.execution.streaming.StreamExecution.org 
> <http://org.apache.spark.sql.execution.streaming.StreamExecution.org>$apache$spark$sql$execution$streaming$StreamExecution$$runStream(StreamExecution.scala:355)
>      at 
> org.apache.spark.sql.execution.streaming.StreamExecution$$anon$1.run(StreamExecution.scala:245)Caused
>  by: org.apache.kafka.common.config.ConfigException: Missing required 
> configuration "partition.assignment.strategy" which has no default value. at 
> org.apache.kafka.common.config.ConfigDef.parse(ConfigDef.java:124)*
>
> I have tried setting up the "partition.assignment.strategy", then also its 
> not working.
>
> Please help.
>
>
> Regards
>
> Amit Joshi
>
>

Reply via email to