have you read
https://spark.apache.org/docs/latest/structured-streaming-kafka-integration.html
?

On Mon, Sep 17, 2018 at 4:46 AM hagersaleh <hager.saleh....@gmail.com>
wrote:

> I write code to connect kafka with spark using python and I run code on
> jupyer
> my code
> import os
> #os.environ['PYSPARK_SUBMIT_ARGS'] = '--jars
>
> /home/hadoop/Desktop/spark-program/kafka/spark-streaming-kafka-0-8-assembly_2.10-2.0.0-preview.jar
> pyspark-shell'
> os.environ['PYSPARK_SUBMIT_ARGS'] = "--packages
> org.apache.spark:spark-sql-kafka-0-10_2.11:2.3.0 pyspark-shell"
>
> os.environ['PYSPARK_SUBMIT_ARGS'] = "--packages
> org.apache.spark:spark-streaming-kafka-0-8_2.11:2.3.0 pyspark-shell"
>
> import pyspark
> from pyspark.streaming.kafka import KafkaUtils
> from pyspark.streaming import StreamingContext
>
> #sc = SparkContext()
> ssc = StreamingContext(sc,1)
>
> broker = "iotmsgs"
> directKafkaStream = KafkaUtils.createDirectStream(ssc, ["test1"],
>                         {"metadata.broker.list": broker})
> directKafkaStream.pprint()
> ssc.start()
>
> error display
> Spark Streaming's Kafka libraries not found in class path. Try one of the
> following.
>
>   1. Include the Kafka library and its dependencies with in the
>      spark-submit command as
>
>      $ bin/spark-submit --packages
> org.apache.spark:spark-streaming-kafka-0-8:2.3.0 ...
>
>   2. Download the JAR of the artifact from Maven Central
> http://search.maven.org/,
>      Group Id = org.apache.spark, Artifact Id =
> spark-streaming-kafka-0-8-assembly, Version = 2.3.0.
>      Then, include the jar in the spark-submit command as
>
>      $ bin/spark-submit --jars <spark-streaming-kafka-0-8-assembly.jar> ...
>
>
>
>
> --
> Sent from: http://apache-spark-developers-list.1001551.n3.nabble.com/
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>
>

Reply via email to