Hi all,
Thanks for the information. I am running Spark streaming in a yarn cluster
and the configuration should be correct. I followed the KafkaWordCount to
write the current code three months ago. It has been working for several
months. The messages are in json format. Actually, this code worked a few
days ago. But now it is not working. Below please find my spark submit
script:
SPARK_BIN=/home/hadoop/spark/bin/
$SPARK_BIN/spark-submit \
--class com.test \
--master yarn-cluster \
--deploy-mode cluster \
--verbose \
--driver-memory 20G \
--executor-memory 20G \
--executor-cores 6 \
--num-executors $2 \
$1 $3 $4 $5
Thanks!
Bill
On Wed, Nov 12, 2014 at 4:53 PM, Shao, Saisai <[email protected]> wrote:
> Did you configure Spark master as local, it should be local[n], n > 1
> for local mode. Beside there’s a Kafka wordcount example in Spark Streaming
> example, you can try that. I’ve tested with latest master, it’s OK.
>
>
>
> Thanks
>
> Jerry
>
>
>
> *From:* Tobias Pfeiffer [mailto:[email protected]]
> *Sent:* Thursday, November 13, 2014 8:45 AM
> *To:* Bill Jay
> *Cc:* [email protected]
> *Subject:* Re: Spark streaming cannot receive any message from Kafka
>
>
>
> Bill,
>
>
>
> However, when I am currently using Spark 1.1.0. the Spark streaming job
> cannot receive any messages from Kafka. I have not made any change to the
> code.
>
>
>
> Do you see any suspicious messages in the log output?
>
>
>
> Tobias
>
>
>