We have streaming jobs and batch jobs running inside the docker containers
with spark driver launched within the container
Now when we open the Resource manager UI http://:8080, and try to
access the application tracking URL of any running job, the page times out
with error:
HTTP ERROR 500
Probl
Hi,
I produced avro data to kafka topic using schema registry and now I want to
use spark streaming to read that data and do some computation in real time.
Can some one please give a sample code for doing that . I couldn't find any
working code online. I am using spark version 2.2.0 and
spark-stre
Hi,
I have a use case for which I want to override the default hdfs replication
factor from my spark code. For this I have set the hadoop replication like
this:
val sc = new SparkContext(conf)
sc.hadoopConfiguration.set('dfs.replication','1').
Now my spark job runs as a cron job in some specific