Hi,

I'm trying to run a phoenix spark job via spark cluster mode to a remote
yarn cluster.

When I do a spark-submit, all jars under SPARK_HOME gets uploaded.

I also need to point the remote hbase jar folder location and other
dependencies for running the job.

Going through the docs, I see setting spark.yarn.jars is one way of doing
it.

For spark.executor.extraClassPath, I see

"Extra classpath entries to prepend to the classpath of executors. This
exists primarily for backwards-compatibility with older versions of
Spark. *Users
typically should not need to set this option*."

Similarly, spark.driver.extraClassPath works only for spark driver.


For running the above job, I need all executors to know hbase and other
dependencies path. So, is setting spark.yarn.jars the preferred way for
submitting spark jobs remotely? Appreciate any input. Thanks!

Reply via email to