You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Kanagha <er...@gmail.com> on 2017/07/10 04:13:46 UTC
spark-submit via cluster mode - setting dependencies classpath!
Hi,
I'm trying to run a phoenix spark job via spark cluster mode to a remote
yarn cluster.
When I do a spark-submit, all jars under SPARK_HOME gets uploaded.
I also need to point the remote hbase jar folder location and other
dependencies for running the job.
Going through the docs, I see setting spark.yarn.jars is one way of doing
it.
For spark.executor.extraClassPath, I see
"Extra classpath entries to prepend to the classpath of executors. This
exists primarily for backwards-compatibility with older versions of
Spark. *Users
typically should not need to set this option*."
Similarly, spark.driver.extraClassPath works only for spark driver.
For running the above job, I need all executors to know hbase and other
dependencies path. So, is setting spark.yarn.jars the preferred way for
submitting spark jobs remotely? Appreciate any input. Thanks!