You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Yin Huai (JIRA)" <ji...@apache.org> on 2016/08/04 02:53:20 UTC

[jira] [Assigned] (SPARK-16887) Add SPARK_DIST_CLASSPATH to LAUNCH_CLASSPATH

     [ https://issues.apache.org/jira/browse/SPARK-16887?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Yin Huai reassigned SPARK-16887:
--------------------------------

    Assignee: Yin Huai

> Add SPARK_DIST_CLASSPATH to LAUNCH_CLASSPATH
> --------------------------------------------
>
>                 Key: SPARK-16887
>                 URL: https://issues.apache.org/jira/browse/SPARK-16887
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Submit
>            Reporter: Yin Huai
>            Assignee: Yin Huai
>
> To deploy Spark, it can be pretty convenient to put all jars (spark jars, hadoop jars, and other libs' jars) that we want to include in the classpath of Spark in the same dir, which may not be spark's assembly dir. So, I am proposing to also add SPARK_DIST_CLASSPATH to the LAUNCH_CLASSPATH.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org