You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Patrick Wendell (JIRA)" <ji...@apache.org> on 2015/04/21 19:50:59 UTC

[jira] [Resolved] (SPARK-7030) spark-submit in cluster mode with supervisor flag, not able to transfer application jar to driver node

     [ https://issues.apache.org/jira/browse/SPARK-7030?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Patrick Wendell resolved SPARK-7030.
------------------------------------
    Resolution: Invalid

This is filed as a bug report, but this is not currently a feature (in Standalone mode we require that the jar be visible everywhere, we don't copy it). We can create a new feature JIRA for adding that, or there may already be one [~bullet650] it would be good to look.

> spark-submit in cluster mode with supervisor flag, not able to transfer application jar to driver node
> ------------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-7030
>                 URL: https://issues.apache.org/jira/browse/SPARK-7030
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core, Spark Submit
>    Affects Versions: 1.3.0
>            Reporter: pankaj
>
> Hi,
> I have 3 node cluster node A,B,C. running master on A ,B and slave on A.B and C. 
> while i run spark-submit from node A using below command.
> /usr/local/spark-1.3.0-bin-hadoop2.4/bin/spark-submit   --class com.test.SparkExample --deploy-mode cluster --supervise --master spark://master.com:7077  file:///home/spark/sparkstreaming-0.0.1-SNAPSHOT.jar
> the driver gets launched to node B and it tries to find application jar at local file system at node B. do we need to transfer application jar on each master node manually.  is the the known bug ? or i am missing something.
> kindly suggest
> Thanks
> Pankaj



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org