You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Andrew Ash (JIRA)" <ji...@apache.org> on 2014/11/14 10:59:33 UTC

[jira] [Commented] (SPARK-904) Not able to Start/Stop Spark Worker from Remote Machine

    [ https://issues.apache.org/jira/browse/SPARK-904?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14212078#comment-14212078 ] 

Andrew Ash commented on SPARK-904:
----------------------------------

[~ayushmishra2005] I suspect you don't have Spark installed on the remote machine -- the {{start-all.sh}} script won't install it for you on remote machines.

If you're still having trouble, please reach out to the spark users list from http://spark.apache.org/community.html which is a better place for these kinds of requests anyway.  I'm closing this issue for now but let me know here if you aren't able to get a resolution on the mailing lists.

Thanks, and good luck with Spark!
Andrew

> Not able to Start/Stop Spark Worker from Remote Machine
> -------------------------------------------------------
>
>                 Key: SPARK-904
>                 URL: https://issues.apache.org/jira/browse/SPARK-904
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 0.7.3
>            Reporter: Ayush
>
> I have two machines A and B. I am trying to run Spark Master on machine A and Spark Worker on machine B. 
> I have set machine B'host name in conf/slaves in my Spark directory. 
> When I am executing start-all.sh to start master and workers, I am getting below message on console:
> abc@abc-vostro:~/spark-scala-2.10$ sudo sh bin/start-all.sh 
> sudo: /etc/sudoers.d is world writable
> starting spark.deploy.master.Master, logging to /home/abc/spark-scala-2.10/bin/../logs/spark-root-spark.deploy.master.Master-1-abc-vostro.out
> 13/09/11 14:54:29 WARN spark.Utils: Your hostname, abc-vostro resolves to a loopback address: 127.0.1.1; using 1XY.1XY.Y.Y instead (on interface wlan2)
> 13/09/11 14:54:29 WARN spark.Utils: Set SPARK_LOCAL_IP if you need to bind to another address
> Master IP: abc-vostro
> cd /home/abc/spark-scala-2.10/bin/.. ; /home/abc/spark-scala-2.10/bin/start-slave.sh 1 spark://abc-vostro:7077
> xyz@1XX.1XX.X.X's password: 
> xyz@1XX.1XX.X.X: bash: line 0: cd: /home/abc/spark-scala-2.10/bin/..: No such file or directory
> xyz@1XX.1XX.X.X: bash: /home/abc/spark-scala-2.10/bin/start-slave.sh: No such file or directory
> Master is started but worker is failed to start. 
> I have set xyz@1XX.1XX.X.X in conf/slaves in my Spark directory. 
> Can anyone help me to resolve this? This is probably something I'm missing any configuration on my end. 
> However When I create Spark Master and Worker on same machine, It is working fine. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org