You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sean Owen (JIRA)" <ji...@apache.org> on 2014/11/08 10:55:33 UTC

[jira] [Resolved] (SPARK-905) Not able to run Job on remote machine

     [ https://issues.apache.org/jira/browse/SPARK-905?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Sean Owen resolved SPARK-905.
-----------------------------
    Resolution: Cannot Reproduce

This looks like something that was either long since fixed, or just a matter of not having the Spark installation set up on each machine. The compute-classpath.sh script does exist in bin/ in the tree and distro.

> Not able to run Job on  remote machine
> --------------------------------------
>
>                 Key: SPARK-905
>                 URL: https://issues.apache.org/jira/browse/SPARK-905
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 0.7.3
>            Reporter: Ayush
>
> I have two machines A and B.  On machine A, I run
>  ./run spark.deploy.master.Master 
> to start Master.
> Master URL is spark://abc-vostro.local:7077. 
> Now on machine B, I run 
> ./run spark.deploy.worker.Worker spark://abc-vostro.local:7077
> Now worker has been registered to  master. 
> Now I want to run a simple job on cluster. 
> Here is SimpleJob.scala
> package spark.examples
>  
> import spark.SparkContext
> import SparkContext._
>  
> object SimpleJob {
>   def main(args: Array[String]) {
>     val logFile = "s3n://<AWS_ACCESS_KEY_ID>:<AWS_SECRET_ACCESS_KEY>/<File Name>"
>     val sc = new SparkContext("spark://abc-vostro.local:7077", "Simple Job",
>       System.getenv("SPARK_HOME"), Seq("/home/abc/spark-scala-2.10/examples/target/scala-2.10/spark-examples_2.10-0.8.0-SNAPSHOT.jar"))
>     val logData = sc.textFile(logFile)
>     val numsa = logData.filter(line => line.contains("a")).count
>     val numsb = logData.filter(line => line.contains("b")).count
>     println("total a : %s, total b : %s".format(numsa, numsb))
>   }
> }  
> This file is located at "/home/abc/spark-scala-2.10/examples/src/main/scala/spark/examples" on machine A.
> Now on machine A, I run sbt/sbt package.
> When I run
>  MASTER=spark://abc-vostro.local:7077 ./run spark.examples.SimpleJob
> to run my job, I am getting below exception on both machines A and B,
> (class java.io.IOException: Cannot run program "/home/abc/spark-scala-2.10/bin/compute-classpath.sh" (in directory "."): error=2, No such file or directory)
> Could you please help me to resolve this? This is probably something I'm missing any configuration on my end. 
> Thanks in advance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org