You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Sunita Arvind <su...@gmail.com> on 2014/11/26 07:54:24 UTC
Spark setup on local windows machine
Hi All,
I just installed a spark on my laptop and trying to get spark-shell to
work. Here is the error I see:
C:\spark\bin>spark-shell
Exception in thread "main" java.util.NoSuchElementException: key not found:
CLAS
SPATH
at scala.collection.MapLike$class.default(MapLike.scala:228)
at scala.collection.AbstractMap.default(Map.scala:58)
at scala.collection.MapLike$class.apply(MapLike.scala:141)
at scala.collection.AbstractMap.apply(Map.scala:58)
at
org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm
itDriverBootstrapper.scala:49)
at
org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi
tDriverBootstrapper.scala)
The classpath seems to be right:
C:\spark\bin>compute-classpath.cmd
;;C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar;
;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucle
us-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar
Manually exporting the classpath to include the assembly jar doesnt help
either.
What could be wrong with this installation? Scala and SBT are installed, in
path and are working fine.
Appreciate your help.
regards
Sunita
Re: Spark setup on local windows machine
Posted by Sunita Arvind <su...@gmail.com>.
Thanks Sameer and Akhil for your help.
I tried both your suggestions however, I still face the same issue. There
was indeed space in the installation path for Scala and Sbt since I had let
the defaults stay and hence the path was C:\Program Files.... . I
reinstalled scala and sbt in c:\ as well as spark.
The spark binaries I am using are built internally from source by our team
and the same binary works for rest of the team.
Here is what the spark-env.cmd looks like for me:
set SCALA_HOME=C:\scala
set
SPARK_CLASSPATH=C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucleus-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar;C:\scala\lib\scala-library.jar;C:\scala\lib\scala-compiler.jar;
I get the same error inspite of this. The compute-classpath.cmd yields
correct results:
C:\spark>bin\spark-shell
Exception in thread "main" java.util.NoSuchElementException: key not found:
CLAS
SPATH
at scala.collection.MapLike$class.default(MapLike.scala:228)
at scala.collection.AbstractMap.default(Map.scala:58)
at scala.collection.MapLike$class.apply(MapLike.scala:141)
at scala.collection.AbstractMap.apply(Map.scala:58)
at
org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm
itDriverBootstrapper.scala:49)
at
org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi
tDriverBootstrapper.scala)
regards
Sunita
On Tue, Nov 25, 2014 at 11:47 PM, Sameer Farooqui <sa...@databricks.com>
wrote:
> Hi Sunita,
>
> This gitbook may also be useful for you to get Spark running in local mode
> on your Windows machine:
> http://blueplastic.gitbooks.io/how-to-light-your-spark-on-a-stick/content/
>
> On Tue, Nov 25, 2014 at 11:09 PM, Akhil Das <ak...@sigmoidanalytics.com>
> wrote:
>
>> You could try following this guidelines
>> http://docs.sigmoidanalytics.com/index.php/How_to_build_SPARK_on_Windows
>>
>> Thanks
>> Best Regards
>>
>> On Wed, Nov 26, 2014 at 12:24 PM, Sunita Arvind <su...@gmail.com>
>> wrote:
>>
>>> Hi All,
>>>
>>> I just installed a spark on my laptop and trying to get spark-shell to
>>> work. Here is the error I see:
>>>
>>> C:\spark\bin>spark-shell
>>> Exception in thread "main" java.util.NoSuchElementException: key not
>>> found: CLAS
>>> SPATH
>>> at scala.collection.MapLike$class.default(MapLike.scala:228)
>>> at scala.collection.AbstractMap.default(Map.scala:58)
>>> at scala.collection.MapLike$class.apply(MapLike.scala:141)
>>> at scala.collection.AbstractMap.apply(Map.scala:58)
>>> at
>>> org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm
>>> itDriverBootstrapper.scala:49)
>>> at
>>> org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi
>>> tDriverBootstrapper.scala)
>>>
>>>
>>> The classpath seems to be right:
>>>
>>> C:\spark\bin>compute-classpath.cmd
>>>
>>> ;;C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar;
>>>
>>> ;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucle
>>> us-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar
>>>
>>> Manually exporting the classpath to include the assembly jar doesnt help
>>> either.
>>>
>>> What could be wrong with this installation? Scala and SBT are installed,
>>> in path and are working fine.
>>>
>>> Appreciate your help.
>>> regards
>>> Sunita
>>>
>>>
>>>
>>
>
Re: Spark setup on local windows machine
Posted by Sameer Farooqui <sa...@databricks.com>.
Hi Sunita,
This gitbook may also be useful for you to get Spark running in local mode
on your Windows machine:
http://blueplastic.gitbooks.io/how-to-light-your-spark-on-a-stick/content/
On Tue, Nov 25, 2014 at 11:09 PM, Akhil Das <ak...@sigmoidanalytics.com>
wrote:
> You could try following this guidelines
> http://docs.sigmoidanalytics.com/index.php/How_to_build_SPARK_on_Windows
>
> Thanks
> Best Regards
>
> On Wed, Nov 26, 2014 at 12:24 PM, Sunita Arvind <su...@gmail.com>
> wrote:
>
>> Hi All,
>>
>> I just installed a spark on my laptop and trying to get spark-shell to
>> work. Here is the error I see:
>>
>> C:\spark\bin>spark-shell
>> Exception in thread "main" java.util.NoSuchElementException: key not
>> found: CLAS
>> SPATH
>> at scala.collection.MapLike$class.default(MapLike.scala:228)
>> at scala.collection.AbstractMap.default(Map.scala:58)
>> at scala.collection.MapLike$class.apply(MapLike.scala:141)
>> at scala.collection.AbstractMap.apply(Map.scala:58)
>> at
>> org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm
>> itDriverBootstrapper.scala:49)
>> at
>> org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi
>> tDriverBootstrapper.scala)
>>
>>
>> The classpath seems to be right:
>>
>> C:\spark\bin>compute-classpath.cmd
>>
>> ;;C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar;
>>
>> ;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucle
>> us-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar
>>
>> Manually exporting the classpath to include the assembly jar doesnt help
>> either.
>>
>> What could be wrong with this installation? Scala and SBT are installed,
>> in path and are working fine.
>>
>> Appreciate your help.
>> regards
>> Sunita
>>
>>
>>
>
Re: Spark setup on local windows machine
Posted by Akhil Das <ak...@sigmoidanalytics.com>.
You could try following this guidelines
http://docs.sigmoidanalytics.com/index.php/How_to_build_SPARK_on_Windows
Thanks
Best Regards
On Wed, Nov 26, 2014 at 12:24 PM, Sunita Arvind <su...@gmail.com>
wrote:
> Hi All,
>
> I just installed a spark on my laptop and trying to get spark-shell to
> work. Here is the error I see:
>
> C:\spark\bin>spark-shell
> Exception in thread "main" java.util.NoSuchElementException: key not
> found: CLAS
> SPATH
> at scala.collection.MapLike$class.default(MapLike.scala:228)
> at scala.collection.AbstractMap.default(Map.scala:58)
> at scala.collection.MapLike$class.apply(MapLike.scala:141)
> at scala.collection.AbstractMap.apply(Map.scala:58)
> at
> org.apache.spark.deploy.SparkSubmitDriverBootstrapper$.main(SparkSubm
> itDriverBootstrapper.scala:49)
> at
> org.apache.spark.deploy.SparkSubmitDriverBootstrapper.main(SparkSubmi
> tDriverBootstrapper.scala)
>
>
> The classpath seems to be right:
>
> C:\spark\bin>compute-classpath.cmd
>
> ;;C:\spark\bin\..\conf;C:\spark\bin\..\lib\spark-assembly-1.1.0-hadoop2.3.0.jar;
>
> ;C:\spark\bin\..\lib\datanucleus-api-jdo-3.2.1.jar;C:\spark\bin\..\lib\datanucle
> us-core-3.2.2.jar;C:\spark\bin\..\lib\datanucleus-rdbms-3.2.1.jar
>
> Manually exporting the classpath to include the assembly jar doesnt help
> either.
>
> What could be wrong with this installation? Scala and SBT are installed,
> in path and are working fine.
>
> Appreciate your help.
> regards
> Sunita
>
>
>