You are viewing a plain text version of this content. The canonical link for it is here.
Posted to user@spark.apache.org by Suhas Satish <su...@gmail.com> on 2014/02/07 03:20:04 UTC

spark 0.9.0 compatible with hadoop 1.0.4 ?

Is spark compatible with Hadoop 1.0.4 ?

I built spark with the variable -
SPARK_HADOOP_VERSION=1.2.1 sbt/sbt assembly

But I got the following jar -
spark-assembly-0.9.0-incubating-hadoop1.0.4.jar

When I ran a spark job, I'm  running  into this issue-

 ./pyspark ~/test/test_spark/spark.script
Found multiple Spark assembly jars in
/home/suhash/git_repos/spark-apache-v0.9.0-incubating/assembly/target/scala-2.10:
Traceback (most recent call last):
  File "/home/ssatish/test/test_spark/spark.script", line 2, in <module>
    sc=SparkContext('local', 'App Name')
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 83, in __init__
    SparkContext._ensure_initialized(self)
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
line 159, in _ensure_initialized
    SparkContext._gateway = launch_gateway()
  File
"/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/java_gateway.py",
line 46, in launch_gateway
    port = int(proc.stdout.readline())
ValueError: invalid literal for int() with base 10:
'spark-assembly-0.9.0-incubating-hadoop1.0.4.jar\n'
Exception AttributeError: "'SparkContext' object has no attribute '_jsc'"
in <bound method SparkContext.__del__ of <pyspark.context.SparkContext
object at 0x15eaa90>> ignored




Thanks,
Suhas.

Re: spark 0.9.0 compatible with hadoop 1.0.4 ?

Posted by Patrick Wendell <pw...@gmail.com>.
It looks like you have multiple copies of the assembly jar (see the
error at the top). Try doing a clean before you build

sbt/sbt clean assembly

On Thu, Feb 6, 2014 at 6:20 PM, Suhas Satish <su...@gmail.com> wrote:
> Is spark compatible with Hadoop 1.0.4 ?
>
> I built spark with the variable -
> SPARK_HADOOP_VERSION=1.2.1 sbt/sbt assembly
>
> But I got the following jar -
> spark-assembly-0.9.0-incubating-hadoop1.0.4.jar
>
> When I ran a spark job, I'm  running  into this issue-
>
>  ./pyspark ~/test/test_spark/spark.script
> Found multiple Spark assembly jars in
> /home/suhash/git_repos/spark-apache-v0.9.0-incubating/assembly/target/scala-2.10:
> Traceback (most recent call last):
>   File "/home/ssatish/test/test_spark/spark.script", line 2, in <module>
>     sc=SparkContext('local', 'App Name')
>   File
> "/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
> line 83, in __init__
>     SparkContext._ensure_initialized(self)
>   File
> "/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/context.py",
> line 159, in _ensure_initialized
>     SparkContext._gateway = launch_gateway()
>   File
> "/home/suhash/git_repos/spark-apache-v0.9.0-incubating/python/pyspark/java_gateway.py",
> line 46, in launch_gateway
>     port = int(proc.stdout.readline())
> ValueError: invalid literal for int() with base 10:
> 'spark-assembly-0.9.0-incubating-hadoop1.0.4.jar\n'
> Exception AttributeError: "'SparkContext' object has no attribute '_jsc'" in
> <bound method SparkContext.__del__ of <pyspark.context.SparkContext object
> at 0x15eaa90>> ignored
>
>
>
>
> Thanks,
> Suhas.
>