You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Rick Bross (JIRA)" <ji...@apache.org> on 2016/06/22 16:37:58 UTC

[jira] [Issue Comment Deleted] (SPARK-15221) error: not found: value sqlContext when starting Spark 1.6.1

     [ https://issues.apache.org/jira/browse/SPARK-15221?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Rick Bross updated SPARK-15221:
-------------------------------
    Comment: was deleted

(was: Actually the error the first time you run is:

java.lang.RuntimeException: java.io.IOException: Filesystem closed
 
And it's the same with the package for Hadoop 1.x

The *second* time through it's a different error, the:

java.lang.RuntimeException: java.lang.RuntimeException: The root scratch dir: /tmp/hive on HDFS should be writable. Current permissions are: rwx--x--x

I tried to create the path "tmp/hive" under /mnt/ephemeral-hdfs and chmod to 777.  No joy.
)

> error: not found: value sqlContext when starting Spark 1.6.1
> ------------------------------------------------------------
>
>                 Key: SPARK-15221
>                 URL: https://issues.apache.org/jira/browse/SPARK-15221
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.6.1
>         Environment: Ubuntu 14.0.4, 8 GB RAM, 1 Processor
>            Reporter: Vijay Parmar
>            Priority: Blocker
>              Labels: build, newbie
>
> When I start Spark (version 1.6.1), at the very end I am getting the following error message:
> <console>:16: error: not found: value sqlContext
>          import sqlContext.implicits._
>                 ^
> <console>:16: error: not found: value sqlContext
>          import sqlContext.sql
> I have gone through some content on the web about editing the /.bashrc file and including the "SPARK_LOCAL_IP=127.0.0.1" under SPARK variables. 
> Also tried editing the /etc/hosts file with :-
>  $ sudo vi /etc/hosts
>  ...
>  127.0.0.1  <HOSTNAME>
>  ...
> but still the issue persists.  Is it the issue with the build or something else?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org