You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Andrew Or (JIRA)" <ji...@apache.org> on 2015/05/08 04:49:59 UTC

[jira] [Updated] (SPARK-7470) Spark shell not having hive crashes SQLContext

     [ https://issues.apache.org/jira/browse/SPARK-7470?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Andrew Or updated SPARK-7470:
-----------------------------
    Priority: Major  (was: Critical)

> Spark shell not having hive crashes SQLContext
> ----------------------------------------------
>
>                 Key: SPARK-7470
>                 URL: https://issues.apache.org/jira/browse/SPARK-7470
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Shell, SQL
>    Affects Versions: 1.3.0
>            Reporter: Andrew Or
>            Assignee: Andrew Or
>
> If hive is not found on my class path, I get the following exception and I don't get to use the SQLContext anymore. In fact, we already catch `java.lang.ClassNotFoundException` in case this happens. We just don't also catch `java.lang.NoClassDefFoundError`.
> {code}
> 15/05/07 17:07:30 INFO BlockManagerMaster: Registered BlockManager
> 15/05/07 17:07:30 INFO EventLoggingListener: Logging events to file:/tmp/spark-events/local-1431043649919
> 15/05/07 17:07:30 INFO SparkILoop: Created spark context..
> Spark context available as sc.
> java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
> 	at java.lang.Class.getDeclaredConstructors0(Native Method)
> 	at java.lang.Class.privateGetDeclaredConstructors(Class.java:2493)
> 	at java.lang.Class.getConstructor0(Class.java:2803)
> 	at java.lang.Class.getConstructor(Class.java:1718)
> 	at org.apache.spark.repl.SparkILoop.createSQLContext(SparkILoop.scala:1026)
> 	at $iwC$$iwC.<init>(<console>:9)
> 	at $iwC.<init>(<console>:18)
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org