You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Vinod KC (JIRA)" <ji...@apache.org> on 2015/05/08 19:38:01 UTC
[jira] [Created] (SPARK-7489) Spark shell crashes when compiled
with scala 2.11 and SPARK_PREPEND_CLASSES=true
Vinod KC created SPARK-7489:
-------------------------------
Summary: Spark shell crashes when compiled with scala 2.11 and SPARK_PREPEND_CLASSES=true
Key: SPARK-7489
URL: https://issues.apache.org/jira/browse/SPARK-7489
Project: Spark
Issue Type: Bug
Components: Spark Shell
Reporter: Vinod KC
Steps followed
>export SPARK_PREPEND_CLASSES=true
>dev/change-version-to-2.11.sh
> sbt/sbt -Pyarn -Phadoop-2.4 -Dscala-2.11 -DskipTests clean assembly
>bin/spark-shell
....................
15/05/08 22:31:35 INFO Main: Created spark context..
Spark context available as sc.
java.lang.NoClassDefFoundError: org/apache/hadoop/hive/conf/HiveConf
at java.lang.Class.getDeclaredConstructors0(Native Method)
at java.lang.Class.privateGetDeclaredConstructors(Class.java:2671)
at java.lang.Class.getConstructor0(Class.java:3075)
at java.lang.Class.getConstructor(Class.java:1825)
at org.apache.spark.repl.Main$.createSQLContext(Main.scala:86)
... 45 elided
Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hive.conf.HiveConf
at java.net.URLClassLoader.findClass(URLClassLoader.java:381)
at java.lang.ClassLoader.loadClass(ClassLoader.java:424)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:331)
at java.lang.ClassLoader.loadClass(ClassLoader.java:357)
... 50 more
<console>:11: error: not found: value sqlContext
import sqlContext.implicits._
^
<console>:11: error: not found: value sqlContext
import sqlContext.sql
There is a similar Resolved JIRA issue -SPARK-7470 and a PR https://github.com/apache/spark/pull/5997 , which handled same issue only in scala 2.10
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org