You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Josiah Samuel Sathiadass (JIRA)" <ji...@apache.org> on 2015/06/17 15:31:00 UTC
[jira] [Created] (SPARK-8410) Hive VersionsSuite RuntimeException
Josiah Samuel Sathiadass created SPARK-8410:
-----------------------------------------------
Summary: Hive VersionsSuite RuntimeException
Key: SPARK-8410
URL: https://issues.apache.org/jira/browse/SPARK-8410
Project: Spark
Issue Type: Question
Components: SQL
Affects Versions: 1.4.0, 1.3.1
Environment: IBM Power system - P7
running Ubuntu 14.04LE
with IBM JDK version 1.7.0
Reporter: Josiah Samuel Sathiadass
Priority: Minor
While testing Spark Project Hive, there are RuntimeExceptions as follows,
VersionsSuite:
- success sanity check *** FAILED ***
java.lang.RuntimeException: [download failed: org.jboss.netty#netty;3.2.2.Final!netty.jar(bundle), download failed: org.codehaus.groovy#groovy-all;2.1.6!groovy-all.jar, download failed: asm#asm;3.2!asm.jar]
at org.apache.spark.deploy.SparkSubmitUtils$.resolveMavenCoordinates(SparkSubmit.scala:978)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$$anonfun$3.apply(IsolatedClientLoader.scala:62)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$$anonfun$3.apply(IsolatedClientLoader.scala:62)
at org.apache.spark.sql.catalyst.util.package$.quietly(package.scala:38)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$.org$apache$spark$sql$hive$client$IsolatedClientLoader$$downloadVersion(IsolatedClientLoader.scala:61)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$$anonfun$1.apply(IsolatedClientLoader.scala:44)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$$anonfun$1.apply(IsolatedClientLoader.scala:44)
at scala.collection.mutable.MapLike$class.getOrElseUpdate(MapLike.scala:189)
at scala.collection.mutable.AbstractMap.getOrElseUpdate(Map.scala:91)
at org.apache.spark.sql.hive.client.IsolatedClientLoader$.forVersion(IsolatedClientLoader.scala:44)
...
The tests are executed with the following set of options,
build/mvn --pl sql/hive --fail-never -Pyarn -Phadoop-2.4 -Dhadoop.version=2.6.0 test
Adding the following dependencies in the "spark/sql/hive/pom.xml" file solves this issue,
< <dependency>
< <groupId>org.jboss.netty</groupId>
< <artifactId>netty</artifactId>
< <version>3.2.2.Final</version>
< <scope>test</scope>
< </dependency>
< <dependency>
< <groupId>org.codehaus.groovy</groupId>
< <artifactId>groovy-all</artifactId>
< <version>2.1.6</version>
< <scope>test</scope>
< </dependency>
<
< <dependency>
< <groupId>asm</groupId>
< <artifactId>asm</artifactId>
< <version>3.2</version>
< <scope>test</scope>
< </dependency>
<
The question is, Is this the correct way to fix this runtimeException ?
If yes, Can a pull request fix this issue permanently ?
If not, suggestions please.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org