You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "DeepakVohra (JIRA)" <ji...@apache.org> on 2015/01/30 01:06:34 UTC

[jira] [Commented] (SPARK-5489) KMeans clustering java.lang.NoSuchMethodError: scala.runtime.IntRef.create (I)Lscala/runtime/IntRef;

    [ https://issues.apache.org/jira/browse/SPARK-5489?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14297937#comment-14297937 ] 

DeepakVohra commented on SPARK-5489:
------------------------------------

Sean,

Made the Scala version the same, but still getting the error.
"For the Scala API, Spark 1.2.0 uses Scala 2.10. "
http://spark.apache.org/docs/1.2.0/

Made Maven dependencies Scala version also 2.10.
<dependencies>
		<dependency>
			<groupId>org.apache.spark</groupId>
			<artifactId>spark-core_2.10</artifactId>
			<version>1.2.0</version>
			<exclusions>
				<exclusion>
					<groupId>org.scala-lang</groupId>
					<artifactId>scala-library</artifactId>
				</exclusion>
				<exclusion>
					<groupId>org.scala-lang</groupId>
					<artifactId>scala-compiler</artifactId>
				</exclusion>
			</exclusions>
		</dependency>

		<dependency>
			<groupId>org.apache.spark</groupId>
			<artifactId>spark-mllib_2.11</artifactId>
			<version>1.2.0</version>
			<exclusions>
				<exclusion>
					<groupId>org.scala-lang</groupId>
					<artifactId>scala-library</artifactId>
				</exclusion>
				<exclusion>
					<groupId>org.scala-lang</groupId>
					<artifactId>scala-compiler</artifactId>
				</exclusion>
			</exclusions>
		</dependency>
		<dependency>
			<groupId>org.scala-lang</groupId>
			<artifactId>scala-library</artifactId>
			<version>2.10.0</version>
		</dependency>
		
			<dependency>
			<groupId>org.scala-lang</groupId>
			<artifactId>scala-compiler</artifactId>
			<version>2.10.0</version>
		</dependency>
	
	</dependencies>

thanks,
Deepak

> KMeans clustering java.lang.NoSuchMethodError: scala.runtime.IntRef.create  (I)Lscala/runtime/IntRef;
> -----------------------------------------------------------------------------------------------------
>
>                 Key: SPARK-5489
>                 URL: https://issues.apache.org/jira/browse/SPARK-5489
>             Project: Spark
>          Issue Type: Bug
>          Components: MLlib
>    Affects Versions: 1.2.0
>         Environment: Spark 1.2 
> Maven
>            Reporter: DeepakVohra
>
> The KMeans clustering generates following error, which also seems to be due version mismatch between Scala used for compiling Spark and Scala in Spark 1.2 Maven dependency. 
> Exception in thread "main" java.lang.NoSuchMethodError: scala.runtime.IntRef.create
> (I)Lscala/runtime/IntRef;
> 	at 
> org.apache.spark.mllib.clustering.KMeans.initKMeansParallel(KMeans.scala:282)
> 	at 
> org.apache.spark.mllib.clustering.KMeans.runAlgorithm(KMeans.scala:155)
> 	at 
> org.apache.spark.mllib.clustering.KMeans.run(KMeans.scala:132)
> 	at 
> org.apache.spark.mllib.clustering.KMeans$.train(KMeans.scala:352)
> 	at 
> org.apache.spark.mllib.clustering.KMeans$.train(KMeans.scala:362)
> 	at 
> org.apache.spark.mllib.clustering.KMeans.train(KMeans.scala)
> 	at 
> clusterer.kmeans.KMeansClusterer.main(KMeansClusterer.java:35)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org