You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@kylin.apache.org by "Zhichao Zhang (Jira)" <ji...@apache.org> on 2020/05/13 10:22:00 UTC
[jira] [Created] (KYLIN-4493) On HDP3 using spark to build
dimension dictionary, it throws NoSuchMethodError.
Zhichao Zhang created KYLIN-4493:
-------------------------------------
Summary: On HDP3 using spark to build dimension dictionary, it throws NoSuchMethodError.
Key: KYLIN-4493
URL: https://issues.apache.org/jira/browse/KYLIN-4493
Project: Kylin
Issue Type: Bug
Components: Spark Engine
Affects Versions: v3.0.1, v3.0.2
Reporter: Zhichao Zhang
Test env:
{code:java}
HDP 3.0.1.0-187
Hadoop 3.1.1
Hive 3.1.0
Kylin 3.0.2
Spark 2.3.2{code}
Problems:
when use spark engine and turn on *'kylin.engine.spark-dimension-dictionary'*, it will throw error:
{code:java}
java.lang.NoSuchMethodError: org.apache.hadoop.hive.metastore.HiveMetaStoreClient.<init>(Lorg/apache/hadoop/conf/Configuration;)V
at org.apache.kylin.source.hive.CLIHiveClient.getMetaStoreClient(CLIHiveClient.java:164)
at org.apache.kylin.source.hive.CLIHiveClient.getHiveTableMeta(CLIHiveClient.java:78)
at org.apache.kylin.source.hive.HiveTable.<init>(HiveTable.java:48)
at org.apache.kylin.source.hive.HiveSource.createReadableTable(HiveSource.java:68)
at org.apache.kylin.source.SourceManager.createReadableTable(SourceManager.java:145)
at org.apache.kylin.engine.spark.SparkBuildDictionary$SnapshotBuildFunction.buildSnapshotTable(SparkBuildDictionary.java:386)
at org.apache.kylin.engine.spark.SparkBuildDictionary$SnapshotBuildFunction.call(SparkBuildDictionary.java:367)
at org.apache.kylin.engine.spark.SparkBuildDictionary$SnapshotBuildFunction.call(SparkBuildDictionary.java:325)
at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)
at org.apache.spark.api.java.JavaPairRDD$$anonfun$pairFunToScalaFun$1.apply(JavaPairRDD.scala:1043)
at scala.collection.Iterator$$anon$11.next(Iterator.scala:409)
at scala.collection.Iterator$$anon$13.hasNext(Iterator.scala:462)
at scala.collection.Iterator$class.foreach(Iterator.scala:893)
at scala.collection.AbstractIterator.foreach(Iterator.scala:1336)
at scala.collection.generic.Growable$class.$plus$plus$eq(Growable.scala:59)
at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:104)
at scala.collection.mutable.ArrayBuffer.$plus$plus$eq(ArrayBuffer.scala:48)
at scala.collection.TraversableOnce$class.to(TraversableOnce.scala:310)
at scala.collection.AbstractIterator.to(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.toBuffer(TraversableOnce.scala:302)
at scala.collection.AbstractIterator.toBuffer(Iterator.scala:1336)
at scala.collection.TraversableOnce$class.toArray(TraversableOnce.scala:289)
at scala.collection.AbstractIterator.toArray(Iterator.scala:1336)
at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:945)
at org.apache.spark.rdd.RDD$$anonfun$collect$1$$anonfun$12.apply(RDD.scala:945)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
at org.apache.spark.SparkContext$$anonfun$runJob$5.apply(SparkContext.scala:2074)
{code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)