You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@iceberg.apache.org by GitBox <gi...@apache.org> on 2021/07/26 06:42:03 UTC

[GitHub] [iceberg] zhangpanbigData opened a new issue #2868: read manifest thought spark-sql throw an exception

zhangpanbigData opened a new issue #2868:
URL: https://github.com/apache/iceberg/issues/2868


   when i read iceberg table manifest managed by hive catalog,it throws a exception;my spark version is spark 3.0.3,hive version is hive 3.1.2,jdk version is 1.8.0_291,iceberg version is the the latest version;below is the exception details:
   
   spark.sql("select * from hive_catalog.icebergdb.icebergtbl.manifests").show(truncate=false)
   java.lang.ClassCastException: java.nio.HeapCharBuffer cannot be cast to java.lang.Void
     at org.apache.iceberg.transforms.VoidTransform.toHumanString(VoidTransform.java:27)
     at org.apache.iceberg.ManifestsTable.partitionSummariesToRows(ManifestsTable.java:114)
     at org.apache.iceberg.ManifestsTable.manifestFileToRow(ManifestsTable.java:97)
     at org.apache.iceberg.ManifestsTable.lambda$task$0(ManifestsTable.java:79)
     at org.apache.iceberg.relocated.com.google.common.collect.Iterators$6.transform(Iterators.java:783)
     at org.apache.iceberg.relocated.com.google.common.collect.TransformedIterator.next(TransformedIterator.java:47)
     at org.apache.iceberg.relocated.com.google.common.collect.Iterators.addAll(Iterators.java:356)
     at org.apache.iceberg.relocated.com.google.common.collect.Lists.newArrayList(Lists.java:143)
     at org.apache.iceberg.relocated.com.google.common.collect.Lists.newArrayList(Lists.java:130)
     at org.apache.iceberg.StaticDataTask.of(StaticDataTask.java:38)
     at org.apache.iceberg.ManifestsTable.task(ManifestsTable.java:76)
     at org.apache.iceberg.StaticTableScan.planFiles(StaticTableScan.java:72)
     at org.apache.iceberg.BaseTableScan.planFiles(BaseTableScan.java:208)
     at org.apache.iceberg.BaseTableScan.planTasks(BaseTableScan.java:241)
     at org.apache.iceberg.spark.source.SparkBatchQueryScan.tasks(SparkBatchQueryScan.java:122)
     at org.apache.iceberg.spark.source.SparkBatchScan.planInputPartitions(SparkBatchScan.java:143)
     at org.apache.spark.sql.execution.datasources.v2.BatchScanExec.partitions$lzycompute(BatchScanExec.scala:43)
     at org.apache.spark.sql.execution.datasources.v2.BatchScanExec.partitions(BatchScanExec.scala:43)
     at org.apache.spark.sql.execution.datasources.v2.DataSourceV2ScanExecBase.supportsColumnar(DataSourceV2ScanExecBase.scala:61)
     at org.apache.spark.sql.execution.datasources.v2.DataSourceV2ScanExecBase.supportsColumnar$(DataSourceV2ScanExecBase.scala:60)
     at org.apache.spark.sql.execution.datasources.v2.BatchScanExec.supportsColumnar(BatchScanExec.scala:29)
     at org.apache.spark.sql.execution.datasources.v2.DataSourceV2Strategy.apply(DataSourceV2Strategy.scala:84)
     at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$1(QueryPlanner.scala:63)
     at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484)
     at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490)
     at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:489)
     at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
     at org.apache.spark.sql.execution.SparkStrategies.plan(SparkStrategies.scala:68)
     at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$3(QueryPlanner.scala:78)
     at scala.collection.TraversableOnce.$anonfun$foldLeft$1(TraversableOnce.scala:162)
     at scala.collection.TraversableOnce.$anonfun$foldLeft$1$adapted(TraversableOnce.scala:162)
     at scala.collection.Iterator.foreach(Iterator.scala:941)
     at scala.collection.Iterator.foreach$(Iterator.scala:941)
     at scala.collection.AbstractIterator.foreach(Iterator.scala:1429)
     at scala.collection.TraversableOnce.foldLeft(TraversableOnce.scala:162)
     at scala.collection.TraversableOnce.foldLeft$(TraversableOnce.scala:160)
     at scala.collection.AbstractIterator.foldLeft(Iterator.scala:1429)
     at org.apache.spark.sql.catalyst.planning.QueryPlanner.$anonfun$plan$2(QueryPlanner.scala:75)
     at scala.collection.Iterator$$anon$11.nextCur(Iterator.scala:484)
     at scala.collection.Iterator$$anon$11.hasNext(Iterator.scala:490)
     at org.apache.spark.sql.catalyst.planning.QueryPlanner.plan(QueryPlanner.scala:93)
     at org.apache.spark.sql.execution.SparkStrategies.plan(SparkStrategies.scala:68)
     at org.apache.spark.sql.execution.QueryExecution$.createSparkPlan(QueryExecution.scala:338)
     at org.apache.spark.sql.execution.QueryExecution.$anonfun$sparkPlan$1(QueryExecution.scala:99)
     at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
     at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:138)
     at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:767)
     at org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:138)
     at org.apache.spark.sql.execution.QueryExecution.sparkPlan$lzycompute(QueryExecution.scala:99)
     at org.apache.spark.sql.execution.QueryExecution.sparkPlan(QueryExecution.scala:92)
     at org.apache.spark.sql.execution.QueryExecution.$anonfun$executedPlan$1(QueryExecution.scala:112)
     at org.apache.spark.sql.catalyst.QueryPlanningTracker.measurePhase(QueryPlanningTracker.scala:111)
     at org.apache.spark.sql.execution.QueryExecution.$anonfun$executePhase$1(QueryExecution.scala:138)
     at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:767)
     at org.apache.spark.sql.execution.QueryExecution.executePhase(QueryExecution.scala:138)
     at org.apache.spark.sql.execution.QueryExecution.executedPlan$lzycompute(QueryExecution.scala:112)
     at org.apache.spark.sql.execution.QueryExecution.executedPlan(QueryExecution.scala:105)
     at org.apache.spark.sql.execution.QueryExecution.$anonfun$writePlans$5(QueryExecution.scala:204)
     at org.apache.spark.sql.catalyst.plans.QueryPlan$.append(QueryPlan.scala:478)
     at org.apache.spark.sql.execution.QueryExecution.org$apache$spark$sql$execution$QueryExecution$$writePlans(QueryExecution.scala:204)
     at org.apache.spark.sql.execution.QueryExecution.toString(QueryExecution.scala:212)
     at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$5(SQLExecution.scala:95)
     at org.apache.spark.sql.execution.SQLExecution$.withSQLConfPropagated(SQLExecution.scala:160)
     at org.apache.spark.sql.execution.SQLExecution$.$anonfun$withNewExecutionId$1(SQLExecution.scala:87)
     at org.apache.spark.sql.SparkSession.withActive(SparkSession.scala:767)
     at org.apache.spark.sql.execution.SQLExecution$.withNewExecutionId(SQLExecution.scala:64)
     at org.apache.spark.sql.Dataset.withAction(Dataset.scala:3616)
     at org.apache.spark.sql.Dataset.head(Dataset.scala:2697)
     at org.apache.spark.sql.Dataset.take(Dataset.scala:2904)
     at org.apache.spark.sql.Dataset.getRows(Dataset.scala:300)
     at org.apache.spark.sql.Dataset.showString(Dataset.scala:337)
     at org.apache.spark.sql.Dataset.show(Dataset.scala:826)
     at org.apache.spark.sql.Dataset.show(Dataset.scala:803)
     ... 47 elided
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org
For additional commands, e-mail: issues-help@iceberg.apache.org


[GitHub] [iceberg] openinx commented on issue #2868: read manifest file thought spark-sql throw an exception

Posted by GitBox <gi...@apache.org>.
openinx commented on issue #2868:
URL: https://github.com/apache/iceberg/issues/2868#issuecomment-886511060


   It's similar to this [issue](https://github.com/apache/iceberg/issues/2710) ,   you can try to use the master branch (which has included the [fix](https://github.com/apache/iceberg/pull/2454) ) to verify this again.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org
For additional commands, e-mail: issues-help@iceberg.apache.org


[GitHub] [iceberg] openinx closed issue #2868: read manifest file thought spark-sql throw an exception

Posted by GitBox <gi...@apache.org>.
openinx closed issue #2868:
URL: https://github.com/apache/iceberg/issues/2868


   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@iceberg.apache.org
For additional commands, e-mail: issues-help@iceberg.apache.org