You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "Ethan Guo (Jira)" <ji...@apache.org> on 2022/06/11 05:50:00 UTC

[jira] [Created] (HUDI-4223) Reading metadata table throws NullPointerException from getLogRecordScanner

Ethan Guo created HUDI-4223:
-------------------------------

             Summary: Reading metadata table throws NullPointerException from getLogRecordScanner
                 Key: HUDI-4223
                 URL: https://issues.apache.org/jira/browse/HUDI-4223
             Project: Apache Hudi
          Issue Type: Bug
          Components: metadata
    Affects Versions: 0.11.0
            Reporter: Ethan Guo
            Assignee: Ethan Guo
             Fix For: 0.11.1


When loading the metadata table in Spark shell using the following code, it throws NullPointerException from getLogRecordScanner
{code:java}
Caused by: java.lang.NullPointerException   at org.apache.hudi.metadata.HoodieBackedTableMetadata.getLogRecordScanner(HoodieBackedTableMetadata.java:484)   at org.apache.hudi.HoodieMergeOnReadRDD$.scanLog(HoodieMergeOnReadRDD.scala:342)   at org.apache.hudi.HoodieMergeOnReadRDD$LogFileIterator.<init>(HoodieMergeOnReadRDD.scala:173)   at org.apache.hudi.HoodieMergeOnReadRDD$RecordMergingFileIterator.<init>(HoodieMergeOnReadRDD.scala:252)   at org.apache.hudi.HoodieMergeOnReadRDD.compute(HoodieMergeOnReadRDD.scala:101)   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)   at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)   at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)   at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)   at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:52)   at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:373)   at org.apache.spark.rdd.RDD.iterator(RDD.scala:337)   at org.apache.spark.scheduler.ResultTask.runTask(ResultTask.scala:90)   at org.apache.spark.scheduler.Task.run(Task.scala:131)   at org.apache.spark.executor.Executor$TaskRunner.$anonfun$run$3(Executor.scala:506)   at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1462)   at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:509)   at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)   at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)   at java.lang.Thread.run(Thread.java:748)  {code}



--
This message was sent by Atlassian Jira
(v8.20.7#820007)