You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by "leesf (Jira)" <ji...@apache.org> on 2022/01/03 07:40:00 UTC

[jira] [Resolved] (HUDI-3134) Fix Insert error after adding columns on Spark 3.2.0

     [ https://issues.apache.org/jira/browse/HUDI-3134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

leesf resolved HUDI-3134.
-------------------------

> Fix Insert error after adding columns on Spark 3.2.0
> ----------------------------------------------------
>
>                 Key: HUDI-3134
>                 URL: https://issues.apache.org/jira/browse/HUDI-3134
>             Project: Apache Hudi
>          Issue Type: Sub-task
>            Reporter: leesf
>            Assignee: leesf
>            Priority: Major
>              Labels: pull-request-available
>
> On Spark 3.2.0, after altering table to add columns, the insert statement will fail with the following exception.
> Caused by: org.apache.hudi.exception.HoodieException: java.util.concurrent.ExecutionException: org.apache.hudi.exception.HoodieException: operation has failed
>   at org.apache.hudi.common.util.queue.BoundedInMemoryExecutor.execute(BoundedInMemoryExecutor.java:147)
>   at org.apache.hudi.table.action.commit.SparkMergeHelper.runMerge(SparkMergeHelper.java:100)
>   ... 31 more
> Caused by: java.util.concurrent.ExecutionException: org.apache.hudi.exception.HoodieException: operation has failed
>   at java.util.concurrent.FutureTask.report(FutureTask.java:122)
>   at java.util.concurrent.FutureTask.get(FutureTask.java:192)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryExecutor.execute(BoundedInMemoryExecutor.java:141)
>   ... 32 more
> Caused by: org.apache.hudi.exception.HoodieException: operation has failed
>   at org.apache.hudi.common.util.queue.BoundedInMemoryQueue.throwExceptionIfFailed(BoundedInMemoryQueue.java:248)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryQueue.readNextRecord(BoundedInMemoryQueue.java:226)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryQueue.access$100(BoundedInMemoryQueue.java:52)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryQueue$QueueIterator.hasNext(BoundedInMemoryQueue.java:278)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryQueueConsumer.consume(BoundedInMemoryQueueConsumer.java:36)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryExecutor.lambda$null$2(BoundedInMemoryExecutor.java:121)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   ... 3 more
> Caused by: java.lang.NoSuchMethodError: org.apache.avro.Schema$Field.defaultValue()Lorg/codehaus/jackson/JsonNode;
>   at org.apache.parquet.avro.AvroRecordConverter.<init>(AvroRecordConverter.java:168)
>   at org.apache.parquet.avro.AvroRecordConverter.<init>(AvroRecordConverter.java:95)
>   at org.apache.parquet.avro.AvroRecordMaterializer.<init>(AvroRecordMaterializer.java:33)
>   at org.apache.parquet.avro.AvroReadSupport.prepareForRead(AvroReadSupport.java:138)
>   at org.apache.parquet.hadoop.InternalParquetRecordReader.initialize(InternalParquetRecordReader.java:185)
>   at org.apache.parquet.hadoop.ParquetReader.initReader(ParquetReader.java:156)
>   at org.apache.parquet.hadoop.ParquetReader.read(ParquetReader.java:135)
>   at org.apache.hudi.common.util.ParquetReaderIterator.hasNext(ParquetReaderIterator.java:49)
>   at org.apache.hudi.common.util.queue.IteratorBasedQueueProducer.produce(IteratorBasedQueueProducer.java:45)
>   at org.apache.hudi.common.util.queue.BoundedInMemoryExecutor.lambda$null$0(BoundedInMemoryExecutor.java:92)
>   at java.util.concurrent.FutureTask.run(FutureTask.java:266)
>   at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>   ... 4 more



--
This message was sent by Atlassian Jira
(v8.20.1#820001)