You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Flink Jira Bot (Jira)" <ji...@apache.org> on 2021/10/30 22:40:01 UTC

[jira] [Updated] (FLINK-20504) NPE when writing to hive and fail over happened

     [ https://issues.apache.org/jira/browse/FLINK-20504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Flink Jira Bot updated FLINK-20504:
-----------------------------------
    Labels: auto-deprioritized-major stale-minor  (was: auto-deprioritized-major)

I am the [Flink Jira Bot|https://github.com/apache/flink-jira-bot/] and I help the community manage its development. I see this issues has been marked as Minor but is unassigned and neither itself nor its Sub-Tasks have been updated for 180 days. I have gone ahead and marked it "stale-minor". If this ticket is still Minor, please either assign yourself or give an update. Afterwards, please remove the label or in 7 days the issue will be deprioritized.


> NPE when writing to hive and fail over happened
> -----------------------------------------------
>
>                 Key: FLINK-20504
>                 URL: https://issues.apache.org/jira/browse/FLINK-20504
>             Project: Flink
>          Issue Type: Bug
>          Components: Connectors / Hive
>    Affects Versions: 1.11.1
>            Reporter: zhuxiaoshang
>            Priority: Minor
>              Labels: auto-deprioritized-major, stale-minor
>
> When writing to hive and fail over happened,I got the following exception
> {code:java}
> java.lang.NullPointerException
> at org.apache.parquet.hadoop.InternalParquetRecordWriter.flushRowGroupToStore(InternalParquetRecordWriter.java:165)
> at org.apache.parquet.hadoop.InternalParquetRecordWriter.close(InternalParquetRecordWriter.java:114)
> at org.apache.parquet.hadoop.ParquetRecordWriter.close(ParquetRecordWriter.java:165)
> at org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:103)
> at org.apache.hadoop.hive.ql.io.parquet.write.ParquetRecordWriterWrapper.close(ParquetRecordWriterWrapper.java:120)
> at org.apache.flink.connectors.hive.write.HiveBulkWriterFactory$1.dispose(HiveBulkWriterFactory.java:61)
> at org.apache.flink.formats.hadoop.bulk.HadoopPathBasedPartFileWriter.dispose(HadoopPathBasedPartFileWriter.java:79)
> at org.apache.flink.streaming.api.functions.sink.filesystem.Bucket.disposePartFile(Bucket.java:235)
> at java.util.HashMap$Values.forEach(HashMap.java:981)
> at org.apache.flink.streaming.api.functions.sink.filesystem.Buckets.close(Buckets.java:318)
> at org.apache.flink.streaming.api.functions.sink.filesystem.StreamingFileSinkHelper.close(StreamingFileSinkHelper.java:108)
> at org.apache.flink.table.filesystem.stream.StreamingFileWriter.dispose(StreamingFileWriter.java:177)
> at org.apache.flink.streaming.runtime.tasks.StreamTask.disposeAllOperators(StreamTask.java:703)
> at org.apache.flink.streaming.runtime.tasks.StreamTask.cleanUpInvoke(StreamTask.java:635)
> at org.apache.flink.streaming.runtime.tasks.StreamTask.invoke(StreamTask.java:542)
> at org.apache.flink.runtime.taskmanager.Task.doRun(Task.java:721)
> at org.apache.flink.runtime.taskmanager.Task.run(Task.java:546)
> at java.lang.Thread.run(Thread.java:748)
> {code}
> But it does not reproduce every time.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)