You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "Leonard Xu (Jira)" <ji...@apache.org> on 2022/11/01 09:38:00 UTC
[jira] [Updated] (FLINK-5518) HadoopInputFormat throws NPE when close() is called before open()
[ https://issues.apache.org/jira/browse/FLINK-5518?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Leonard Xu updated FLINK-5518:
------------------------------
Labels: beginner easyfix starter (was: beginner easyfix newbie)
> HadoopInputFormat throws NPE when close() is called before open()
> -----------------------------------------------------------------
>
> Key: FLINK-5518
> URL: https://issues.apache.org/jira/browse/FLINK-5518
> Project: Flink
> Issue Type: Bug
> Components: API / DataSet
> Affects Versions: 1.1.4, 1.2.0
> Reporter: Jakub Havlik
> Assignee: Jakub Havlik
> Priority: Major
> Labels: beginner, easyfix, starter
> Fix For: 1.1.5, 1.2.0, 1.3.0
>
>
> When developing a simple Flink applications reading ORC files it crashes with NullPointerException when number of instances/executor threads is higher then the number of files because it is trying to close a HadoopInputFormat which is trying to close RecordReader which was not yet initialized as there is no file for which it should have been opened. The issue is caused when
> {code:java}
> public void run(SourceContext<OUT> ctx) throws Exception {
> try {
> ...
> while (isRunning) {
> format.open(splitIterator.next());
> ...
> } finally {
> format.close();
> ...
> }
> {code}
> in file {{InputFormatSourceFunction.java}} which calls
> {code:java}
> public void close() throws IOException {
> // enforce sequential close() calls
> synchronized (CLOSE_MUTEX) {
> this.recordReader.close();
> }
> }
> {code}
> from {{HadoopInputFormatBase.java}}.
> As there is just this one implementation of the {{close()}} method it may be enough just to add a null check for the {{this.recordReader}} in there.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)