You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Hyukjin Kwon (JIRA)" <ji...@apache.org> on 2019/05/21 04:13:08 UTC

[jira] [Resolved] (SPARK-20045) Make sure SparkHadoopMapReduceWriter is resilient to failures of writers and committers

     [ https://issues.apache.org/jira/browse/SPARK-20045?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Hyukjin Kwon resolved SPARK-20045.
----------------------------------
    Resolution: Incomplete

> Make sure SparkHadoopMapReduceWriter is resilient to failures of writers and committers
> ---------------------------------------------------------------------------------------
>
>                 Key: SPARK-20045
>                 URL: https://issues.apache.org/jira/browse/SPARK-20045
>             Project: Spark
>          Issue Type: Improvement
>          Components: Spark Core
>    Affects Versions: 2.2.0
>            Reporter: Steve Loughran
>            Priority: Minor
>              Labels: bulk-closed
>
> Following on from SPARK-20038: review {{SparkHadoopMapReduceWriter}} and ensure that it's failure handling code is itself resilient to follow on failures, especially in things like {{writer.close()}} and the abortTask/abortJob calls. That's to ensure as robust a cleanup as possible, and to stop the original exception getting lost.
> At a quick glance 
> 1. {{executeTask()}}'s catch logic should catch & log any failure in the {{writer.close()}} 
> 2. The Hadoop commit protocol's abort* operation's can throw IOEs. Again, they need to be caught and logged
> Should be testable with mocking, and worthwhile giving how important commit protocol resilience is.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org