You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Davies Liu (JIRA)" <ji...@apache.org> on 2015/07/11 04:32:04 UTC

[jira] [Assigned] (SPARK-7735) Raise Exception on non-zero exit from pyspark pipe commands

     [ https://issues.apache.org/jira/browse/SPARK-7735?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Davies Liu reassigned SPARK-7735:
---------------------------------

    Assignee: Davies Liu

> Raise Exception on non-zero exit from pyspark pipe commands
> -----------------------------------------------------------
>
>                 Key: SPARK-7735
>                 URL: https://issues.apache.org/jira/browse/SPARK-7735
>             Project: Spark
>          Issue Type: Bug
>          Components: PySpark
>    Affects Versions: 1.3.0, 1.3.1
>            Reporter: Scott Taylor
>            Assignee: Davies Liu
>            Priority: Minor
>              Labels: newbie, patch
>             Fix For: 1.5.0
>
>
> In pyspark errors are ignored when using the rdd.pipe function. This is different to the scala behaviour where abnormal exit of the piped command is raised. I have submitted a pull request on github which I believe will bring the pyspark behaviour closer to the scala behaviour.
> A simple case of where this bug may be problematic is using a network bash utility to perform computations on an rdd. Currently, network errors will be ignored and blank results returned when it would be more desirable to raise an exception so that spark can retry the failed task.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org