You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Holden Karau (Jira)" <ji...@apache.org> on 2020/10/24 17:01:00 UTC

[jira] [Resolved] (SPARK-30821) Executor pods with multiple containers will not be rescheduled unless all containers fail

     [ https://issues.apache.org/jira/browse/SPARK-30821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Holden Karau resolved SPARK-30821.
----------------------------------
    Fix Version/s: 3.1.0
                   3.0.2
       Resolution: Fixed

> Executor pods with multiple containers will not be rescheduled unless all containers fail
> -----------------------------------------------------------------------------------------
>
>                 Key: SPARK-30821
>                 URL: https://issues.apache.org/jira/browse/SPARK-30821
>             Project: Spark
>          Issue Type: Improvement
>          Components: Kubernetes, Spark Core
>    Affects Versions: 3.1.0
>            Reporter: Kevin Hogeland
>            Assignee: Apache Spark
>            Priority: Major
>             Fix For: 3.0.2, 3.1.0
>
>
> Since the restart policy of launched pods is Never, additional handling is required for pods that may have sidecar containers. The executor should be considered failed if any containers have terminated and have a non-zero exit code, but Spark currently only checks the pod phase. The pod phase will remain "running" as long as _any_ pods are still running. Kubernetes sidecar support in 1.18/1.19 does not address this situation, as sidecar containers are excluded from pod phase calculation.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org