You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Andrew Ash (JIRA)" <ji...@apache.org> on 2014/11/14 09:58:34 UTC
[jira] [Commented] (SPARK-672) Executor gets stuck in a "zombie"
state after running out of memory
[ https://issues.apache.org/jira/browse/SPARK-672?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14212020#comment-14212020 ]
Andrew Ash commented on SPARK-672:
----------------------------------
[~mbautin] when an executor JVM is under extremely heavy GC load, it will often lock up, not even responding to normal kill commands and requiring a kill -9 to shut down. I agree that Spark could behave better in these situations.
What is your preference for handling these issues -- maybe give the executor a timeout and kill -9 it after a certain period of time of non-responsiveness?
> Executor gets stuck in a "zombie" state after running out of memory
> -------------------------------------------------------------------
>
> Key: SPARK-672
> URL: https://issues.apache.org/jira/browse/SPARK-672
> Project: Spark
> Issue Type: Bug
> Reporter: Mikhail Bautin
> Attachments: executor_jstack.txt, executor_stderr.txt, standalone_worker_jstack.txt
>
>
> As a result of running a workload, an executor ran out of memory, but the executor process stayed up. Also (not sure this is related) the standalone worker process stayed up but disappeared from the master web UI.
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)
---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org