You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Matt Cheah (JIRA)" <ji...@apache.org> on 2014/10/09 22:49:33 UTC

[jira] [Comment Edited] (SPARK-3835) Spark applications that are killed should show up as "KILLED" or "CANCELLED" in the Spark UI

    [ https://issues.apache.org/jira/browse/SPARK-3835?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14165722#comment-14165722 ] 

Matt Cheah edited comment on SPARK-3835 at 10/9/14 8:48 PM:
------------------------------------------------------------

Any updates on this? I've tried tackling it myself but I'm actually not sure how possible this is - killing a JVM just causes a DisassociatedEvent to be fired... but a DisassociatedEvent is also fired if SparkContext.stop() is called, making it hard to tell if a context was stopped gracefully or forcefully.


was (Author: mcheah):
Any updates on this?

> Spark applications that are killed should show up as "KILLED" or "CANCELLED" in the Spark UI
> --------------------------------------------------------------------------------------------
>
>                 Key: SPARK-3835
>                 URL: https://issues.apache.org/jira/browse/SPARK-3835
>             Project: Spark
>          Issue Type: Improvement
>          Components: Web UI
>    Affects Versions: 1.1.0
>            Reporter: Matt Cheah
>              Labels: UI
>
> Spark applications that crash or are killed are listed as FINISHED in the Spark UI.
> It looks like the Master only passes back a list of "Running" applications and a list of "Completed" applications, All of the applications under "Completed" have status "FINISHED", however if they were killed manually they should show "CANCELLED", or if they failed they should read "FAILED".



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org