You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Sean Owen (JIRA)" <ji...@apache.org> on 2015/05/29 20:49:17 UTC

[jira] [Resolved] (SPARK-7951) Events Missing with custom sparkListener

     [ https://issues.apache.org/jira/browse/SPARK-7951?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Sean Owen resolved SPARK-7951.
------------------------------
    Resolution: Invalid

I don't think this clearly describes a problem, and sounds like a question for user@ first. Please see https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark

> Events Missing with custom sparkListener
> ----------------------------------------
>
>                 Key: SPARK-7951
>                 URL: https://issues.apache.org/jira/browse/SPARK-7951
>             Project: Spark
>          Issue Type: Bug
>          Components: Spark Core
>    Affects Versions: 1.3.0
>            Reporter: Anubhav Srivastava
>              Labels: patch
>   Original Estimate: 48h
>  Remaining Estimate: 48h
>
> If we try to add our own SparkListener, we miss some of the events. I went through the code for spark-core, figured out that initial events like onApplicationStart() etc are fired when SparkContext() is initialised. Default eventListeners are getting those events as they are added while the sparkContext object is created, but when we add our own listener by calling sc().addListener(), while firing initial events our listener is not present at that time so its not getting those events.
>  In some cases you may be able to get those events as when asyncListenerBus is created, it spawns a new thread so in case the execution of that thread takes time, our listener might get added by that time so we may receive those events but that's not fair.
> I Can fix it. In-fact the fix is working fine for me. But just wanted to know is that done purposely?? If yes then what's the actual reason???



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org