You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Ella Peltonen (JIRA)" <ji...@apache.org> on 2016/07/14 17:23:20 UTC

[jira] [Comment Edited] (SPARK-12945) ERROR LiveListenerBus: Listener JobProgressListener threw an exception

    [ https://issues.apache.org/jira/browse/SPARK-12945?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15377322#comment-15377322 ] 

Ella Peltonen edited comment on SPARK-12945 at 7/14/16 5:23 PM:
----------------------------------------------------------------

I got that error too multiple times from different jobs. I had UI disabled, so that might not be the reason. What helped was to reduce a number of partitions: I just coalesced my dataset in the beginning of the program from over 30.000 partitions to 300, and everything worked fine (but slowly).


was (Author: snorrie):
I got that error too multiple times from different jobs. I had UI disabled, so that might not be the reason. What helped was to reduce a number of partitions: I just coalesced my dataset in the beginning of the program from over 30.000 partitions to 300, and everything worked fine (but slowly).

- Ella // Uni. Helsinki

> ERROR LiveListenerBus: Listener JobProgressListener threw an exception
> ----------------------------------------------------------------------
>
>                 Key: SPARK-12945
>                 URL: https://issues.apache.org/jira/browse/SPARK-12945
>             Project: Spark
>          Issue Type: Bug
>          Components: Web UI
>    Affects Versions: 1.6.0
>         Environment: Linux, yarn-client
>            Reporter: Tristan
>            Priority: Minor
>
> Seeing this a lot; not sure if it is a problem or spurious error (I recall this was an ignorable issue in previous version). The UI seems to be working fine:
> ERROR LiveListenerBus: Listener JobProgressListener threw an exception
> java.lang.NullPointerException
>         at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:361)
>         at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:360)
>         at scala.collection.immutable.List.foreach(List.scala:318)
>         at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:32)
>         at scala.collection.mutable.ListBuffer.foreach(ListBuffer.scala:45)
>         at org.apache.spark.ui.jobs.JobProgressListener.onTaskEnd(JobProgressListener.scala:360)
>         at org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:42)
>         at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
>         at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
>         at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:55)
>         at org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(AsynchronousListenerBus.scala:80)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:64)
>         at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1180)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org