You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@spark.apache.org by "Michael Huang (JIRA)" <ji...@apache.org> on 2018/01/06 04:53:00 UTC

[jira] [Comment Edited] (SPARK-12945) ERROR LiveListenerBus: Listener JobProgressListener threw an exception

    [ https://issues.apache.org/jira/browse/SPARK-12945?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16314369#comment-16314369 ] 

Michael Huang edited comment on SPARK-12945 at 1/6/18 4:52 AM:
---------------------------------------------------------------

We started seeing these exceptions after dialing down Spark UI params as below, in order to reduce Driver Heap usage. Spark 1.6.

{noformat}
--conf spark.ui.retainedJobs=50 \
--conf spark.ui.retainedStages=50 \
--conf spark.worker.ui.retainedExecutors=50 \
--conf spark.worker.ui.retainedDrivers=50 \
--conf spark.sql.ui.retainedExecutions=0 \
--conf spark.streaming.ui.retainedBatches=50 \
{noformat}

{noformat}
18/01/06 01:37:23 ERROR LiveListenerBus: Listener JobProgressListener threw an exception
java.lang.NullPointerException
	at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:361)
	at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:360)
	at scala.collection.immutable.List.foreach(List.scala:318)
	at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:32)
	at scala.collection.mutable.ListBuffer.foreach(ListBuffer.scala:45)
	at org.apache.spark.ui.jobs.JobProgressListener.onTaskEnd(JobProgressListener.scala:360)
	at org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:42)
	at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
	at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
	at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:55)
	at org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
	at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(AsynchronousListenerBus.scala:80)
	at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
	at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
	at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
	at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:64)
	at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1181)
	at org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)
{noformat}


was (Author: mhuang74):
We started seeing these exceptions after dialing down Spark UI params as below, in order to reduce Driver Heap usage.

{noformat}
--conf spark.ui.retainedJobs=50 \
--conf spark.ui.retainedStages=50 \
--conf spark.worker.ui.retainedExecutors=50 \
--conf spark.worker.ui.retainedDrivers=50 \
--conf spark.sql.ui.retainedExecutions=0 \
--conf spark.streaming.ui.retainedBatches=50 \
{noformat}

> ERROR LiveListenerBus: Listener JobProgressListener threw an exception
> ----------------------------------------------------------------------
>
>                 Key: SPARK-12945
>                 URL: https://issues.apache.org/jira/browse/SPARK-12945
>             Project: Spark
>          Issue Type: Bug
>          Components: Web UI
>    Affects Versions: 1.6.0, 2.0.2
>         Environment: Linux, yarn-client
>            Reporter: Tristan
>            Priority: Minor
>
> Seeing this a lot; not sure if it is a problem or spurious error (I recall this was an ignorable issue in previous version). The UI seems to be working fine:
> ERROR LiveListenerBus: Listener JobProgressListener threw an exception
> java.lang.NullPointerException
>         at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:361)
>         at org.apache.spark.ui.jobs.JobProgressListener$$anonfun$onTaskEnd$1.apply(JobProgressListener.scala:360)
>         at scala.collection.immutable.List.foreach(List.scala:318)
>         at scala.collection.generic.TraversableForwarder$class.foreach(TraversableForwarder.scala:32)
>         at scala.collection.mutable.ListBuffer.foreach(ListBuffer.scala:45)
>         at org.apache.spark.ui.jobs.JobProgressListener.onTaskEnd(JobProgressListener.scala:360)
>         at org.apache.spark.scheduler.SparkListenerBus$class.onPostEvent(SparkListenerBus.scala:42)
>         at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
>         at org.apache.spark.scheduler.LiveListenerBus.onPostEvent(LiveListenerBus.scala:31)
>         at org.apache.spark.util.ListenerBus$class.postToAll(ListenerBus.scala:55)
>         at org.apache.spark.util.AsynchronousListenerBus.postToAll(AsynchronousListenerBus.scala:37)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(AsynchronousListenerBus.scala:80)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1$$anonfun$apply$mcV$sp$1.apply(AsynchronousListenerBus.scala:65)
>         at scala.util.DynamicVariable.withValue(DynamicVariable.scala:57)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1$$anonfun$run$1.apply$mcV$sp(AsynchronousListenerBus.scala:64)
>         at org.apache.spark.util.Utils$.tryOrStopSparkContext(Utils.scala:1180)
>         at org.apache.spark.util.AsynchronousListenerBus$$anon$1.run(AsynchronousListenerBus.scala:63)



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@spark.apache.org
For additional commands, e-mail: issues-help@spark.apache.org