You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@zeppelin.apache.org by "Philipp Dallig (Jira)" <ji...@apache.org> on 2020/04/29 08:14:00 UTC

[jira] [Created] (ZEPPELIN-4781) SparkUI-Link is not displayed in ZeppelinUI

Philipp Dallig created ZEPPELIN-4781:
----------------------------------------

             Summary: SparkUI-Link is not displayed in ZeppelinUI
                 Key: ZEPPELIN-4781
                 URL: https://issues.apache.org/jira/browse/ZEPPELIN-4781
             Project: Zeppelin
          Issue Type: Bug
          Components: spark
    Affects Versions: 0.9.0
            Reporter: Philipp Dallig


Hi
I remember Zeppelin showing a link to SparkUI when a Spark interpreter is running. Now the link is gone.
versions: 
 - zeppelin -> master branch
 - spark -> spark3 (Preview version)

I am a little engaged in this problem. I added the following code to my local installation to debug the problem.

{code:java}
diff --git a/spark/spark3-shims/src/main/scala/org/apache/zeppelin/spark/Spark3Shims.java b/spark/spark3-shims/src/main/scala/org/apache/zeppelin/spark/Spark3Shims.java
index b213041df..8930bc036 100644
--- a/spark/spark3-shims/src/main/scala/org/apache/zeppelin/spark/Spark3Shims.java
+++ b/spark/spark3-shims/src/main/scala/org/apache/zeppelin/spark/Spark3Shims.java
@@ -20,7 +20,9 @@ package org.apache.zeppelin.spark;
 
 import org.apache.commons.lang.StringUtils;
 import org.apache.spark.SparkContext;
+import org.apache.spark.SparkFirehoseListener;
 import org.apache.spark.scheduler.SparkListener;
+import org.apache.spark.scheduler.SparkListenerEvent;
 import org.apache.spark.scheduler.SparkListenerJobStart;
 import org.apache.spark.sql.Dataset;
 import org.apache.spark.sql.Row;
@@ -30,6 +32,8 @@ import org.apache.spark.sql.types.StructType;
 import org.apache.zeppelin.interpreter.InterpreterContext;
 import org.apache.zeppelin.interpreter.ResultMessages;
 import org.apache.zeppelin.tabledata.TableDataUtils;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
 
 import java.util.ArrayList;
 import java.util.List;
@@ -39,25 +43,39 @@ public class Spark3Shims extends SparkShims {
 
   private SparkSession sparkSession;
 
+  private static final Logger LOGGER = LoggerFactory.getLogger(Spark3Shims.class);
+
   public Spark3Shims(Properties properties, Object entryPoint) {
     super(properties);
     this.sparkSession = (SparkSession) entryPoint;
   }
 
+  @Override
   public void setupSparkListener(final String master,
                                  final String sparkWebUrl,
                                  final InterpreterContext context) {
     SparkContext sc = SparkContext.getOrCreate();
+    LOGGER.info("Spark3Shims - setupSparkListener");
     sc.addSparkListener(new SparkListener() {
       @Override
       public void onJobStart(SparkListenerJobStart jobStart) {
 
+        LOGGER.info("SparkUI");
+        LOGGER.info("spark.ui.enabled: {}", sc.getConf().getBoolean("spark.ui.enabled", true));
+        LOGGER.info("zeppelin.spark.ui.hidden: {}", Boolean.parseBoolean(properties.getProperty("zeppelin.spark.ui.hidden", "false")));
         if (sc.getConf().getBoolean("spark.ui.enabled", true) &&
             !Boolean.parseBoolean(properties.getProperty("zeppelin.spark.ui.hidden", "false"))) {
           buildSparkJobUrl(master, sparkWebUrl, jobStart.jobId(), jobStart.properties(), context);
         }
       }
     });
+    sc.addSparkListener(new SparkFirehoseListener( ) {
+      @Override
+      public void onEvent(SparkListenerEvent event) {
+        LOGGER.info("Event: {}", event.getClass());
+          super.onEvent(event);
+        }
+     });
   }
 
   @Override
{code}

The spark interpreter has now additional output.
{code:none}
INFO [2020-04-29 09:53:36,628] ({FIFOScheduler-interpreter_1800946569-Worker-1} SparkShims.java[loadShims]:61) - Initializing shims for Spark 3.x
INFO [2020-04-29 09:53:36,629] ({FIFOScheduler-interpreter_1800946569-Worker-1} Spark3Shims.java[setupSparkListener]:58) - Spark3Shims - setupSparkListener
INFO [2020-04-29 09:53:37,602] ({FIFOScheduler-interpreter_1800946569-Worker-1} AbstractScheduler.java[runJob]:152) - Job paragraph_1585936744837_-1110526008 finished by scheduler interpreter_1800946569
INFO [2020-04-29 09:53:46,570] ({spark-listener-group-shared} Spark3Shims.java[onEvent]:75) - Event: class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate
INFO [2020-04-29 09:53:47,840] ({spark-listener-group-shared} Spark3Shims.java[onEvent]:75) - Event: class org.apache.spark.scheduler.SparkListenerExecutorMetricsUpdate;
{code}

I hope that someone can help to fix this problem. It seems that the Event `jobStart` is never called by Spark.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)