You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@zeppelin.apache.org by zj...@apache.org on 2020/07/21 02:10:11 UTC

[zeppelin] branch master updated: Revert "[ZEPPELIN-4950]. Support for manually specifying the Java version of Spark Interpreter Scala REPL"

This is an automated email from the ASF dual-hosted git repository.

zjffdu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/zeppelin.git


The following commit(s) were added to refs/heads/master by this push:
     new 0b85cac  Revert "[ZEPPELIN-4950].  Support for manually specifying the Java version of Spark Interpreter Scala REPL"
0b85cac is described below

commit 0b85cac1abfda3c85defc185d33b2d846c370525
Author: Jeff Zhang <zj...@apache.org>
AuthorDate: Tue Jul 21 10:09:54 2020 +0800

    Revert "[ZEPPELIN-4950].  Support for manually specifying the Java version of Spark Interpreter Scala REPL"
    
    This reverts commit 6045079b5f5cb89def74bef55518351d4966c2dd.
---
 docs/interpreter/spark.md                                             | 4 ----
 .../scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala    | 3 ---
 .../scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala    | 2 --
 .../scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala    | 3 ---
 4 files changed, 12 deletions(-)

diff --git a/docs/interpreter/spark.md b/docs/interpreter/spark.md
index 21069b4..5fc9305 100644
--- a/docs/interpreter/spark.md
+++ b/docs/interpreter/spark.md
@@ -199,10 +199,6 @@ You can also set other Spark properties which are not listed in the table. For a
     <td>false</td>
     <td>whether use yarn proxy url as spark weburl, e.g. http://localhost:8088/proxy/application_1583396598068_0004</td>
   </tr>
-  <td>spark.repl.target</td>
-      <td>jvm-1.8</td>
-      <td>Manually specifying the Java version of Spark Interpreter Scala REPL,Available options:[jvm-1.5, jvm-1.6, jvm-1.7, jvm-1.8] </td>
-    </tr>
 </table>
 
 Without any configuration, Spark interpreter works out of box in local mode. But if you want to connect to your Spark cluster, you'll need to follow below two simple steps.
diff --git a/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala b/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala
index f59f137..0eac200 100644
--- a/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala
+++ b/spark/scala-2.10/src/main/scala/org/apache/zeppelin/spark/SparkScala210Interpreter.scala
@@ -67,13 +67,10 @@ class SparkScala210Interpreter(override val conf: SparkConf,
       sparkHttpServer = server
       conf.set("spark.repl.class.uri", uri)
     }
-    val target = conf.get("spark.repl.target", "jvm-1.8")
 
     val settings = new Settings()
     settings.embeddedDefaults(sparkInterpreterClassLoader)
     settings.usejavacp.value = true
-    settings.target.value = target
-
     this.userJars = getUserJars()
     LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator))
     settings.classpath.value = userJars.mkString(File.pathSeparator)
diff --git a/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala b/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala
index d2fb971..cb5a016 100644
--- a/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala
+++ b/spark/scala-2.11/src/main/scala/org/apache/zeppelin/spark/SparkScala211Interpreter.scala
@@ -66,14 +66,12 @@ class SparkScala211Interpreter(override val conf: SparkConf,
       sparkHttpServer = server
       conf.set("spark.repl.class.uri", uri)
     }
-    val target = conf.get("spark.repl.target", "jvm-1.8")
 
     val settings = new Settings()
     settings.processArguments(List("-Yrepl-class-based",
       "-Yrepl-outdir", s"${outputDir.getAbsolutePath}"), true)
     settings.embeddedDefaults(sparkInterpreterClassLoader)
     settings.usejavacp.value = true
-    settings.target.value = target
 
     this.userJars = getUserJars()
     LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator))
diff --git a/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala b/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala
index 7f35125..2b04a1d 100644
--- a/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala
+++ b/spark/scala-2.12/src/main/scala/org/apache/zeppelin/spark/SparkScala212Interpreter.scala
@@ -60,15 +60,12 @@ class SparkScala212Interpreter(override val conf: SparkConf,
     LOGGER.info("Scala shell repl output dir: " + outputDir.getAbsolutePath)
     outputDir.deleteOnExit()
     conf.set("spark.repl.class.outputDir", outputDir.getAbsolutePath)
-    val target = conf.get("spark.repl.target", "jvm-1.8")
 
     val settings = new Settings()
     settings.processArguments(List("-Yrepl-class-based",
       "-Yrepl-outdir", s"${outputDir.getAbsolutePath}"), true)
     settings.embeddedDefaults(sparkInterpreterClassLoader)
     settings.usejavacp.value = true
-    settings.target.value = target
-
     this.userJars = getUserJars()
     LOGGER.info("UserJars: " + userJars.mkString(File.pathSeparator))
     settings.classpath.value = userJars.mkString(File.pathSeparator)