You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@zeppelin.apache.org by jo...@apache.org on 2018/08/26 04:44:18 UTC

[35/49] zeppelin git commit: [ZEPPELIN-3740] Adopt `google-java-format` and `fmt-maven-plugin`

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java
index 20c336d..c33db1f 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/IPySparkInterpreterTest.java
@@ -17,8 +17,19 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Matchers.any;
+import static org.mockito.Mockito.mock;
+import static org.mockito.Mockito.reset;
+import static org.mockito.Mockito.verify;
 
 import com.google.common.io.Files;
+import java.io.IOException;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.Map;
+import java.util.Properties;
 import org.apache.zeppelin.interpreter.Interpreter;
 import org.apache.zeppelin.interpreter.InterpreterContext;
 import org.apache.zeppelin.interpreter.InterpreterException;
@@ -32,23 +43,11 @@ import org.apache.zeppelin.interpreter.thrift.InterpreterCompletion;
 import org.apache.zeppelin.python.IPythonInterpreterTest;
 import org.junit.Test;
 
-import java.io.IOException;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Map;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Matchers.any;
-import static org.mockito.Mockito.mock;
-import static org.mockito.Mockito.reset;
-import static org.mockito.Mockito.verify;
-
 public class IPySparkInterpreterTest extends IPythonInterpreterTest {
 
   private InterpreterGroup intpGroup;
-  private RemoteInterpreterEventClient mockIntpEventClient = mock(RemoteInterpreterEventClient.class);
+  private RemoteInterpreterEventClient mockIntpEventClient =
+      mock(RemoteInterpreterEventClient.class);
 
   @Override
   protected Properties initIntpProperties() {
@@ -67,15 +66,14 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     return p;
   }
 
-
   @Override
   protected void startInterpreter(Properties properties) throws InterpreterException {
     InterpreterContext context = getInterpreterContext();
     context.setIntpEventClient(mockIntpEventClient);
     InterpreterContext.set(context);
 
-    LazyOpenInterpreter sparkInterpreter = new LazyOpenInterpreter(
-        new SparkInterpreter(properties));
+    LazyOpenInterpreter sparkInterpreter =
+        new LazyOpenInterpreter(new SparkInterpreter(properties));
     intpGroup = new InterpreterGroup();
     intpGroup.put("session_1", new ArrayList<Interpreter>());
     intpGroup.get("session_1").add(sparkInterpreter);
@@ -93,7 +91,6 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     interpreter.open();
   }
 
-
   @Override
   public void tearDown() throws InterpreterException {
     intpGroup.close();
@@ -106,7 +103,8 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     testPySpark(interpreter, mockIntpEventClient);
   }
 
-  public static void testPySpark(final Interpreter interpreter, RemoteInterpreterEventClient mockIntpEventClient)
+  public static void testPySpark(
+      final Interpreter interpreter, RemoteInterpreterEventClient mockIntpEventClient)
       throws InterpreterException, IOException, InterruptedException {
     reset(mockIntpEventClient);
     // rdd
@@ -120,7 +118,8 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     result = interpreter.interpret("sc.range(1,10).sum()", context);
     Thread.sleep(100);
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-    List<InterpreterResultMessage> interpreterResultMessages = context.out.toInterpreterResultMessage();
+    List<InterpreterResultMessage> interpreterResultMessages =
+        context.out.toInterpreterResultMessage();
     assertEquals("45", interpreterResultMessages.get(0).getData().trim());
     // spark job url is sent
     verify(mockIntpEventClient).onParaInfosReceived(any(Map.class));
@@ -128,69 +127,73 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     // spark sql
     context = createInterpreterContext(mockIntpEventClient);
     if (!isSpark2(sparkVersion)) {
-      result = interpreter.interpret("df = sqlContext.createDataFrame([(1,'a'),(2,'b')])\ndf.show()", context);
+      result =
+          interpreter.interpret(
+              "df = sqlContext.createDataFrame([(1,'a'),(2,'b')])\ndf.show()", context);
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       interpreterResultMessages = context.out.toInterpreterResultMessage();
       assertEquals(
-          "+---+---+\n" +
-              "| _1| _2|\n" +
-              "+---+---+\n" +
-              "|  1|  a|\n" +
-              "|  2|  b|\n" +
-              "+---+---+", interpreterResultMessages.get(0).getData().trim());
+          "+---+---+\n"
+              + "| _1| _2|\n"
+              + "+---+---+\n"
+              + "|  1|  a|\n"
+              + "|  2|  b|\n"
+              + "+---+---+",
+          interpreterResultMessages.get(0).getData().trim());
 
       context = createInterpreterContext(mockIntpEventClient);
       result = interpreter.interpret("z.show(df)", context);
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       interpreterResultMessages = context.out.toInterpreterResultMessage();
-      assertEquals(
-          "_1	_2\n" +
-              "1	a\n" +
-              "2	b", interpreterResultMessages.get(0).getData().trim());
+      assertEquals("_1	_2\n" + "1	a\n" + "2	b", interpreterResultMessages.get(0).getData().trim());
     } else {
-      result = interpreter.interpret("df = spark.createDataFrame([(1,'a'),(2,'b')])\ndf.show()", context);
+      result =
+          interpreter.interpret(
+              "df = spark.createDataFrame([(1,'a'),(2,'b')])\ndf.show()", context);
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       interpreterResultMessages = context.out.toInterpreterResultMessage();
       assertEquals(
-          "+---+---+\n" +
-              "| _1| _2|\n" +
-              "+---+---+\n" +
-              "|  1|  a|\n" +
-              "|  2|  b|\n" +
-              "+---+---+", interpreterResultMessages.get(0).getData().trim());
+          "+---+---+\n"
+              + "| _1| _2|\n"
+              + "+---+---+\n"
+              + "|  1|  a|\n"
+              + "|  2|  b|\n"
+              + "+---+---+",
+          interpreterResultMessages.get(0).getData().trim());
 
       context = createInterpreterContext(mockIntpEventClient);
       result = interpreter.interpret("z.show(df)", context);
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       interpreterResultMessages = context.out.toInterpreterResultMessage();
-      assertEquals(
-          "_1	_2\n" +
-              "1	a\n" +
-              "2	b", interpreterResultMessages.get(0).getData().trim());
+      assertEquals("_1	_2\n" + "1	a\n" + "2	b", interpreterResultMessages.get(0).getData().trim());
     }
     // cancel
     if (interpreter instanceof IPySparkInterpreter) {
       final InterpreterContext context2 = createInterpreterContext(mockIntpEventClient);
 
-      Thread thread = new Thread() {
-        @Override
-        public void run() {
-          InterpreterResult result = null;
-          try {
-            result = interpreter.interpret("import time\nsc.range(1,10).foreach(lambda x: time.sleep(1))", context2);
-          } catch (InterpreterException e) {
-            e.printStackTrace();
-          }
-          assertEquals(InterpreterResult.Code.ERROR, result.code());
-          List<InterpreterResultMessage> interpreterResultMessages = null;
-          try {
-            interpreterResultMessages = context2.out.toInterpreterResultMessage();
-            assertTrue(interpreterResultMessages.get(0).getData().contains("KeyboardInterrupt"));
-          } catch (IOException e) {
-            e.printStackTrace();
-          }
-        }
-      };
+      Thread thread =
+          new Thread() {
+            @Override
+            public void run() {
+              InterpreterResult result = null;
+              try {
+                result =
+                    interpreter.interpret(
+                        "import time\nsc.range(1,10).foreach(lambda x: time.sleep(1))", context2);
+              } catch (InterpreterException e) {
+                e.printStackTrace();
+              }
+              assertEquals(InterpreterResult.Code.ERROR, result.code());
+              List<InterpreterResultMessage> interpreterResultMessages = null;
+              try {
+                interpreterResultMessages = context2.out.toInterpreterResultMessage();
+                assertTrue(
+                    interpreterResultMessages.get(0).getData().contains("KeyboardInterrupt"));
+              } catch (IOException e) {
+                e.printStackTrace();
+              }
+            }
+          };
       thread.start();
 
       // sleep 1 second to wait for the spark job starts
@@ -200,7 +203,8 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     }
 
     // completions
-    List<InterpreterCompletion> completions = interpreter.completion("sc.ran", 6, createInterpreterContext(mockIntpEventClient));
+    List<InterpreterCompletion> completions =
+        interpreter.completion("sc.ran", 6, createInterpreterContext(mockIntpEventClient));
     assertEquals(1, completions.size());
     assertEquals("range", completions.get(0).getValue());
 
@@ -208,7 +212,8 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     assertTrue(completions.size() > 0);
     completions.contains(new InterpreterCompletion("range", "range", ""));
 
-    completions = interpreter.completion("1+1\nsc.", 7, createInterpreterContext(mockIntpEventClient));
+    completions =
+        interpreter.completion("1+1\nsc.", 7, createInterpreterContext(mockIntpEventClient));
     assertTrue(completions.size() > 0);
     completions.contains(new InterpreterCompletion("range", "range", ""));
 
@@ -218,20 +223,22 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
 
     // pyspark streaming
     context = createInterpreterContext(mockIntpEventClient);
-    result = interpreter.interpret(
-        "from pyspark.streaming import StreamingContext\n" +
-            "import time\n" +
-            "ssc = StreamingContext(sc, 1)\n" +
-            "rddQueue = []\n" +
-            "for i in range(5):\n" +
-            "    rddQueue += [ssc.sparkContext.parallelize([j for j in range(1, 1001)], 10)]\n" +
-            "inputStream = ssc.queueStream(rddQueue)\n" +
-            "mappedStream = inputStream.map(lambda x: (x % 10, 1))\n" +
-            "reducedStream = mappedStream.reduceByKey(lambda a, b: a + b)\n" +
-            "reducedStream.pprint()\n" +
-            "ssc.start()\n" +
-            "time.sleep(6)\n" +
-            "ssc.stop(stopSparkContext=False, stopGraceFully=True)", context);
+    result =
+        interpreter.interpret(
+            "from pyspark.streaming import StreamingContext\n"
+                + "import time\n"
+                + "ssc = StreamingContext(sc, 1)\n"
+                + "rddQueue = []\n"
+                + "for i in range(5):\n"
+                + "    rddQueue += [ssc.sparkContext.parallelize([j for j in range(1, 1001)], 10)]\n"
+                + "inputStream = ssc.queueStream(rddQueue)\n"
+                + "mappedStream = inputStream.map(lambda x: (x % 10, 1))\n"
+                + "reducedStream = mappedStream.reduceByKey(lambda a, b: a + b)\n"
+                + "reducedStream.pprint()\n"
+                + "ssc.start()\n"
+                + "time.sleep(6)\n"
+                + "ssc.stop(stopSparkContext=False, stopGraceFully=True)",
+            context);
     Thread.sleep(1000);
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     interpreterResultMessages = context.out.toInterpreterResultMessage();
@@ -243,7 +250,8 @@ public class IPySparkInterpreterTest extends IPythonInterpreterTest {
     return sparkVersion.startsWith("'2.") || sparkVersion.startsWith("u'2.");
   }
 
-  private static InterpreterContext createInterpreterContext(RemoteInterpreterEventClient mockRemoteEventClient) {
+  private static InterpreterContext createInterpreterContext(
+      RemoteInterpreterEventClient mockRemoteEventClient) {
     return InterpreterContext.builder()
         .setNoteId("noteId")
         .setParagraphId("paragraphId")

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java
index ea19866..2722bfb 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkInterpreterTest.java
@@ -17,7 +17,24 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Matchers.any;
+import static org.mockito.Mockito.mock;
+import static org.mockito.Mockito.never;
+import static org.mockito.Mockito.verify;
+
 import com.google.common.io.Files;
+import java.io.File;
+import java.io.FileOutputStream;
+import java.io.IOException;
+import java.net.URL;
+import java.nio.channels.Channels;
+import java.nio.channels.ReadableByteChannel;
+import java.util.ArrayList;
+import java.util.List;
+import java.util.Map;
+import java.util.Properties;
 import org.apache.zeppelin.display.AngularObjectRegistry;
 import org.apache.zeppelin.display.ui.CheckBox;
 import org.apache.zeppelin.display.ui.Password;
@@ -37,26 +54,6 @@ import org.junit.After;
 import org.junit.Ignore;
 import org.junit.Test;
 
-import java.io.File;
-import java.io.FileOutputStream;
-import java.io.IOException;
-import java.net.URL;
-import java.nio.channels.Channels;
-import java.nio.channels.ReadableByteChannel;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Map;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Matchers.any;
-import static org.mockito.Mockito.mock;
-import static org.mockito.Mockito.never;
-import static org.mockito.Mockito.reset;
-import static org.mockito.Mockito.verify;
-
-
 public class NewSparkInterpreterTest {
 
   private SparkInterpreter interpreter;
@@ -67,10 +64,12 @@ public class NewSparkInterpreterTest {
   // catch the interpreter output in onUpdate
   private InterpreterResultMessageOutput messageOutput;
 
-  private RemoteInterpreterEventClient mockRemoteEventClient = mock(RemoteInterpreterEventClient.class);
+  private RemoteInterpreterEventClient mockRemoteEventClient =
+      mock(RemoteInterpreterEventClient.class);
 
   @Test
-  public void testSparkInterpreter() throws IOException, InterruptedException, InterpreterException {
+  public void testSparkInterpreter()
+      throws IOException, InterruptedException, InterpreterException {
     Properties properties = new Properties();
     properties.setProperty("spark.master", "local");
     properties.setProperty("spark.app.name", "test");
@@ -79,11 +78,12 @@ public class NewSparkInterpreterTest {
     properties.setProperty("zeppelin.spark.useNew", "true");
     properties.setProperty("zeppelin.spark.uiWebUrl", "fake_spark_weburl");
 
-    InterpreterContext context = InterpreterContext.builder()
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mockRemoteEventClient)
-        .setAngularObjectRegistry(new AngularObjectRegistry("spark", null))
-        .build();
+    InterpreterContext context =
+        InterpreterContext.builder()
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mockRemoteEventClient)
+            .setAngularObjectRegistry(new AngularObjectRegistry("spark", null))
+            .build();
     InterpreterContext.set(context);
 
     interpreter = new SparkInterpreter(properties);
@@ -93,7 +93,8 @@ public class NewSparkInterpreterTest {
 
     assertEquals("fake_spark_weburl", interpreter.getSparkUIUrl());
 
-    InterpreterResult result = interpreter.interpret("val a=\"hello world\"", getInterpreterContext());
+    InterpreterResult result =
+        interpreter.interpret("val a=\"hello world\"", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals("a: String = hello world\n", output);
 
@@ -120,11 +121,13 @@ public class NewSparkInterpreterTest {
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     // single line comment
-    result = interpreter.interpret("print(\"hello world\")/*comment here*/", getInterpreterContext());
+    result =
+        interpreter.interpret("print(\"hello world\")/*comment here*/", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals("hello world", output);
 
-    result = interpreter.interpret("/*comment here*/\nprint(\"hello world\")", getInterpreterContext());
+    result =
+        interpreter.interpret("/*comment here*/\nprint(\"hello world\")", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     // multiple line comment
@@ -132,27 +135,32 @@ public class NewSparkInterpreterTest {
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     // test function
-    result = interpreter.interpret("def add(x:Int, y:Int)\n{ return x+y }", getInterpreterContext());
+    result =
+        interpreter.interpret("def add(x:Int, y:Int)\n{ return x+y }", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     result = interpreter.interpret("print(add(1,2))", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-    result = interpreter.interpret("/*line 1 \n line 2*/print(\"hello world\")", getInterpreterContext());
+    result =
+        interpreter.interpret(
+            "/*line 1 \n line 2*/print(\"hello world\")", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-
     // Companion object with case class
-    result = interpreter.interpret("import scala.math._\n" +
-        "object Circle {\n" +
-        "  private def calculateArea(radius: Double): Double = Pi * pow(radius, 2.0)\n" +
-        "}\n" +
-        "case class Circle(radius: Double) {\n" +
-        "  import Circle._\n" +
-        "  def area: Double = calculateArea(radius)\n" +
-        "}\n" +
-        "\n" +
-        "val circle1 = new Circle(5.0)", getInterpreterContext());
+    result =
+        interpreter.interpret(
+            "import scala.math._\n"
+                + "object Circle {\n"
+                + "  private def calculateArea(radius: Double): Double = Pi * pow(radius, 2.0)\n"
+                + "}\n"
+                + "case class Circle(radius: Double) {\n"
+                + "  import Circle._\n"
+                + "  def area: Double = calculateArea(radius)\n"
+                + "}\n"
+                + "\n"
+                + "val circle1 = new Circle(5.0)",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     // spark rdd operation
@@ -164,19 +172,22 @@ public class NewSparkInterpreterTest {
     verify(mockRemoteEventClient).onParaInfosReceived(any(Map.class));
 
     // case class
-    result = interpreter.interpret("val bankText = sc.textFile(\"bank.csv\")", getInterpreterContext());
+    result =
+        interpreter.interpret("val bankText = sc.textFile(\"bank.csv\")", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-    result = interpreter.interpret(
-        "case class Bank(age:Integer, job:String, marital : String, education : String, balance : Integer)\n" +
-            "val bank = bankText.map(s=>s.split(\";\")).filter(s => s(0)!=\"\\\"age\\\"\").map(\n" +
-            "    s => Bank(s(0).toInt, \n" +
-            "            s(1).replaceAll(\"\\\"\", \"\"),\n" +
-            "            s(2).replaceAll(\"\\\"\", \"\"),\n" +
-            "            s(3).replaceAll(\"\\\"\", \"\"),\n" +
-            "            s(5).replaceAll(\"\\\"\", \"\").toInt\n" +
-            "        )\n" +
-            ").toDF()", getInterpreterContext());
+    result =
+        interpreter.interpret(
+            "case class Bank(age:Integer, job:String, marital : String, education : String, balance : Integer)\n"
+                + "val bank = bankText.map(s=>s.split(\";\")).filter(s => s(0)!=\"\\\"age\\\"\").map(\n"
+                + "    s => Bank(s(0).toInt, \n"
+                + "            s(1).replaceAll(\"\\\"\", \"\"),\n"
+                + "            s(2).replaceAll(\"\\\"\", \"\"),\n"
+                + "            s(3).replaceAll(\"\\\"\", \"\"),\n"
+                + "            s(5).replaceAll(\"\\\"\", \"\").toInt\n"
+                + "        )\n"
+                + ").toDF()",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     // spark version
@@ -189,32 +200,36 @@ public class NewSparkInterpreterTest {
       result = interpreter.interpret("sqlContext", getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-      result = interpreter.interpret(
-          "val df = sqlContext.createDataFrame(Seq((1,\"a\"),(2, null)))\n" +
-              "df.show()", getInterpreterContext());
+      result =
+          interpreter.interpret(
+              "val df = sqlContext.createDataFrame(Seq((1,\"a\"),(2, null)))\n" + "df.show()",
+              getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-      assertTrue(output.contains(
-          "+---+----+\n" +
-              "| _1|  _2|\n" +
-              "+---+----+\n" +
-              "|  1|   a|\n" +
-              "|  2|null|\n" +
-              "+---+----+"));
+      assertTrue(
+          output.contains(
+              "+---+----+\n"
+                  + "| _1|  _2|\n"
+                  + "+---+----+\n"
+                  + "|  1|   a|\n"
+                  + "|  2|null|\n"
+                  + "+---+----+"));
     } else if (version.contains("String = 2.")) {
       result = interpreter.interpret("spark", getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-      result = interpreter.interpret(
-          "val df = spark.createDataFrame(Seq((1,\"a\"),(2, null)))\n" +
-              "df.show()", getInterpreterContext());
+      result =
+          interpreter.interpret(
+              "val df = spark.createDataFrame(Seq((1,\"a\"),(2, null)))\n" + "df.show()",
+              getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-      assertTrue(output.contains(
-          "+---+----+\n" +
-              "| _1|  _2|\n" +
-              "+---+----+\n" +
-              "|  1|   a|\n" +
-              "|  2|null|\n" +
-              "+---+----+"));
+      assertTrue(
+          output.contains(
+              "+---+----+\n"
+                  + "| _1|  _2|\n"
+                  + "+---+----+\n"
+                  + "|  1|   a|\n"
+                  + "|  2|null|\n"
+                  + "+---+----+"));
     }
 
     // ZeppelinContext
@@ -242,7 +257,10 @@ public class NewSparkInterpreterTest {
     assertEquals("pwd", pwd.getName());
 
     context = getInterpreterContext();
-    result = interpreter.interpret("z.checkbox(\"checkbox_1\", Seq(\"value_2\"), Seq((\"value_1\", \"name_1\"), (\"value_2\", \"name_2\")))", context);
+    result =
+        interpreter.interpret(
+            "z.checkbox(\"checkbox_1\", Seq(\"value_2\"), Seq((\"value_1\", \"name_1\"), (\"value_2\", \"name_2\")))",
+            context);
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals(1, context.getGui().getForms().size());
     assertTrue(context.getGui().getForms().get("checkbox_1") instanceof CheckBox);
@@ -257,13 +275,17 @@ public class NewSparkInterpreterTest {
     assertEquals("name_2", checkBox.getOptions()[1].getDisplayName());
 
     context = getInterpreterContext();
-    result = interpreter.interpret("z.select(\"select_1\", Seq(\"value_2\"), Seq((\"value_1\", \"name_1\"), (\"value_2\", \"name_2\")))", context);
+    result =
+        interpreter.interpret(
+            "z.select(\"select_1\", Seq(\"value_2\"), Seq((\"value_1\", \"name_1\"), (\"value_2\", \"name_2\")))",
+            context);
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals(1, context.getGui().getForms().size());
     assertTrue(context.getGui().getForms().get("select_1") instanceof Select);
     Select select = (Select) context.getGui().getForms().get("select_1");
     assertEquals("select_1", select.getName());
-    // TODO(zjffdu) it seems a bug of GUI, the default value should be 'value_2', but it is List(value_2)
+    // TODO(zjffdu) it seems a bug of GUI, the default value should be 'value_2', but it is
+    // List(value_2)
     // assertEquals("value_2", select.getDefaultValue());
     assertEquals(2, select.getOptions().length);
     assertEquals("value_1", select.getOptions()[0].getValue());
@@ -271,9 +293,9 @@ public class NewSparkInterpreterTest {
     assertEquals("value_2", select.getOptions()[1].getValue());
     assertEquals("name_2", select.getOptions()[1].getDisplayName());
 
-
     // completions
-    List<InterpreterCompletion> completions = interpreter.completion("a.", 2, getInterpreterContext());
+    List<InterpreterCompletion> completions =
+        interpreter.completion("a.", 2, getInterpreterContext());
     assertTrue(completions.size() > 0);
 
     completions = interpreter.completion("a.isEm", 6, getInterpreterContext());
@@ -284,43 +306,57 @@ public class NewSparkInterpreterTest {
     assertEquals(1, completions.size());
     assertEquals("range", completions.get(0).name);
 
-
     // Zeppelin-Display
-    result = interpreter.interpret("import org.apache.zeppelin.display.angular.notebookscope._\n" +
-        "import AngularElem._", getInterpreterContext());
+    result =
+        interpreter.interpret(
+            "import org.apache.zeppelin.display.angular.notebookscope._\n" + "import AngularElem._",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
-    result = interpreter.interpret("<div style=\"color:blue\">\n" +
-        "<h4>Hello Angular Display System</h4>\n" +
-        "</div>.display", getInterpreterContext());
+    result =
+        interpreter.interpret(
+            "<div style=\"color:blue\">\n"
+                + "<h4>Hello Angular Display System</h4>\n"
+                + "</div>.display",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals(InterpreterResult.Type.ANGULAR, messageOutput.getType());
-    assertTrue(messageOutput.toInterpreterResultMessage().getData().contains("Hello Angular Display System"));
-
-    result = interpreter.interpret("<div class=\"btn btn-success\">\n" +
-        "  Click me\n" +
-        "</div>.onClick{() =>\n" +
-        "  println(\"hello world\")\n" +
-        "}.display", getInterpreterContext());
+    assertTrue(
+        messageOutput
+            .toInterpreterResultMessage()
+            .getData()
+            .contains("Hello Angular Display System"));
+
+    result =
+        interpreter.interpret(
+            "<div class=\"btn btn-success\">\n"
+                + "  Click me\n"
+                + "</div>.onClick{() =>\n"
+                + "  println(\"hello world\")\n"
+                + "}.display",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals(InterpreterResult.Type.ANGULAR, messageOutput.getType());
     assertTrue(messageOutput.toInterpreterResultMessage().getData().contains("Click me"));
 
     // getProgress
     final InterpreterContext context2 = getInterpreterContext();
-    Thread interpretThread = new Thread() {
-      @Override
-      public void run() {
-        InterpreterResult result = null;
-        try {
-          result = interpreter.interpret(
-              "val df = sc.parallelize(1 to 10, 2).foreach(e=>Thread.sleep(1000))", context2);
-        } catch (InterpreterException e) {
-          e.printStackTrace();
-        }
-        assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-      }
-    };
+    Thread interpretThread =
+        new Thread() {
+          @Override
+          public void run() {
+            InterpreterResult result = null;
+            try {
+              result =
+                  interpreter.interpret(
+                      "val df = sc.parallelize(1 to 10, 2).foreach(e=>Thread.sleep(1000))",
+                      context2);
+            } catch (InterpreterException e) {
+              e.printStackTrace();
+            }
+            assertEquals(InterpreterResult.Code.SUCCESS, result.code());
+          }
+        };
     interpretThread.start();
     boolean nonZeroProgress = false;
     int progress = 0;
@@ -336,20 +372,23 @@ public class NewSparkInterpreterTest {
 
     // cancel
     final InterpreterContext context3 = getInterpreterContext();
-    interpretThread = new Thread() {
-      @Override
-      public void run() {
-        InterpreterResult result = null;
-        try {
-          result = interpreter.interpret(
-              "val df = sc.parallelize(1 to 10, 2).foreach(e=>Thread.sleep(1000))", context3);
-        } catch (InterpreterException e) {
-          e.printStackTrace();
-        }
-        assertEquals(InterpreterResult.Code.ERROR, result.code());
-        assertTrue(output.contains("cancelled"));
-      }
-    };
+    interpretThread =
+        new Thread() {
+          @Override
+          public void run() {
+            InterpreterResult result = null;
+            try {
+              result =
+                  interpreter.interpret(
+                      "val df = sc.parallelize(1 to 10, 2).foreach(e=>Thread.sleep(1000))",
+                      context3);
+            } catch (InterpreterException e) {
+              e.printStackTrace();
+            }
+            assertEquals(InterpreterResult.Code.ERROR, result.code());
+            assertTrue(output.contains("cancelled"));
+          }
+        };
 
     interpretThread.start();
     // sleep 1 second to wait for the spark job start
@@ -367,7 +406,9 @@ public class NewSparkInterpreterTest {
     properties.setProperty("zeppelin.spark.useNew", "true");
 
     // download spark-avro jar
-    URL website = new URL("http://repo1.maven.org/maven2/com/databricks/spark-avro_2.11/3.2.0/spark-avro_2.11-3.2.0.jar");
+    URL website =
+        new URL(
+            "http://repo1.maven.org/maven2/com/databricks/spark-avro_2.11/3.2.0/spark-avro_2.11-3.2.0.jar");
     ReadableByteChannel rbc = Channels.newChannel(website.openStream());
     File avroJarFile = new File("spark-avro_2.11-3.2.0.jar");
     FileOutputStream fos = new FileOutputStream(avroJarFile);
@@ -380,11 +421,13 @@ public class NewSparkInterpreterTest {
     interpreter.setInterpreterGroup(mock(InterpreterGroup.class));
     interpreter.open();
 
-    InterpreterResult result = interpreter.interpret("import com.databricks.spark.avro._", getInterpreterContext());
+    InterpreterResult result =
+        interpreter.interpret("import com.databricks.spark.avro._", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
   }
 
-  //TODO(zjffdu) This unit test will fail due to classpath issue, should enable it after the classpath issue is fixed.
+  // TODO(zjffdu) This unit test will fail due to classpath issue, should enable it after the
+  // classpath issue is fixed.
   @Ignore
   public void testDepInterpreter() throws InterpreterException {
     Properties properties = new Properties();
@@ -406,7 +449,8 @@ public class NewSparkInterpreterTest {
 
     depInterpreter.open();
     InterpreterResult result =
-        depInterpreter.interpret("z.load(\"com.databricks:spark-avro_2.11:3.2.0\")", getInterpreterContext());
+        depInterpreter.interpret(
+            "z.load(\"com.databricks:spark-avro_2.11:3.2.0\")", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
 
     interpreter.open();
@@ -430,7 +474,8 @@ public class NewSparkInterpreterTest {
     interpreter.setInterpreterGroup(mock(InterpreterGroup.class));
     interpreter.open();
 
-    InterpreterResult result = interpreter.interpret("val a=\"hello world\"", getInterpreterContext());
+    InterpreterResult result =
+        interpreter.interpret("val a=\"hello world\"", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     // no output for define new variable
     assertEquals("", output);
@@ -532,19 +577,17 @@ public class NewSparkInterpreterTest {
 
   private InterpreterContext getInterpreterContext() {
     output = "";
-    InterpreterContext context = InterpreterContext.builder()
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mockRemoteEventClient)
-        .setAngularObjectRegistry(new AngularObjectRegistry("spark", null))
-        .build();
+    InterpreterContext context =
+        InterpreterContext.builder()
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mockRemoteEventClient)
+            .setAngularObjectRegistry(new AngularObjectRegistry("spark", null))
+            .build();
     context.out =
         new InterpreterOutput(
-
             new InterpreterOutputListener() {
               @Override
-              public void onUpdateAll(InterpreterOutput out) {
-
-              }
+              public void onUpdateAll(InterpreterOutput out) {}
 
               @Override
               public void onAppend(int index, InterpreterResultMessageOutput out, byte[] line) {

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java
index ed91ffe..cac3295 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/NewSparkSqlInterpreterTest.java
@@ -17,6 +17,12 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Mockito.mock;
+
+import java.util.LinkedList;
+import java.util.Properties;
 import org.apache.zeppelin.display.AngularObjectRegistry;
 import org.apache.zeppelin.interpreter.Interpreter;
 import org.apache.zeppelin.interpreter.InterpreterContext;
@@ -31,13 +37,6 @@ import org.junit.AfterClass;
 import org.junit.BeforeClass;
 import org.junit.Test;
 
-import java.util.LinkedList;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Mockito.mock;
-
 public class NewSparkSqlInterpreterTest {
 
   private static SparkSqlInterpreter sqlInterpreter;
@@ -64,15 +63,16 @@ public class NewSparkSqlInterpreterTest {
     intpGroup.get("session_1").add(sparkInterpreter);
     intpGroup.get("session_1").add(sqlInterpreter);
 
-    context = InterpreterContext.builder()
-        .setNoteId("noteId")
-        .setParagraphId("paragraphId")
-        .setParagraphTitle("title")
-        .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
-        .setResourcePool(new LocalResourcePool("id"))
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
-        .build();
+    context =
+        InterpreterContext.builder()
+            .setNoteId("noteId")
+            .setParagraphId("paragraphId")
+            .setParagraphTitle("title")
+            .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
+            .setResourcePool(new LocalResourcePool("id"))
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
+            .build();
     InterpreterContext.set(context);
 
     sparkInterpreter.open();
@@ -88,10 +88,13 @@ public class NewSparkSqlInterpreterTest {
   @Test
   public void test() throws InterpreterException {
     sparkInterpreter.interpret("case class Test(name:String, age:Int)", context);
-    sparkInterpreter.interpret("val test = sc.parallelize(Seq(Test(\"moon\", 33), Test(\"jobs\", 51), Test(\"gates\", 51), Test(\"park\", 34)))", context);
+    sparkInterpreter.interpret(
+        "val test = sc.parallelize(Seq(Test(\"moon\", 33), Test(\"jobs\", 51), Test(\"gates\", 51), Test(\"park\", 34)))",
+        context);
     sparkInterpreter.interpret("test.toDF.registerTempTable(\"test\")", context);
 
-    InterpreterResult ret = sqlInterpreter.interpret("select name, age from test where age < 40", context);
+    InterpreterResult ret =
+        sqlInterpreter.interpret("select name, age from test where age < 40", context);
     assertEquals(InterpreterResult.Code.SUCCESS, ret.code());
     assertEquals(Type.TABLE, ret.message().get(0).getType());
     assertEquals("name\tage\nmoon\t33\npark\t34\n", ret.message().get(0).getData());
@@ -100,7 +103,11 @@ public class NewSparkSqlInterpreterTest {
     assertEquals(InterpreterResult.Code.ERROR, ret.code());
     assertTrue(ret.message().get(0).getData().length() > 0);
 
-    assertEquals(InterpreterResult.Code.SUCCESS, sqlInterpreter.interpret("select case when name='aa' then name else name end from test", context).code());
+    assertEquals(
+        InterpreterResult.Code.SUCCESS,
+        sqlInterpreter
+            .interpret("select case when name='aa' then name else name end from test", context)
+            .code());
   }
 
   @Test
@@ -129,17 +136,14 @@ public class NewSparkSqlInterpreterTest {
         "val schema = StructType(Seq(StructField(\"name\", StringType, false),StructField(\"age\" , IntegerType, true),StructField(\"other\" , StringType, false)))",
         context);
     sparkInterpreter.interpret(
-        "val csv = sc.parallelize(Seq((\"jobs, 51, apple\"), (\"gates, , microsoft\")))",
-        context);
+        "val csv = sc.parallelize(Seq((\"jobs, 51, apple\"), (\"gates, , microsoft\")))", context);
     sparkInterpreter.interpret(
-        "val raw = csv.map(_.split(\",\")).map(p => Row(p(0),toInt(p(1)),p(2)))",
-        context);
-    sparkInterpreter.interpret("val people = sqlContext.createDataFrame(raw, schema)",
-        context);
+        "val raw = csv.map(_.split(\",\")).map(p => Row(p(0),toInt(p(1)),p(2)))", context);
+    sparkInterpreter.interpret("val people = sqlContext.createDataFrame(raw, schema)", context);
     sparkInterpreter.interpret("people.toDF.registerTempTable(\"people\")", context);
 
-    InterpreterResult ret = sqlInterpreter.interpret(
-        "select name, age from people where name = 'gates'", context);
+    InterpreterResult ret =
+        sqlInterpreter.interpret("select name, age from people where name = 'gates'", context);
     assertEquals(InterpreterResult.Code.SUCCESS, ret.code());
     assertEquals(Type.TABLE, ret.message().get(0).getType());
     assertEquals("name\tage\ngates\tnull\n", ret.message().get(0).getData());
@@ -161,34 +165,38 @@ public class NewSparkSqlInterpreterTest {
   @Test
   public void testConcurrentSQL() throws InterpreterException, InterruptedException {
     if (sparkInterpreter.getSparkVersion().isSpark2()) {
-      sparkInterpreter.interpret("spark.udf.register(\"sleep\", (e:Int) => {Thread.sleep(e*1000); e})", context);
+      sparkInterpreter.interpret(
+          "spark.udf.register(\"sleep\", (e:Int) => {Thread.sleep(e*1000); e})", context);
     } else {
-      sparkInterpreter.interpret("sqlContext.udf.register(\"sleep\", (e:Int) => {Thread.sleep(e*1000); e})", context);
+      sparkInterpreter.interpret(
+          "sqlContext.udf.register(\"sleep\", (e:Int) => {Thread.sleep(e*1000); e})", context);
     }
 
-    Thread thread1 = new Thread() {
-      @Override
-      public void run() {
-        try {
-          InterpreterResult result = sqlInterpreter.interpret("select sleep(10)", context);
-          assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-        } catch (InterpreterException e) {
-          e.printStackTrace();
-        }
-      }
-    };
-
-    Thread thread2 = new Thread() {
-      @Override
-      public void run() {
-        try {
-          InterpreterResult result = sqlInterpreter.interpret("select sleep(10)", context);
-          assertEquals(InterpreterResult.Code.SUCCESS, result.code());
-        } catch (InterpreterException e) {
-          e.printStackTrace();
-        }
-      }
-    };
+    Thread thread1 =
+        new Thread() {
+          @Override
+          public void run() {
+            try {
+              InterpreterResult result = sqlInterpreter.interpret("select sleep(10)", context);
+              assertEquals(InterpreterResult.Code.SUCCESS, result.code());
+            } catch (InterpreterException e) {
+              e.printStackTrace();
+            }
+          }
+        };
+
+    Thread thread2 =
+        new Thread() {
+          @Override
+          public void run() {
+            try {
+              InterpreterResult result = sqlInterpreter.interpret("select sleep(10)", context);
+              assertEquals(InterpreterResult.Code.SUCCESS, result.code());
+            } catch (InterpreterException e) {
+              e.printStackTrace();
+            }
+          }
+        };
 
     // start running 2 spark sql, each would sleep 10 seconds, the totally running time should
     // be less than 20 seconds, which means they run concurrently.
@@ -198,8 +206,6 @@ public class NewSparkSqlInterpreterTest {
     thread1.join();
     thread2.join();
     long end = System.currentTimeMillis();
-    assertTrue("running time must be less than 20 seconds", ((end - start)/1000) < 20);
-
+    assertTrue("running time must be less than 20 seconds", ((end - start) / 1000) < 20);
   }
-
 }

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java
index 8ae66b2..12e3252 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkInterpreterTest.java
@@ -17,6 +17,14 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Mockito.mock;
+
+import java.io.IOException;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Properties;
 import org.apache.spark.SparkConf;
 import org.apache.zeppelin.display.AngularObjectRegistry;
 import org.apache.zeppelin.interpreter.Interpreter;
@@ -41,21 +49,10 @@ import org.junit.runners.MethodSorters;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import java.io.IOException;
-import java.util.LinkedList;
-import java.util.List;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNotNull;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Mockito.mock;
-
 @FixMethodOrder(MethodSorters.NAME_ASCENDING)
 public class OldSparkInterpreterTest {
 
-  @ClassRule
-  public static TemporaryFolder tmpDir = new TemporaryFolder();
+  @ClassRule public static TemporaryFolder tmpDir = new TemporaryFolder();
 
   static SparkInterpreter repl;
   static InterpreterGroup intpGroup;
@@ -63,8 +60,7 @@ public class OldSparkInterpreterTest {
   static Logger LOGGER = LoggerFactory.getLogger(OldSparkInterpreterTest.class);
 
   /**
-   * Get spark version number as a numerical value.
-   * eg. 1.1.x => 11, 1.2.x => 12, 1.3.x => 13 ...
+   * Get spark version number as a numerical value. eg. 1.1.x => 11, 1.2.x => 12, 1.3.x => 13 ...
    */
   public static int getSparkVersionNumber(SparkInterpreter repl) {
     if (repl == null) {
@@ -91,15 +87,16 @@ public class OldSparkInterpreterTest {
   @BeforeClass
   public static void setUp() throws Exception {
     intpGroup = new InterpreterGroup();
-    context = InterpreterContext.builder()
-        .setNoteId("noteId")
-        .setParagraphId("paragraphId")
-        .setParagraphTitle("title")
-        .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
-        .setResourcePool(new LocalResourcePool("id"))
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
-        .build();
+    context =
+        InterpreterContext.builder()
+            .setNoteId("noteId")
+            .setParagraphId("paragraphId")
+            .setParagraphTitle("title")
+            .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
+            .setResourcePool(new LocalResourcePool("id"))
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
+            .build();
     InterpreterContext.set(context);
 
     intpGroup.put("note", new LinkedList<Interpreter>());
@@ -108,10 +105,10 @@ public class OldSparkInterpreterTest {
     intpGroup.get("note").add(repl);
     repl.open();
     // The first para interpretdr will set the Eventclient wrapper
-    //SparkInterpreter.interpret(String, InterpreterContext) ->
-    //SparkInterpreter.populateSparkWebUrl(InterpreterContext) ->
-    //ZeppelinContext.setEventClient(RemoteEventClientWrapper)
-    //running a dummy to ensure that we dont have any race conditions among tests
+    // SparkInterpreter.interpret(String, InterpreterContext) ->
+    // SparkInterpreter.populateSparkWebUrl(InterpreterContext) ->
+    // ZeppelinContext.setEventClient(RemoteEventClientWrapper)
+    // running a dummy to ensure that we dont have any race conditions among tests
     repl.interpret("sc", context);
   }
 
@@ -122,14 +119,14 @@ public class OldSparkInterpreterTest {
 
   @Test
   public void testBasicIntp() throws InterpreterException {
-    assertEquals(InterpreterResult.Code.SUCCESS,
-        repl.interpret("val a = 1\nval b = 2", context).code());
+    assertEquals(
+        InterpreterResult.Code.SUCCESS, repl.interpret("val a = 1\nval b = 2", context).code());
 
     // when interpret incomplete expression
     InterpreterResult incomplete = repl.interpret("val a = \"\"\"", context);
     assertEquals(InterpreterResult.Code.INCOMPLETE, incomplete.code());
     assertTrue(incomplete.message().get(0).getData().length() > 0); // expecting some error
-                                                   // message
+    // message
 
     /*
      * assertEquals(1, repl.getValue("a")); assertEquals(2, repl.getValue("b"));
@@ -153,30 +150,41 @@ public class OldSparkInterpreterTest {
 
   @Test
   public void testNextLineComments() throws InterpreterException {
-    assertEquals(InterpreterResult.Code.SUCCESS, repl.interpret("\"123\"\n/*comment here\n*/.toInt", context).code());
+    assertEquals(
+        InterpreterResult.Code.SUCCESS,
+        repl.interpret("\"123\"\n/*comment here\n*/.toInt", context).code());
   }
 
   @Test
   public void testNextLineCompanionObject() throws InterpreterException {
-    String code = "class Counter {\nvar value: Long = 0\n}\n // comment\n\n object Counter {\n def apply(x: Long) = new Counter()\n}";
+    String code =
+        "class Counter {\nvar value: Long = 0\n}\n // comment\n\n object Counter {\n def apply(x: Long) = new Counter()\n}";
     assertEquals(InterpreterResult.Code.SUCCESS, repl.interpret(code, context).code());
   }
 
   @Test
   public void testEndWithComment() throws InterpreterException {
-    assertEquals(InterpreterResult.Code.SUCCESS, repl.interpret("val c=1\n//comment", context).code());
+    assertEquals(
+        InterpreterResult.Code.SUCCESS, repl.interpret("val c=1\n//comment", context).code());
   }
 
   @Test
   public void testCreateDataFrame() throws InterpreterException {
     if (getSparkVersionNumber(repl) >= 13) {
       repl.interpret("case class Person(name:String, age:Int)\n", context);
-      repl.interpret("val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n", context);
+      repl.interpret(
+          "val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n",
+          context);
       repl.interpret("people.toDF.count", context);
-      assertEquals(new Long(4), context.getResourcePool().get(
-          context.getNoteId(),
-          context.getParagraphId(),
-          WellKnownResourceName.ZeppelinReplResult.toString()).get());
+      assertEquals(
+          new Long(4),
+          context
+              .getResourcePool()
+              .get(
+                  context.getNoteId(),
+                  context.getParagraphId(),
+                  WellKnownResourceName.ZeppelinReplResult.toString())
+              .get());
     }
   }
 
@@ -184,23 +192,26 @@ public class OldSparkInterpreterTest {
   public void testZShow() throws InterpreterException {
     String code = "";
     repl.interpret("case class Person(name:String, age:Int)\n", context);
-    repl.interpret("val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n", context);
+    repl.interpret(
+        "val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n",
+        context);
     if (getSparkVersionNumber(repl) < 13) {
       repl.interpret("people.registerTempTable(\"people\")", context);
       code = "z.show(sqlc.sql(\"select * from people\"))";
     } else {
       code = "z.show(people.toDF)";
     }
-      assertEquals(Code.SUCCESS, repl.interpret(code, context).code());
+    assertEquals(Code.SUCCESS, repl.interpret(code, context).code());
   }
 
   @Test
   public void testSparkSql() throws IOException, InterpreterException {
     repl.interpret("case class Person(name:String, age:Int)\n", context);
-    repl.interpret("val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n", context);
+    repl.interpret(
+        "val people = sc.parallelize(Seq(Person(\"moon\", 33), Person(\"jobs\", 51), Person(\"gates\", 51), Person(\"park\", 34)))\n",
+        context);
     assertEquals(Code.SUCCESS, repl.interpret("people.take(3)", context).code());
 
-
     if (getSparkVersionNumber(repl) <= 11) { // spark 1.2 or later does not allow create multiple
       // SparkContext in the same jvm by default.
       // create new interpreter
@@ -210,7 +221,9 @@ public class OldSparkInterpreterTest {
       repl2.open();
 
       repl2.interpret("case class Man(name:String, age:Int)", context);
-      repl2.interpret("val man = sc.parallelize(Seq(Man(\"moon\", 33), Man(\"jobs\", 51), Man(\"gates\", 51), Man(\"park\", 34)))", context);
+      repl2.interpret(
+          "val man = sc.parallelize(Seq(Man(\"moon\", 33), Man(\"jobs\", 51), Man(\"gates\", 51), Man(\"park\", 34)))",
+          context);
       assertEquals(Code.SUCCESS, repl2.interpret("man.take(3)", context).code());
       repl2.close();
     }
@@ -218,8 +231,9 @@ public class OldSparkInterpreterTest {
 
   @Test
   public void testReferencingUndefinedVal() throws InterpreterException {
-    InterpreterResult result = repl.interpret("def category(min: Int) = {"
-        + "    if (0 <= value) \"error\"" + "}", context);
+    InterpreterResult result =
+        repl.interpret(
+            "def category(min: Int) = {" + "    if (0 <= value) \"error\"" + "}", context);
     assertEquals(Code.ERROR, result.code());
   }
 
@@ -232,23 +246,26 @@ public class OldSparkInterpreterTest {
       String value = (String) intpProperty.get(key);
       LOGGER.debug(String.format("[%s]: [%s]", key, value));
       if (key.startsWith("spark.") && value.isEmpty()) {
-        assertTrue(String.format("configuration starting from 'spark.' should not be empty. [%s]", key), !sparkConf.contains(key) || !sparkConf.get(key).isEmpty());
+        assertTrue(
+            String.format("configuration starting from 'spark.' should not be empty. [%s]", key),
+            !sparkConf.contains(key) || !sparkConf.get(key).isEmpty());
       }
     }
   }
 
   @Test
-  public void shareSingleSparkContext() throws InterruptedException, IOException, InterpreterException {
+  public void shareSingleSparkContext()
+      throws InterruptedException, IOException, InterpreterException {
     // create another SparkInterpreter
     SparkInterpreter repl2 = new SparkInterpreter(getSparkTestProperties(tmpDir));
     repl2.setInterpreterGroup(intpGroup);
     intpGroup.get("note").add(repl2);
     repl2.open();
 
-    assertEquals(Code.SUCCESS,
-        repl.interpret("print(sc.parallelize(1 to 10).count())", context).code());
-    assertEquals(Code.SUCCESS,
-        repl2.interpret("print(sc.parallelize(1 to 10).count())", context).code());
+    assertEquals(
+        Code.SUCCESS, repl.interpret("print(sc.parallelize(1 to 10).count())", context).code());
+    assertEquals(
+        Code.SUCCESS, repl2.interpret("print(sc.parallelize(1 to 10).count())", context).code());
 
     repl2.close();
   }
@@ -297,17 +314,17 @@ public class OldSparkInterpreterTest {
   @Test
   public void testMultilineCompletion() throws InterpreterException {
     String buf = "val x = 1\nsc.";
-	List<InterpreterCompletion> completions = repl.completion(buf, buf.length(), null);
+    List<InterpreterCompletion> completions = repl.completion(buf, buf.length(), null);
     assertTrue(completions.size() > 0);
   }
 
   @Test
   public void testMultilineCompletionNewVar() throws InterpreterException {
     Assume.assumeFalse("this feature does not work with scala 2.10", Utils.isScala2_10());
-    Assume.assumeTrue("This feature does not work with scala < 2.11.8", Utils.isCompilerAboveScala2_11_7());
+    Assume.assumeTrue(
+        "This feature does not work with scala < 2.11.8", Utils.isCompilerAboveScala2_11_7());
     String buf = "val x = sc\nx.";
-	  List<InterpreterCompletion> completions = repl.completion(buf, buf.length(), null);
+    List<InterpreterCompletion> completions = repl.completion(buf, buf.length(), null);
     assertTrue(completions.size() > 0);
   }
-
 }

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java
index fa1e257..f69eb0c 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/OldSparkSqlInterpreterTest.java
@@ -17,6 +17,12 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Mockito.mock;
+
+import java.util.LinkedList;
+import java.util.Properties;
 import org.apache.zeppelin.display.AngularObjectRegistry;
 import org.apache.zeppelin.interpreter.Interpreter;
 import org.apache.zeppelin.interpreter.InterpreterContext;
@@ -33,17 +39,9 @@ import org.junit.ClassRule;
 import org.junit.Test;
 import org.junit.rules.TemporaryFolder;
 
-import java.util.LinkedList;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Mockito.mock;
-
 public class OldSparkSqlInterpreterTest {
 
-  @ClassRule
-  public static TemporaryFolder tmpDir = new TemporaryFolder();
+  @ClassRule public static TemporaryFolder tmpDir = new TemporaryFolder();
 
   static SparkSqlInterpreter sql;
   static SparkInterpreter repl;
@@ -74,15 +72,16 @@ public class OldSparkSqlInterpreterTest {
     sql.setInterpreterGroup(intpGroup);
     sql.open();
 
-    context = InterpreterContext.builder()
-        .setNoteId("noteId")
-        .setParagraphId("paragraphId")
-        .setParagraphTitle("title")
-        .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
-        .setResourcePool(new LocalResourcePool("id"))
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
-        .build();
+    context =
+        InterpreterContext.builder()
+            .setNoteId("noteId")
+            .setParagraphId("paragraphId")
+            .setParagraphTitle("title")
+            .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
+            .setResourcePool(new LocalResourcePool("id"))
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
+            .build();
   }
 
   @AfterClass
@@ -98,7 +97,9 @@ public class OldSparkSqlInterpreterTest {
   @Test
   public void test() throws InterpreterException {
     repl.interpret("case class Test(name:String, age:Int)", context);
-    repl.interpret("val test = sc.parallelize(Seq(Test(\"moon\", 33), Test(\"jobs\", 51), Test(\"gates\", 51), Test(\"park\", 34)))", context);
+    repl.interpret(
+        "val test = sc.parallelize(Seq(Test(\"moon\", 33), Test(\"jobs\", 51), Test(\"gates\", 51), Test(\"park\", 34)))",
+        context);
     if (isDataFrameSupported()) {
       repl.interpret("test.toDF.registerTempTable(\"test\")", context);
     } else {
@@ -114,7 +115,10 @@ public class OldSparkSqlInterpreterTest {
     assertEquals(InterpreterResult.Code.ERROR, ret.code());
     assertTrue(ret.message().get(0).getData().length() > 0);
 
-    assertEquals(InterpreterResult.Code.SUCCESS, sql.interpret("select case when name==\"aa\" then name else name end from test", context).code());
+    assertEquals(
+        InterpreterResult.Code.SUCCESS,
+        sql.interpret("select case when name==\"aa\" then name else name end from test", context)
+            .code());
   }
 
   @Test
@@ -149,23 +153,19 @@ public class OldSparkSqlInterpreterTest {
         "val schema = StructType(Seq(StructField(\"name\", StringType, false),StructField(\"age\" , IntegerType, true),StructField(\"other\" , StringType, false)))",
         context);
     repl.interpret(
-        "val csv = sc.parallelize(Seq((\"jobs, 51, apple\"), (\"gates, , microsoft\")))",
-        context);
+        "val csv = sc.parallelize(Seq((\"jobs, 51, apple\"), (\"gates, , microsoft\")))", context);
     repl.interpret(
-        "val raw = csv.map(_.split(\",\")).map(p => Row(p(0),toInt(p(1)),p(2)))",
-        context);
+        "val raw = csv.map(_.split(\",\")).map(p => Row(p(0),toInt(p(1)),p(2)))", context);
     if (isDataFrameSupported()) {
-      repl.interpret("val people = sqlContext.createDataFrame(raw, schema)",
-          context);
+      repl.interpret("val people = sqlContext.createDataFrame(raw, schema)", context);
       repl.interpret("people.toDF.registerTempTable(\"people\")", context);
     } else {
-      repl.interpret("val people = sqlContext.applySchema(raw, schema)",
-          context);
+      repl.interpret("val people = sqlContext.applySchema(raw, schema)", context);
       repl.interpret("people.registerTempTable(\"people\")", context);
     }
 
-    InterpreterResult ret = sql.interpret(
-        "select name, age from people where name = 'gates'", context);
+    InterpreterResult ret =
+        sql.interpret("select name, age from people where name = 'gates'", context);
     System.err.println("RET=" + ret.message());
     assertEquals(InterpreterResult.Code.SUCCESS, ret.code());
     assertEquals(Type.TABLE, ret.message().get(0).getType());

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java
index 5a05ad5..c365160 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterMatplotlibTest.java
@@ -17,6 +17,15 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertNotSame;
+import static org.junit.Assert.assertTrue;
+import static org.mockito.Mockito.mock;
+
+import java.io.IOException;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Properties;
 import org.apache.zeppelin.display.AngularObjectRegistry;
 import org.apache.zeppelin.interpreter.Interpreter;
 import org.apache.zeppelin.interpreter.InterpreterContext;
@@ -37,21 +46,10 @@ import org.junit.runners.MethodSorters;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import java.io.IOException;
-import java.util.LinkedList;
-import java.util.List;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertNotSame;
-import static org.junit.Assert.assertTrue;
-import static org.mockito.Mockito.mock;
-
 @FixMethodOrder(MethodSorters.NAME_ASCENDING)
 public class PySparkInterpreterMatplotlibTest {
 
-  @ClassRule
-  public static TemporaryFolder tmpDir = new TemporaryFolder();
+  @ClassRule public static TemporaryFolder tmpDir = new TemporaryFolder();
 
   static SparkInterpreter sparkInterpreter;
   static PySparkInterpreter pyspark;
@@ -61,21 +59,21 @@ public class PySparkInterpreterMatplotlibTest {
 
   public static class AltPySparkInterpreter extends PySparkInterpreter {
     /**
-     * Since pyspark output is  sent to an outputstream rather than
-     * being directly provided by interpret(), this subclass is created to
-     * override interpret() to append the result from the outputStream
-     * for the sake of convenience in testing.
+     * Since pyspark output is sent to an outputstream rather than being directly provided by
+     * interpret(), this subclass is created to override interpret() to append the result from the
+     * outputStream for the sake of convenience in testing.
      */
     public AltPySparkInterpreter(Properties property) {
       super(property);
     }
 
     /**
-     * This code is mainly copied from RemoteInterpreterServer.java which
-     * normally handles this in real use cases.
+     * This code is mainly copied from RemoteInterpreterServer.java which normally handles this in
+     * real use cases.
      */
     @Override
-    public InterpreterResult interpret(String st, InterpreterContext context) throws InterpreterException {
+    public InterpreterResult interpret(String st, InterpreterContext context)
+        throws InterpreterException {
       context.out.clear();
       InterpreterResult result = super.interpret(st, context);
       List<InterpreterResultMessage> resultMessages = null;
@@ -106,8 +104,7 @@ public class PySparkInterpreterMatplotlibTest {
   }
 
   /**
-   * Get spark version number as a numerical value.
-   * eg. 1.1.x => 11, 1.2.x => 12, 1.3.x => 13 ...
+   * Get spark version number as a numerical value. eg. 1.1.x => 11, 1.2.x => 12, 1.3.x => 13 ...
    */
   public static int getSparkVersionNumber() {
     if (sparkInterpreter == null) {
@@ -123,12 +120,13 @@ public class PySparkInterpreterMatplotlibTest {
   public static void setUp() throws Exception {
     intpGroup = new InterpreterGroup();
     intpGroup.put("note", new LinkedList<Interpreter>());
-    context = InterpreterContext.builder()
-        .setNoteId("note")
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
-        .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
-        .build();
+    context =
+        InterpreterContext.builder()
+            .setNoteId("note")
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mock(RemoteInterpreterEventClient.class))
+            .setAngularObjectRegistry(new AngularObjectRegistry(intpGroup.getId(), null))
+            .build();
     InterpreterContext.set(context);
 
     sparkInterpreter = new SparkInterpreter(getPySparkTestProperties());
@@ -183,7 +181,8 @@ public class PySparkInterpreterMatplotlibTest {
     InterpreterResult ret2;
     ret = pyspark.interpret("import matplotlib.pyplot as plt", context);
     ret = pyspark.interpret("plt.close()", context);
-    ret = pyspark.interpret("z.configure_mpl(interactive=False, close=True, angular=False)", context);
+    ret =
+        pyspark.interpret("z.configure_mpl(interactive=False, close=True, angular=False)", context);
     ret = pyspark.interpret("plt.plot([1, 2, 3])", context);
     ret1 = pyspark.interpret("plt.show()", context);
 
@@ -210,7 +209,9 @@ public class PySparkInterpreterMatplotlibTest {
     InterpreterResult ret2;
     ret = pyspark.interpret("import matplotlib.pyplot as plt", context);
     ret = pyspark.interpret("plt.close()", context);
-    ret = pyspark.interpret("z.configure_mpl(interactive=False, close=False, angular=False)", context);
+    ret =
+        pyspark.interpret(
+            "z.configure_mpl(interactive=False, close=False, angular=False)", context);
     ret = pyspark.interpret("plt.plot([1, 2, 3])", context);
     ret1 = pyspark.interpret("plt.show()", context);
 
@@ -235,7 +236,8 @@ public class PySparkInterpreterMatplotlibTest {
     InterpreterResult ret;
     ret = pyspark.interpret("import matplotlib.pyplot as plt", context);
     ret = pyspark.interpret("plt.close()", context);
-    ret = pyspark.interpret("z.configure_mpl(interactive=False, close=False, angular=True)", context);
+    ret =
+        pyspark.interpret("z.configure_mpl(interactive=False, close=False, angular=True)", context);
     ret = pyspark.interpret("plt.plot([1, 2, 3])", context);
     ret = pyspark.interpret("plt.show()", context);
     assertEquals(ret.message().toString(), InterpreterResult.Code.SUCCESS, ret.code());

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java
index 64f1ff5..0a3677c 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/PySparkInterpreterTest.java
@@ -17,8 +17,12 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.mockito.Mockito.mock;
 
 import com.google.common.io.Files;
+import java.io.IOException;
+import java.util.LinkedList;
+import java.util.Properties;
 import org.apache.zeppelin.interpreter.Interpreter;
 import org.apache.zeppelin.interpreter.InterpreterContext;
 import org.apache.zeppelin.interpreter.InterpreterException;
@@ -29,16 +33,10 @@ import org.apache.zeppelin.interpreter.remote.RemoteInterpreterEventClient;
 import org.apache.zeppelin.python.PythonInterpreterTest;
 import org.junit.Test;
 
-import java.io.IOException;
-import java.util.LinkedList;
-import java.util.Properties;
-
-import static org.mockito.Mockito.mock;
-
-
 public class PySparkInterpreterTest extends PythonInterpreterTest {
 
-  private RemoteInterpreterEventClient mockRemoteEventClient = mock(RemoteInterpreterEventClient.class);
+  private RemoteInterpreterEventClient mockRemoteEventClient =
+      mock(RemoteInterpreterEventClient.class);
 
   @Override
   public void setUp() throws InterpreterException {
@@ -59,10 +57,11 @@ public class PySparkInterpreterTest extends PythonInterpreterTest {
     intpGroup = new InterpreterGroup();
     intpGroup.put("note", new LinkedList<Interpreter>());
 
-    InterpreterContext context = InterpreterContext.builder()
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setIntpEventClient(mockRemoteEventClient)
-        .build();
+    InterpreterContext context =
+        InterpreterContext.builder()
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setIntpEventClient(mockRemoteEventClient)
+            .build();
     InterpreterContext.set(context);
     LazyOpenInterpreter sparkInterpreter =
         new LazyOpenInterpreter(new SparkInterpreter(properties));

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java
index fb9ad62..e685d8a 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkRInterpreterTest.java
@@ -17,6 +17,17 @@
 
 package org.apache.zeppelin.spark;
 
+import static org.junit.Assert.assertEquals;
+import static org.junit.Assert.assertTrue;
+import static org.junit.Assert.fail;
+import static org.mockito.Matchers.any;
+import static org.mockito.Mockito.atLeastOnce;
+import static org.mockito.Mockito.mock;
+import static org.mockito.Mockito.verify;
+
+import java.util.HashMap;
+import java.util.Map;
+import java.util.Properties;
 import org.apache.zeppelin.interpreter.InterpreterContext;
 import org.apache.zeppelin.interpreter.InterpreterException;
 import org.apache.zeppelin.interpreter.InterpreterGroup;
@@ -28,23 +39,12 @@ import org.junit.After;
 import org.junit.Before;
 import org.junit.Test;
 
-import java.util.HashMap;
-import java.util.Map;
-import java.util.Properties;
-
-import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertTrue;
-import static org.junit.Assert.fail;
-import static org.mockito.Matchers.any;
-import static org.mockito.Mockito.atLeastOnce;
-import static org.mockito.Mockito.mock;
-import static org.mockito.Mockito.verify;
-
 public class SparkRInterpreterTest {
 
   private SparkRInterpreter sparkRInterpreter;
   private SparkInterpreter sparkInterpreter;
-  private RemoteInterpreterEventClient mockRemoteIntpEventClient = mock(RemoteInterpreterEventClient.class);
+  private RemoteInterpreterEventClient mockRemoteIntpEventClient =
+      mock(RemoteInterpreterEventClient.class);
 
   @Before
   public void setUp() throws InterpreterException {
@@ -63,8 +63,10 @@ public class SparkRInterpreterTest {
     sparkInterpreter = new SparkInterpreter(properties);
 
     InterpreterGroup interpreterGroup = new InterpreterGroup();
-    interpreterGroup.addInterpreterToSession(new LazyOpenInterpreter(sparkRInterpreter), "session_1");
-    interpreterGroup.addInterpreterToSession(new LazyOpenInterpreter(sparkInterpreter), "session_1");
+    interpreterGroup.addInterpreterToSession(
+        new LazyOpenInterpreter(sparkRInterpreter), "session_1");
+    interpreterGroup.addInterpreterToSession(
+        new LazyOpenInterpreter(sparkInterpreter), "session_1");
     sparkRInterpreter.setInterpreterGroup(interpreterGroup);
     sparkInterpreter.setInterpreterGroup(interpreterGroup);
 
@@ -79,7 +81,6 @@ public class SparkRInterpreterTest {
   @Test
   public void testSparkRInterpreter() throws InterpreterException, InterruptedException {
 
-
     InterpreterResult result = sparkRInterpreter.interpret("1+1", getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertTrue(result.message().get(0).getData().contains("2"));
@@ -88,7 +89,9 @@ public class SparkRInterpreterTest {
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     if (result.message().get(0).getData().contains("2.")) {
       // spark 2.x
-      result = sparkRInterpreter.interpret("df <- as.DataFrame(faithful)\nhead(df)", getInterpreterContext());
+      result =
+          sparkRInterpreter.interpret(
+              "df <- as.DataFrame(faithful)\nhead(df)", getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       assertTrue(result.message().get(0).getData().contains("eruptions waiting"));
       // spark job url is sent
@@ -96,30 +99,36 @@ public class SparkRInterpreterTest {
 
       // cancel
       final InterpreterContext context = getInterpreterContext();
-      Thread thread = new Thread() {
-        @Override
-        public void run() {
-          try {
-            InterpreterResult result = sparkRInterpreter.interpret("ldf <- dapplyCollect(\n" +
-                "         df,\n" +
-                "         function(x) {\n" +
-                "           Sys.sleep(3)\n" +
-                "           x <- cbind(x, \"waiting_secs\" = x$waiting * 60)\n" +
-                "         })\n" +
-                "head(ldf, 3)", context);
-            assertTrue(result.message().get(0).getData().contains("cancelled"));
-          } catch (InterpreterException e) {
-            fail("Should not throw InterpreterException");
-          }
-        }
-      };
+      Thread thread =
+          new Thread() {
+            @Override
+            public void run() {
+              try {
+                InterpreterResult result =
+                    sparkRInterpreter.interpret(
+                        "ldf <- dapplyCollect(\n"
+                            + "         df,\n"
+                            + "         function(x) {\n"
+                            + "           Sys.sleep(3)\n"
+                            + "           x <- cbind(x, \"waiting_secs\" = x$waiting * 60)\n"
+                            + "         })\n"
+                            + "head(ldf, 3)",
+                        context);
+                assertTrue(result.message().get(0).getData().contains("cancelled"));
+              } catch (InterpreterException e) {
+                fail("Should not throw InterpreterException");
+              }
+            }
+          };
       thread.setName("Cancel-Thread");
       thread.start();
       Thread.sleep(1000);
       sparkRInterpreter.cancel(context);
     } else {
       // spark 1.x
-      result = sparkRInterpreter.interpret("df <- createDataFrame(sqlContext, faithful)\nhead(df)", getInterpreterContext());
+      result =
+          sparkRInterpreter.interpret(
+              "df <- createDataFrame(sqlContext, faithful)\nhead(df)", getInterpreterContext());
       assertEquals(InterpreterResult.Code.SUCCESS, result.code());
       assertTrue(result.message().get(0).getData().contains("eruptions waiting"));
       // spark job url is sent
@@ -136,8 +145,11 @@ public class SparkRInterpreterTest {
     assertTrue(result.message().get(0).getData().contains("<img src="));
     assertTrue(result.message().get(0).getData().contains("width=\"100\""));
 
-    result = sparkRInterpreter.interpret("library(ggplot2)\n" +
-        "ggplot(diamonds, aes(x=carat, y=price, color=cut)) + geom_point()", getInterpreterContext());
+    result =
+        sparkRInterpreter.interpret(
+            "library(ggplot2)\n"
+                + "ggplot(diamonds, aes(x=carat, y=price, color=cut)) + geom_point()",
+            getInterpreterContext());
     assertEquals(InterpreterResult.Code.SUCCESS, result.code());
     assertEquals(1, result.message().size());
     assertEquals(InterpreterResult.Type.HTML, result.message().get(0).getType());
@@ -151,14 +163,14 @@ public class SparkRInterpreterTest {
   }
 
   private InterpreterContext getInterpreterContext() {
-    InterpreterContext context = InterpreterContext.builder()
-        .setNoteId("note_1")
-        .setParagraphId("paragraph_1")
-        .setIntpEventClient(mockRemoteIntpEventClient)
-        .setInterpreterOut(new InterpreterOutput(null))
-        .setLocalProperties(new HashMap<>())
-        .build();
+    InterpreterContext context =
+        InterpreterContext.builder()
+            .setNoteId("note_1")
+            .setParagraphId("paragraph_1")
+            .setIntpEventClient(mockRemoteIntpEventClient)
+            .setInterpreterOut(new InterpreterOutput(null))
+            .setLocalProperties(new HashMap<>())
+            .build();
     return context;
   }
 }
-

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkShimsTest.java
----------------------------------------------------------------------
diff --git a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkShimsTest.java b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkShimsTest.java
index 48d0055..d6170b7 100644
--- a/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkShimsTest.java
+++ b/spark/interpreter/src/test/java/org/apache/zeppelin/spark/SparkShimsTest.java
@@ -41,7 +41,6 @@ import org.junit.runners.Parameterized.Parameter;
 import org.junit.runners.Parameterized.Parameters;
 import org.mockito.ArgumentCaptor;
 import org.mockito.Captor;
-import org.mockito.Mock;
 import org.powermock.core.classloader.annotations.PowerMockIgnore;
 import org.powermock.core.classloader.annotations.PrepareForTest;
 import org.powermock.modules.junit4.PowerMockRunner;
@@ -91,9 +90,8 @@ public class SparkShimsTest {
       SparkShims sparkShims =
           new SparkShims(new Properties()) {
             @Override
-            public void setupSparkListener(String master,
-                                           String sparkWebUrl,
-                                           InterpreterContext context) {}
+            public void setupSparkListener(
+                String master, String sparkWebUrl, InterpreterContext context) {}
 
             @Override
             public String showDataFrame(Object obj, int maxResult) {

http://git-wip-us.apache.org/repos/asf/zeppelin/blob/55f6c91c/zeppelin-examples/zeppelin-example-clock/src/main/java/org/apache/zeppelin/example/app/clock/Clock.java
----------------------------------------------------------------------
diff --git a/zeppelin-examples/zeppelin-example-clock/src/main/java/org/apache/zeppelin/example/app/clock/Clock.java b/zeppelin-examples/zeppelin-example-clock/src/main/java/org/apache/zeppelin/example/app/clock/Clock.java
index bee8cf1..e22e003 100644
--- a/zeppelin-examples/zeppelin-example-clock/src/main/java/org/apache/zeppelin/example/app/clock/Clock.java
+++ b/zeppelin-examples/zeppelin-example-clock/src/main/java/org/apache/zeppelin/example/app/clock/Clock.java
@@ -16,6 +16,9 @@
  */
 package org.apache.zeppelin.example.app.clock;
 
+import java.io.IOException;
+import java.text.SimpleDateFormat;
+import java.util.Date;
 import org.apache.zeppelin.helium.Application;
 import org.apache.zeppelin.helium.ApplicationContext;
 import org.apache.zeppelin.helium.ApplicationException;
@@ -24,14 +27,7 @@ import org.apache.zeppelin.resource.*;
 import org.slf4j.Logger;
 import org.slf4j.LoggerFactory;
 
-import java.io.IOException;
-import java.text.SimpleDateFormat;
-import java.util.Date;
-
-/**
- * Basic example application.
- * Get java.util.Date from resource pool and display it
- */
+/** Basic example application. Get java.util.Date from resource pool and display it */
 public class Clock extends Application {
   private final Logger logger = LoggerFactory.getLogger(Clock.class);
 
@@ -60,31 +56,30 @@ public class Clock extends Application {
     }
   }
 
-
   public void start() {
-    updateThread = new Thread() {
-      public void run() {
-        while (!shutdown) {
-          // format date
-          SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
-
-          // put formatted string to angular object.
-          context().getAngularObjectRegistry().add("date", df.format(date));
-
-          try {
-            Thread.sleep(1000);
-          } catch (InterruptedException e) {
-            // nothing todo
+    updateThread =
+        new Thread() {
+          public void run() {
+            while (!shutdown) {
+              // format date
+              SimpleDateFormat df = new SimpleDateFormat("yyyy-MM-dd HH:mm:ss");
+
+              // put formatted string to angular object.
+              context().getAngularObjectRegistry().add("date", df.format(date));
+
+              try {
+                Thread.sleep(1000);
+              } catch (InterruptedException e) {
+                // nothing todo
+              }
+              date = new Date(date.getTime() + 1000);
+            }
           }
-          date = new Date(date.getTime() + 1000);
-        }
-      }
-    };
+        };
 
     updateThread.start();
   }
 
-
   @Override
   public void unload() throws ApplicationException {
     shutdown = true;
@@ -96,16 +91,13 @@ public class Clock extends Application {
     context().getAngularObjectRegistry().remove("date");
   }
 
-  /**
-   * Development mode
-   */
+  /** Development mode */
   public static void main(String[] args) throws Exception {
     LocalResourcePool pool = new LocalResourcePool("dev");
     pool.put("date", new Date());
 
-    ZeppelinApplicationDevServer devServer = new ZeppelinApplicationDevServer(
-        Clock.class.getName(),
-        pool.getAll());
+    ZeppelinApplicationDevServer devServer =
+        new ZeppelinApplicationDevServer(Clock.class.getName(), pool.getAll());
 
     devServer.start();
     devServer.join();