You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@zeppelin.apache.org by zj...@apache.org on 2020/02/19 05:43:20 UTC

[zeppelin] branch master updated: [hotfix] Fix z.show(df, show_index) in PySpark & PyFlink

This is an automated email from the ASF dual-hosted git repository.

zjffdu pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/zeppelin.git


The following commit(s) were added to refs/heads/master by this push:
     new 8b5bab0  [hotfix] Fix z.show(df, show_index) in PySpark & PyFlink
8b5bab0 is described below

commit 8b5bab08d627feeeb080b73895fc5531332abd79
Author: Jeff Zhang <zj...@apache.org>
AuthorDate: Tue Feb 18 12:33:18 2020 +0800

    [hotfix] Fix z.show(df, show_index) in PySpark & PyFlink
---
 flink/src/main/resources/python/zeppelin_ipyflink.py                | 4 ++--
 flink/src/main/resources/python/zeppelin_pyflink.py                 | 4 ++--
 python/src/main/resources/python/zeppelin_context.py                | 6 +++++-
 spark/interpreter/src/main/resources/python/zeppelin_ipyspark.py    | 4 ++--
 spark/interpreter/src/main/resources/python/zeppelin_pyspark.py     | 4 ++--
 .../src/test/java/org/apache/zeppelin/notebook/ParagraphTest.java   | 4 ++++
 6 files changed, 17 insertions(+), 9 deletions(-)

diff --git a/flink/src/main/resources/python/zeppelin_ipyflink.py b/flink/src/main/resources/python/zeppelin_ipyflink.py
index c6dfa47..3752a4f 100644
--- a/flink/src/main/resources/python/zeppelin_ipyflink.py
+++ b/flink/src/main/resources/python/zeppelin_ipyflink.py
@@ -54,11 +54,11 @@ class IPyFlinkZeppelinContext(PyZeppelinContext):
     def __init__(self, z, gateway):
         super(IPyFlinkZeppelinContext, self).__init__(z, gateway)
 
-    def show(self, obj):
+    def show(self, obj, **kwargs):
         from pyflink.table import Table
         if isinstance(obj, Table):
             print(self.z.showData(obj._j_table))
         else:
-            super(IPyFlinkZeppelinContext, self).show(obj)
+            super(IPyFlinkZeppelinContext, self).show(obj, **kwargs)
 
 z = __zeppelin__ = IPyFlinkZeppelinContext(intp.getZeppelinContext(), gateway)
diff --git a/flink/src/main/resources/python/zeppelin_pyflink.py b/flink/src/main/resources/python/zeppelin_pyflink.py
index 86e1a50..a6bd4e9 100644
--- a/flink/src/main/resources/python/zeppelin_pyflink.py
+++ b/flink/src/main/resources/python/zeppelin_pyflink.py
@@ -46,12 +46,12 @@ class PyFlinkZeppelinContext(PyZeppelinContext):
   def __init__(self, z, gateway):
     super(PyFlinkZeppelinContext, self).__init__(z, gateway)
 
-  def show(self, obj):
+  def show(self, obj, **kwargs):
     from pyflink.table import Table
     if isinstance(obj, Table):
       print(self.z.showData(obj._j_table))
     else:
-      super(PyFlinkZeppelinContext, self).show(obj)
+      super(PyFlinkZeppelinContext, self).show(obj, **kwargs)
 
 z = __zeppelin__ = PyFlinkZeppelinContext(intp.getZeppelinContext(), gateway)
 __zeppelin__._setup_matplotlib()
diff --git a/python/src/main/resources/python/zeppelin_context.py b/python/src/main/resources/python/zeppelin_context.py
index 6d4263e..f4cf269 100644
--- a/python/src/main/resources/python/zeppelin_context.py
+++ b/python/src/main/resources/python/zeppelin_context.py
@@ -189,9 +189,13 @@ class PyZeppelinContext(object):
     def normalizeColumn(self, column):
         return column.replace("\t", " ").replace("\r\n", " ").replace("\n", " ")
 
-    def show_dataframe(self, df, show_index=False, **kwargs):
+    def show_dataframe(self, df, **kwargs):
         """Pretty prints DF using Table Display System
         """
+        show_index = False
+        if 'show_index' in kwargs:
+            show_index = kwargs['show_index']
+
         exceed_limit = len(df) > self.max_result
         header_buf = StringIO("")
         if show_index:
diff --git a/spark/interpreter/src/main/resources/python/zeppelin_ipyspark.py b/spark/interpreter/src/main/resources/python/zeppelin_ipyspark.py
index 07aad9b..5d199bc 100644
--- a/spark/interpreter/src/main/resources/python/zeppelin_ipyspark.py
+++ b/spark/interpreter/src/main/resources/python/zeppelin_ipyspark.py
@@ -61,11 +61,11 @@ class IPySparkZeppelinContext(PyZeppelinContext):
     def __init__(self, z, gateway):
         super(IPySparkZeppelinContext, self).__init__(z, gateway)
 
-    def show(self, obj):
+    def show(self, obj, **kwargs):
         from pyspark.sql import DataFrame
         if isinstance(obj, DataFrame):
             print(self.z.showData(obj._jdf))
         else:
-            super(IPySparkZeppelinContext, self).show(obj)
+            super(IPySparkZeppelinContext, self).show(obj, **kwargs)
 
 z = __zeppelin__ = IPySparkZeppelinContext(intp.getZeppelinContext(), gateway)
diff --git a/spark/interpreter/src/main/resources/python/zeppelin_pyspark.py b/spark/interpreter/src/main/resources/python/zeppelin_pyspark.py
index 1917349..7132b8a 100644
--- a/spark/interpreter/src/main/resources/python/zeppelin_pyspark.py
+++ b/spark/interpreter/src/main/resources/python/zeppelin_pyspark.py
@@ -57,12 +57,12 @@ class PySparkZeppelinContext(PyZeppelinContext):
   def __init__(self, z, gateway):
     super(PySparkZeppelinContext, self).__init__(z, gateway)
 
-  def show(self, obj):
+  def show(self, obj, **kwargs):
     from pyspark.sql import DataFrame
     if isinstance(obj, DataFrame):
       print(self.z.showData(obj._jdf))
     else:
-      super(PySparkZeppelinContext, self).show(obj)
+      super(PySparkZeppelinContext, self).show(obj, **kwargs)
 
 z = __zeppelin__ = PySparkZeppelinContext(intp.getZeppelinContext(), gateway)
 __zeppelin__._setup_matplotlib()
diff --git a/zeppelin-zengine/src/test/java/org/apache/zeppelin/notebook/ParagraphTest.java b/zeppelin-zengine/src/test/java/org/apache/zeppelin/notebook/ParagraphTest.java
index 03f428e..5dd6301 100644
--- a/zeppelin-zengine/src/test/java/org/apache/zeppelin/notebook/ParagraphTest.java
+++ b/zeppelin-zengine/src/test/java/org/apache/zeppelin/notebook/ParagraphTest.java
@@ -307,6 +307,10 @@ public class ParagraphTest extends AbstractInterpreterTest {
     spyParagraph.jobRun();
     Paragraph p1 = spyParagraph.getUserParagraph(user1.getUser());
 
+    mockInterpreterResult = mock(InterpreterResult.class);
+    when(mockInterpreter.interpret(anyString(), Mockito.<InterpreterContext>any())).thenReturn(mockInterpreterResult);
+    when(mockInterpreterResult.code()).thenReturn(Code.SUCCESS);
+
     List<InterpreterResultMessage> result2 = Lists.newArrayList();
     result2.add(new InterpreterResultMessage(Type.TEXT, "result2"));
     when(mockInterpreterResult.message()).thenReturn(result2);