You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2020/05/28 01:28:16 UTC

[spark] branch master updated: [SPARK-25351][PYTHON][TEST][FOLLOWUP] Fix test assertions to be consistent

This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new 8bbb666  [SPARK-25351][PYTHON][TEST][FOLLOWUP] Fix test assertions to be consistent
8bbb666 is described below

commit 8bbb666622e042c1533da294ac7b504b6aaa694a
Author: Bryan Cutler <cu...@gmail.com>
AuthorDate: Thu May 28 10:27:15 2020 +0900

    [SPARK-25351][PYTHON][TEST][FOLLOWUP] Fix test assertions to be consistent
    
    ### What changes were proposed in this pull request?
    Followup to make assertions from recent test consistent with the rest of the module
    
    ### Why are the changes needed?
    
    Better to use assertions from `unittest` and be consistent
    
    ### Does this PR introduce _any_ user-facing change?
    
    No
    
    ### How was this patch tested?
    
    Existing tests
    
    Closes #28659 from BryanCutler/arrow-category-test-fix-SPARK-25351.
    
    Authored-by: Bryan Cutler <cu...@gmail.com>
    Signed-off-by: HyukjinKwon <gu...@apache.org>
---
 python/pyspark/sql/tests/test_arrow.py             | 9 +++++----
 python/pyspark/sql/tests/test_pandas_udf_scalar.py | 9 +++------
 2 files changed, 8 insertions(+), 10 deletions(-)

diff --git a/python/pyspark/sql/tests/test_arrow.py b/python/pyspark/sql/tests/test_arrow.py
index c3c9fb0..c59765d 100644
--- a/python/pyspark/sql/tests/test_arrow.py
+++ b/python/pyspark/sql/tests/test_arrow.py
@@ -435,11 +435,12 @@ class ArrowTests(ReusedSQLTestCase):
         assert_frame_equal(result_spark, result_arrow)
 
         # ensure original category elements are string
-        assert isinstance(category_first_element, str)
+        self.assertIsInstance(category_first_element, str)
         # spark data frame and arrow execution mode enabled data frame type must match pandas
-        assert spark_type == arrow_type == 'string'
-        assert isinstance(arrow_first_category_element, str)
-        assert isinstance(spark_first_category_element, str)
+        self.assertEqual(spark_type, 'string')
+        self.assertEqual(arrow_type, 'string')
+        self.assertIsInstance(arrow_first_category_element, str)
+        self.assertIsInstance(spark_first_category_element, str)
 
 
 @unittest.skipIf(
diff --git a/python/pyspark/sql/tests/test_pandas_udf_scalar.py b/python/pyspark/sql/tests/test_pandas_udf_scalar.py
index ae6b8d5..2d38efd 100644
--- a/python/pyspark/sql/tests/test_pandas_udf_scalar.py
+++ b/python/pyspark/sql/tests/test_pandas_udf_scalar.py
@@ -910,13 +910,10 @@ class ScalarPandasUDFTests(ReusedSQLTestCase):
 
         spark_type = df.dtypes[1][1]
         # spark data frame and arrow execution mode enabled data frame type must match pandas
-        assert spark_type == 'string'
+        self.assertEqual(spark_type, 'string')
 
-        # Check result value of column 'B' must be equal to column 'A'
-        for i in range(0, len(result_spark["A"])):
-            assert result_spark["A"][i] == result_spark["B"][i]
-            assert isinstance(result_spark["A"][i], str)
-            assert isinstance(result_spark["B"][i], str)
+        # Check result of column 'B' must be equal to column 'A' in type and values
+        pd.testing.assert_series_equal(result_spark["A"], result_spark["B"], check_names=False)
 
     @unittest.skipIf(sys.version_info[:2] < (3, 5), "Type hints are supported from Python 3.5.")
     def test_type_annotation(self):


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org