You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2022/08/26 00:36:06 UTC

[spark] branch master updated: Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"

This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new a590141d0ba Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"
a590141d0ba is described below

commit a590141d0ba0aab2f014e49e4beb77ddf189cda6
Author: Kazuyuki Tanimura <kt...@apple.com>
AuthorDate: Fri Aug 26 09:35:46 2022 +0900

    Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"
    
    ### What changes were proposed in this pull request?
    This PR reverts commit 44f30a04dad2baa471b505f95c6a29992ee7ca72.
    
    ### Why are the changes needed?
    [Based on the follow up discussion](https://github.com/apache/spark/pull/37544), concluded  `JDBCWithAQESuite` is not necessary
    
    ### Does this PR introduce _any_ user-facing change?
    No
    
    ### How was this patch tested?
    Existing tests
    
    Closes #37666 from kazuyukitanimura/SPARK-40110-revert.
    
    Authored-by: Kazuyuki Tanimura <kt...@apple.com>
    Signed-off-by: Hyukjin Kwon <gu...@apache.org>
---
 .../org/apache/spark/sql/jdbc/JDBCSuite.scala      | 32 ++++++----------------
 1 file changed, 9 insertions(+), 23 deletions(-)

diff --git a/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
index 8eda0c288a3..b87fee6cec2 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
@@ -34,7 +34,6 @@ import org.apache.spark.sql.catalyst.parser.CatalystSqlParser
 import org.apache.spark.sql.catalyst.plans.logical.ShowCreateTable
 import org.apache.spark.sql.catalyst.util.{CaseInsensitiveMap, DateTimeTestUtils}
 import org.apache.spark.sql.execution.{DataSourceScanExec, ExtendedMode}
-import org.apache.spark.sql.execution.adaptive.{AdaptiveSparkPlanExec, AdaptiveSparkPlanHelper, DisableAdaptiveExecutionSuite, EnableAdaptiveExecutionSuite}
 import org.apache.spark.sql.execution.command.{ExplainCommand, ShowCreateTableCommand}
 import org.apache.spark.sql.execution.datasources.LogicalRelation
 import org.apache.spark.sql.execution.datasources.jdbc.{JDBCOptions, JDBCPartition, JDBCRelation, JdbcUtils}
@@ -45,8 +44,7 @@ import org.apache.spark.sql.test.SharedSparkSession
 import org.apache.spark.sql.types._
 import org.apache.spark.util.Utils
 
-class JDBCSuite extends QueryTest with SharedSparkSession
-  with AdaptiveSparkPlanHelper with DisableAdaptiveExecutionSuite {
+class JDBCSuite extends QueryTest with SharedSparkSession {
   import testImplicits._
 
   val url = "jdbc:h2:mem:testdb0"
@@ -300,15 +298,10 @@ class JDBCSuite extends QueryTest with SharedSparkSession
     val parentPlan = df.queryExecution.executedPlan
     // Check if SparkPlan Filter is removed in a physical plan and
     // the plan only has PhysicalRDD to scan JDBCRelation.
-    val child = if (df.sqlContext.conf.adaptiveExecutionEnabled) {
-      assert(parentPlan.isInstanceOf[AdaptiveSparkPlanExec])
-      parentPlan.asInstanceOf[AdaptiveSparkPlanExec].executedPlan
-    } else {
-      assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
-      parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec].child
-    }
-    assert(child.isInstanceOf[org.apache.spark.sql.execution.DataSourceScanExec])
-    assert(child.asInstanceOf[DataSourceScanExec].nodeName.contains("JDBCRelation"))
+    assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
+    val node = parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec]
+    assert(node.child.isInstanceOf[org.apache.spark.sql.execution.DataSourceScanExec])
+    assert(node.child.asInstanceOf[DataSourceScanExec].nodeName.contains("JDBCRelation"))
     df
   }
 
@@ -316,14 +309,9 @@ class JDBCSuite extends QueryTest with SharedSparkSession
     val parentPlan = df.queryExecution.executedPlan
     // Check if SparkPlan Filter is not removed in a physical plan because JDBCRDD
     // cannot compile given predicates.
-    val child = if (df.sqlContext.conf.adaptiveExecutionEnabled) {
-      assert(parentPlan.isInstanceOf[AdaptiveSparkPlanExec])
-      parentPlan.asInstanceOf[AdaptiveSparkPlanExec].executedPlan
-    } else {
-      assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
-      parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec].child
-    }
-    assert(child.isInstanceOf[org.apache.spark.sql.execution.FilterExec])
+    assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
+    val node = parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec]
+    assert(node.child.isInstanceOf[org.apache.spark.sql.execution.FilterExec])
     df
   }
 
@@ -1779,7 +1767,7 @@ class JDBCSuite extends QueryTest with SharedSparkSession
 
     def getRowCount(df: DataFrame): Long = {
       val queryExecution = df.queryExecution
-      val rawPlan = collect(queryExecution.executedPlan) {
+      val rawPlan = queryExecution.executedPlan.collect {
         case p: DataSourceScanExec => p
       } match {
         case Seq(p) => p
@@ -1976,5 +1964,3 @@ class JDBCSuite extends QueryTest with SharedSparkSession
     }
   }
 }
-
-class JDBCWithAQESuite extends JDBCSuite with EnableAdaptiveExecutionSuite


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org