You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by gu...@apache.org on 2022/08/26 00:36:06 UTC
[spark] branch master updated: Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"
This is an automated email from the ASF dual-hosted git repository.
gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new a590141d0ba Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"
a590141d0ba is described below
commit a590141d0ba0aab2f014e49e4beb77ddf189cda6
Author: Kazuyuki Tanimura <kt...@apple.com>
AuthorDate: Fri Aug 26 09:35:46 2022 +0900
Revert "[SPARK-40110][SQL][TESTS] Add JDBCWithAQESuite"
### What changes were proposed in this pull request?
This PR reverts commit 44f30a04dad2baa471b505f95c6a29992ee7ca72.
### Why are the changes needed?
[Based on the follow up discussion](https://github.com/apache/spark/pull/37544), concluded `JDBCWithAQESuite` is not necessary
### Does this PR introduce _any_ user-facing change?
No
### How was this patch tested?
Existing tests
Closes #37666 from kazuyukitanimura/SPARK-40110-revert.
Authored-by: Kazuyuki Tanimura <kt...@apple.com>
Signed-off-by: Hyukjin Kwon <gu...@apache.org>
---
.../org/apache/spark/sql/jdbc/JDBCSuite.scala | 32 ++++++----------------
1 file changed, 9 insertions(+), 23 deletions(-)
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
index 8eda0c288a3..b87fee6cec2 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/jdbc/JDBCSuite.scala
@@ -34,7 +34,6 @@ import org.apache.spark.sql.catalyst.parser.CatalystSqlParser
import org.apache.spark.sql.catalyst.plans.logical.ShowCreateTable
import org.apache.spark.sql.catalyst.util.{CaseInsensitiveMap, DateTimeTestUtils}
import org.apache.spark.sql.execution.{DataSourceScanExec, ExtendedMode}
-import org.apache.spark.sql.execution.adaptive.{AdaptiveSparkPlanExec, AdaptiveSparkPlanHelper, DisableAdaptiveExecutionSuite, EnableAdaptiveExecutionSuite}
import org.apache.spark.sql.execution.command.{ExplainCommand, ShowCreateTableCommand}
import org.apache.spark.sql.execution.datasources.LogicalRelation
import org.apache.spark.sql.execution.datasources.jdbc.{JDBCOptions, JDBCPartition, JDBCRelation, JdbcUtils}
@@ -45,8 +44,7 @@ import org.apache.spark.sql.test.SharedSparkSession
import org.apache.spark.sql.types._
import org.apache.spark.util.Utils
-class JDBCSuite extends QueryTest with SharedSparkSession
- with AdaptiveSparkPlanHelper with DisableAdaptiveExecutionSuite {
+class JDBCSuite extends QueryTest with SharedSparkSession {
import testImplicits._
val url = "jdbc:h2:mem:testdb0"
@@ -300,15 +298,10 @@ class JDBCSuite extends QueryTest with SharedSparkSession
val parentPlan = df.queryExecution.executedPlan
// Check if SparkPlan Filter is removed in a physical plan and
// the plan only has PhysicalRDD to scan JDBCRelation.
- val child = if (df.sqlContext.conf.adaptiveExecutionEnabled) {
- assert(parentPlan.isInstanceOf[AdaptiveSparkPlanExec])
- parentPlan.asInstanceOf[AdaptiveSparkPlanExec].executedPlan
- } else {
- assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
- parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec].child
- }
- assert(child.isInstanceOf[org.apache.spark.sql.execution.DataSourceScanExec])
- assert(child.asInstanceOf[DataSourceScanExec].nodeName.contains("JDBCRelation"))
+ assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
+ val node = parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec]
+ assert(node.child.isInstanceOf[org.apache.spark.sql.execution.DataSourceScanExec])
+ assert(node.child.asInstanceOf[DataSourceScanExec].nodeName.contains("JDBCRelation"))
df
}
@@ -316,14 +309,9 @@ class JDBCSuite extends QueryTest with SharedSparkSession
val parentPlan = df.queryExecution.executedPlan
// Check if SparkPlan Filter is not removed in a physical plan because JDBCRDD
// cannot compile given predicates.
- val child = if (df.sqlContext.conf.adaptiveExecutionEnabled) {
- assert(parentPlan.isInstanceOf[AdaptiveSparkPlanExec])
- parentPlan.asInstanceOf[AdaptiveSparkPlanExec].executedPlan
- } else {
- assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
- parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec].child
- }
- assert(child.isInstanceOf[org.apache.spark.sql.execution.FilterExec])
+ assert(parentPlan.isInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec])
+ val node = parentPlan.asInstanceOf[org.apache.spark.sql.execution.WholeStageCodegenExec]
+ assert(node.child.isInstanceOf[org.apache.spark.sql.execution.FilterExec])
df
}
@@ -1779,7 +1767,7 @@ class JDBCSuite extends QueryTest with SharedSparkSession
def getRowCount(df: DataFrame): Long = {
val queryExecution = df.queryExecution
- val rawPlan = collect(queryExecution.executedPlan) {
+ val rawPlan = queryExecution.executedPlan.collect {
case p: DataSourceScanExec => p
} match {
case Seq(p) => p
@@ -1976,5 +1964,3 @@ class JDBCSuite extends QueryTest with SharedSparkSession
}
}
}
-
-class JDBCWithAQESuite extends JDBCSuite with EnableAdaptiveExecutionSuite
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org