You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2016/12/19 19:33:50 UTC

spark git commit: Fix test case for SubquerySuite.

Repository: spark
Updated Branches:
  refs/heads/branch-2.0 b41668349 -> 2a5ab1490


Fix test case for SubquerySuite.


Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/2a5ab149
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/2a5ab149
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/2a5ab149

Branch: refs/heads/branch-2.0
Commit: 2a5ab149005f305b7bfcb160c5023e1a83d2dce1
Parents: b416683
Author: Reynold Xin <rx...@databricks.com>
Authored: Mon Dec 19 11:33:31 2016 -0800
Committer: Reynold Xin <rx...@databricks.com>
Committed: Mon Dec 19 11:33:31 2016 -0800

----------------------------------------------------------------------
 sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)
----------------------------------------------------------------------


http://git-wip-us.apache.org/repos/asf/spark/blob/2a5ab149/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
index eff8894..dfbfa3f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
@@ -59,7 +59,8 @@ class SubquerySuite extends QueryTest with SharedSQLContext {
       "(select id from range(2) union all select id from range(2))")
 
     // The depth first traversal of the plan tree
-    val dfs = Seq("Project", "Filter", "Union", "Project", "Range", "Project", "Range", "Range")
+    val dfs = Seq("Project", "Filter", "SubqueryAlias", "Union", "Project", "Range", "Project",
+      "Range", "Range")
     val numbered = df.queryExecution.analyzed.numberedTreeString.split("\n")
 
     // There should be 8 plan nodes in total


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org