You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by ma...@apache.org on 2014/07/15 23:06:49 UTC
git commit: [SPARK-2474][SQL] For a registered table in
OverrideCatalog,
the Analyzer failed to resolve references in the format of
"tableName.fieldName"
Repository: spark
Updated Branches:
refs/heads/master bcd0c30c7 -> 8af46d584
[SPARK-2474][SQL] For a registered table in OverrideCatalog, the Analyzer failed to resolve references in the format of "tableName.fieldName"
Please refer to JIRA (https://issues.apache.org/jira/browse/SPARK-2474) for how to reproduce the problem and my understanding of the root cause.
Author: Yin Huai <hu...@cse.ohio-state.edu>
Closes #1406 from yhuai/SPARK-2474 and squashes the following commits:
96b1627 [Yin Huai] Merge remote-tracking branch 'upstream/master' into SPARK-2474
af36d65 [Yin Huai] Fix comment.
be86ba9 [Yin Huai] Correct SQL console settings.
c43ad00 [Yin Huai] Wrap the relation in a Subquery named by the table name in OverrideCatalog.lookupRelation.
a5c2145 [Yin Huai] Support sql/console.
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/8af46d58
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/8af46d58
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/8af46d58
Branch: refs/heads/master
Commit: 8af46d58464b96471825ce376c3e11c8b1108c0e
Parents: bcd0c30
Author: Yin Huai <hu...@cse.ohio-state.edu>
Authored: Tue Jul 15 14:06:45 2014 -0700
Committer: Michael Armbrust <mi...@databricks.com>
Committed: Tue Jul 15 14:06:45 2014 -0700
----------------------------------------------------------------------
project/SparkBuild.scala | 24 ++++++++++++++++++++
.../spark/sql/catalyst/analysis/Catalog.scala | 3 ++-
2 files changed, 26 insertions(+), 1 deletion(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/8af46d58/project/SparkBuild.scala
----------------------------------------------------------------------
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index 754d54e..5461d25 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -166,6 +166,9 @@ object SparkBuild extends PomBuild {
/* Enable unidoc only for the root spark project */
enable(Unidoc.settings)(spark)
+ /* Spark SQL Core console settings */
+ enable(SQL.settings)(sql)
+
/* Hive console settings */
enable(Hive.settings)(hive)
@@ -179,6 +182,27 @@ object SparkBuild extends PomBuild {
}
+object SQL {
+
+ lazy val settings = Seq(
+
+ initialCommands in console :=
+ """
+ |import org.apache.spark.sql.catalyst.analysis._
+ |import org.apache.spark.sql.catalyst.dsl._
+ |import org.apache.spark.sql.catalyst.errors._
+ |import org.apache.spark.sql.catalyst.expressions._
+ |import org.apache.spark.sql.catalyst.plans.logical._
+ |import org.apache.spark.sql.catalyst.rules._
+ |import org.apache.spark.sql.catalyst.types._
+ |import org.apache.spark.sql.catalyst.util._
+ |import org.apache.spark.sql.execution
+ |import org.apache.spark.sql.test.TestSQLContext._
+ |import org.apache.spark.sql.parquet.ParquetTestData""".stripMargin
+ )
+
+}
+
object Hive {
lazy val settings = Seq(
http://git-wip-us.apache.org/repos/asf/spark/blob/8af46d58/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala
----------------------------------------------------------------------
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala
index 0d05d98..616f1e2 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/analysis/Catalog.scala
@@ -113,11 +113,12 @@ trait OverrideCatalog extends Catalog {
alias: Option[String] = None): LogicalPlan = {
val (dbName, tblName) = processDatabaseAndTableName(databaseName, tableName)
val overriddenTable = overrides.get((dbName, tblName))
+ val tableWithQualifers = overriddenTable.map(r => Subquery(tblName, r))
// If an alias was specified by the lookup, wrap the plan in a subquery so that attributes are
// properly qualified with this alias.
val withAlias =
- overriddenTable.map(r => alias.map(a => Subquery(a, r)).getOrElse(r))
+ tableWithQualifers.map(r => alias.map(a => Subquery(a, r)).getOrElse(r))
withAlias.getOrElse(super.lookupRelation(dbName, tblName, alias))
}