You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by rx...@apache.org on 2015/04/27 06:08:53 UTC
spark git commit: [SQL][Minor] rename DataTypeParser.apply to
DataTypeParser.parse
Repository: spark
Updated Branches:
refs/heads/master ca55dc95b -> d188b8bad
[SQL][Minor] rename DataTypeParser.apply to DataTypeParser.parse
rename DataTypeParser.apply to DataTypeParser.parse to make it more clear and readable.
/cc rxin
Author: wangfei <wa...@huawei.com>
Closes #5710 from scwf/apply and squashes the following commits:
c319977 [wangfei] rename apply to parse
Project: http://git-wip-us.apache.org/repos/asf/spark/repo
Commit: http://git-wip-us.apache.org/repos/asf/spark/commit/d188b8ba
Tree: http://git-wip-us.apache.org/repos/asf/spark/tree/d188b8ba
Diff: http://git-wip-us.apache.org/repos/asf/spark/diff/d188b8ba
Branch: refs/heads/master
Commit: d188b8bad82836bf654e57f9dd4e1ddde1d530f4
Parents: ca55dc9
Author: wangfei <wa...@huawei.com>
Authored: Sun Apr 26 21:08:47 2015 -0700
Committer: Reynold Xin <rx...@databricks.com>
Committed: Sun Apr 26 21:08:47 2015 -0700
----------------------------------------------------------------------
.../scala/org/apache/spark/sql/catalyst/planning/patterns.scala | 2 +-
.../main/scala/org/apache/spark/sql/types/DataTypeParser.scala | 2 +-
.../scala/org/apache/spark/sql/types/DataTypeParserSuite.scala | 4 ++--
sql/core/src/main/scala/org/apache/spark/sql/Column.scala | 2 +-
.../scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala | 2 +-
5 files changed, 6 insertions(+), 6 deletions(-)
----------------------------------------------------------------------
http://git-wip-us.apache.org/repos/asf/spark/blob/d188b8ba/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
----------------------------------------------------------------------
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
index 9c8c643..4574934 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/planning/patterns.scala
@@ -92,7 +92,7 @@ object PhysicalOperation extends PredicateHelper {
}
def collectAliases(fields: Seq[Expression]): Map[Attribute, Expression] = fields.collect {
- case a @ Alias(child, _) => a.toAttribute.asInstanceOf[Attribute] -> child
+ case a @ Alias(child, _) => a.toAttribute -> child
}.toMap
def substitute(aliases: Map[Attribute, Expression])(expr: Expression): Expression = {
http://git-wip-us.apache.org/repos/asf/spark/blob/d188b8ba/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataTypeParser.scala
----------------------------------------------------------------------
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataTypeParser.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataTypeParser.scala
index 5163f05..04f3379 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataTypeParser.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DataTypeParser.scala
@@ -108,7 +108,7 @@ private[sql] object DataTypeParser {
override val lexical = new SqlLexical
}
- def apply(dataTypeString: String): DataType = dataTypeParser.toDataType(dataTypeString)
+ def parse(dataTypeString: String): DataType = dataTypeParser.toDataType(dataTypeString)
}
/** The exception thrown from the [[DataTypeParser]]. */
http://git-wip-us.apache.org/repos/asf/spark/blob/d188b8ba/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DataTypeParserSuite.scala
----------------------------------------------------------------------
diff --git a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DataTypeParserSuite.scala b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DataTypeParserSuite.scala
index 1691252..3e7cf7c 100644
--- a/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DataTypeParserSuite.scala
+++ b/sql/catalyst/src/test/scala/org/apache/spark/sql/types/DataTypeParserSuite.scala
@@ -23,13 +23,13 @@ class DataTypeParserSuite extends FunSuite {
def checkDataType(dataTypeString: String, expectedDataType: DataType): Unit = {
test(s"parse ${dataTypeString.replace("\n", "")}") {
- assert(DataTypeParser(dataTypeString) === expectedDataType)
+ assert(DataTypeParser.parse(dataTypeString) === expectedDataType)
}
}
def unsupported(dataTypeString: String): Unit = {
test(s"$dataTypeString is not supported") {
- intercept[DataTypeException](DataTypeParser(dataTypeString))
+ intercept[DataTypeException](DataTypeParser.parse(dataTypeString))
}
}
http://git-wip-us.apache.org/repos/asf/spark/blob/d188b8ba/sql/core/src/main/scala/org/apache/spark/sql/Column.scala
----------------------------------------------------------------------
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/Column.scala b/sql/core/src/main/scala/org/apache/spark/sql/Column.scala
index edb229c..33f9d0b 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/Column.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/Column.scala
@@ -647,7 +647,7 @@ class Column(protected[sql] val expr: Expression) extends Logging {
*
* @group expr_ops
*/
- def cast(to: String): Column = cast(DataTypeParser(to))
+ def cast(to: String): Column = cast(DataTypeParser.parse(to))
/**
* Returns an ordering used in sorting.
http://git-wip-us.apache.org/repos/asf/spark/blob/d188b8ba/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
----------------------------------------------------------------------
diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
index f1c0bd9..4d222cf 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
@@ -871,7 +871,7 @@ private[hive] case class MetastoreRelation
private[hive] object HiveMetastoreTypes {
- def toDataType(metastoreType: String): DataType = DataTypeParser(metastoreType)
+ def toDataType(metastoreType: String): DataType = DataTypeParser.parse(metastoreType)
def toMetastoreType(dt: DataType): String = dt match {
case ArrayType(elementType, _) => s"array<${toMetastoreType(elementType)}>"
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org