You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@spark.apache.org by do...@apache.org on 2021/02/26 20:20:14 UTC
[spark] branch branch-3.0 updated: Revert "[SPARK-34543][SQL]
Respect the `spark.sql.caseSensitive` config while resolving partition spec
in v1 `SET LOCATION`"
This is an automated email from the ASF dual-hosted git repository.
dongjoon pushed a commit to branch branch-3.0
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/branch-3.0 by this push:
new baa2e6d Revert "[SPARK-34543][SQL] Respect the `spark.sql.caseSensitive` config while resolving partition spec in v1 `SET LOCATION`"
baa2e6d is described below
commit baa2e6d44c06de4390246187ab83ef6165902e82
Author: Dongjoon Hyun <dh...@apple.com>
AuthorDate: Fri Feb 26 12:15:53 2021 -0800
Revert "[SPARK-34543][SQL] Respect the `spark.sql.caseSensitive` config while resolving partition spec in v1 `SET LOCATION`"
This reverts commit 7b532922bb72a8fc9ba933673f055763cedfbfa4.
---
.../scala/org/apache/spark/sql/execution/command/ddl.scala | 7 +------
.../org/apache/spark/sql/execution/command/DDLSuite.scala | 11 -----------
2 files changed, 1 insertion(+), 17 deletions(-)
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala
index 01e10bb..a70d32e 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/command/ddl.scala
@@ -825,12 +825,7 @@ case class AlterTableSetLocationCommand(
DDLUtils.verifyPartitionProviderIsHive(
sparkSession, table, "ALTER TABLE ... SET LOCATION")
// Partition spec is specified, so we set the location only for this partition
- val normalizedSpec = PartitioningUtils.normalizePartitionSpec(
- spec,
- table.partitionSchema,
- table.identifier.quotedString,
- sparkSession.sessionState.conf.resolver)
- val part = catalog.getPartition(table.identifier, normalizedSpec)
+ val part = catalog.getPartition(table.identifier, spec)
val newPart = part.copy(storage = part.storage.copy(locationUri = Some(locUri)))
catalog.alterPartitions(table.identifier, Seq(newPart))
case None =>
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
index ee8f211..21d7acf 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
@@ -1479,17 +1479,6 @@ abstract class DDLSuite extends QueryTest with SQLTestUtils {
// set table partition location
sql("ALTER TABLE dbx.tab1 PARTITION (a='1', b='2') SET LOCATION '/path/to/part/ways'")
verifyLocation(new URI("/path/to/part/ways"), Some(partSpec))
- // set location for partition spec in the upper case
- withSQLConf(SQLConf.CASE_SENSITIVE.key -> "false") {
- sql("ALTER TABLE dbx.tab1 PARTITION (A='1', B='2') SET LOCATION '/path/to/part/ways2'")
- verifyLocation(new URI("/path/to/part/ways2"), Some(partSpec))
- }
- withSQLConf(SQLConf.CASE_SENSITIVE.key -> "true") {
- val errMsg = intercept[AnalysisException] {
- sql("ALTER TABLE dbx.tab1 PARTITION (A='1', B='2') SET LOCATION '/path/to/part/ways3'")
- }.getMessage
- assert(errMsg.contains("not a valid partition column"))
- }
// set table location without explicitly specifying database
catalog.setCurrentDatabase("dbx")
sql("ALTER TABLE tab1 SET LOCATION '/swanky/steak/place'")
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@spark.apache.org
For additional commands, e-mail: commits-help@spark.apache.org