You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by me...@apache.org on 2022/06/25 02:45:07 UTC
[hudi] 01/01: Revert "[TEST][DO_NOT_MERGE]fix random failed for ci (#5948)"
This is an automated email from the ASF dual-hosted git repository.
mengtao pushed a commit to branch revert-5948-testflaky
in repository https://gitbox.apache.org/repos/asf/hudi.git
commit 5cdf95a03006a0e7c9f923fa468774341e41c8e3
Author: xiarixiaoyao <me...@qq.com>
AuthorDate: Sat Jun 25 10:45:00 2022 +0800
Revert "[TEST][DO_NOT_MERGE]fix random failed for ci (#5948)"
This reverts commit e8fbd4daf49802f60f800ccc92e66369d44f07f6.
---
.../src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala | 7 +++----
1 file changed, 3 insertions(+), 4 deletions(-)
diff --git a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
index d6ec645920..4160c34b0c 100644
--- a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
+++ b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
@@ -166,9 +166,7 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
DataSource.apply(
sparkSession = sparkSession,
paths = extraReadPaths,
- // Here we should specify the schema to the latest commit schema since
- // the table schema evolution.
- userSpecifiedSchema = userSchema.orElse(Some(tableStructSchema)),
+ userSpecifiedSchema = userSchema,
className = formatClassName,
// Since we're reading the table as just collection of files we have to make sure
// we only read the latest version of every Hudi's file-group, which might be compacted, clustered, etc.
@@ -177,7 +175,8 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
// We rely on [[HoodieROTablePathFilter]], to do proper filtering to assure that
options = optParams ++ Map(
"mapreduce.input.pathFilter.class" -> classOf[HoodieROTablePathFilter].getName
- )
+ ),
+ partitionColumns = partitionColumns
)
.resolveRelation()
.asInstanceOf[HadoopFsRelation]