You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by me...@apache.org on 2022/06/25 02:45:07 UTC

[hudi] 01/01: Revert "[TEST][DO_NOT_MERGE]fix random failed for ci (#5948)"

This is an automated email from the ASF dual-hosted git repository.

mengtao pushed a commit to branch revert-5948-testflaky
in repository https://gitbox.apache.org/repos/asf/hudi.git

commit 5cdf95a03006a0e7c9f923fa468774341e41c8e3
Author: xiarixiaoyao <me...@qq.com>
AuthorDate: Sat Jun 25 10:45:00 2022 +0800

    Revert "[TEST][DO_NOT_MERGE]fix random failed for ci (#5948)"
    
    This reverts commit e8fbd4daf49802f60f800ccc92e66369d44f07f6.
---
 .../src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala      | 7 +++----
 1 file changed, 3 insertions(+), 4 deletions(-)

diff --git a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
index d6ec645920..4160c34b0c 100644
--- a/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
+++ b/hudi-spark-datasource/hudi-spark-common/src/main/scala/org/apache/hudi/BaseFileOnlyRelation.scala
@@ -166,9 +166,7 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
       DataSource.apply(
         sparkSession = sparkSession,
         paths = extraReadPaths,
-        // Here we should specify the schema to the latest commit schema since
-        // the table schema evolution.
-        userSpecifiedSchema = userSchema.orElse(Some(tableStructSchema)),
+        userSpecifiedSchema = userSchema,
         className = formatClassName,
         // Since we're reading the table as just collection of files we have to make sure
         // we only read the latest version of every Hudi's file-group, which might be compacted, clustered, etc.
@@ -177,7 +175,8 @@ class BaseFileOnlyRelation(sqlContext: SQLContext,
         // We rely on [[HoodieROTablePathFilter]], to do proper filtering to assure that
         options = optParams ++ Map(
           "mapreduce.input.pathFilter.class" -> classOf[HoodieROTablePathFilter].getName
-        )
+        ),
+        partitionColumns = partitionColumns
       )
         .resolveRelation()
         .asInstanceOf[HadoopFsRelation]