You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by ud...@apache.org on 2021/08/19 20:28:01 UTC
[hudi] branch release-0.9.0 updated: [HUDI-2322] Use correct meta
columns while preparing dataset for bulk insert (#3504)
This is an automated email from the ASF dual-hosted git repository.
uditme pushed a commit to branch release-0.9.0
in repository https://gitbox.apache.org/repos/asf/hudi.git
The following commit(s) were added to refs/heads/release-0.9.0 by this push:
new 73c0e1e [HUDI-2322] Use correct meta columns while preparing dataset for bulk insert (#3504)
73c0e1e is described below
commit 73c0e1e4432d97958ab1fec6ce9350ecd5bc656b
Author: Sagar Sumit <sa...@gmail.com>
AuthorDate: Thu Aug 19 21:37:12 2021 +0530
[HUDI-2322] Use correct meta columns while preparing dataset for bulk insert (#3504)
---
.../main/java/org/apache/hudi/HoodieDatasetBulkInsertHelper.java | 8 ++++++--
1 file changed, 6 insertions(+), 2 deletions(-)
diff --git a/hudi-spark-datasource/hudi-spark/src/main/java/org/apache/hudi/HoodieDatasetBulkInsertHelper.java b/hudi-spark-datasource/hudi-spark/src/main/java/org/apache/hudi/HoodieDatasetBulkInsertHelper.java
index 0ccd33d..47e97aa 100644
--- a/hudi-spark-datasource/hudi-spark/src/main/java/org/apache/hudi/HoodieDatasetBulkInsertHelper.java
+++ b/hudi-spark-datasource/hudi-spark/src/main/java/org/apache/hudi/HoodieDatasetBulkInsertHelper.java
@@ -145,10 +145,14 @@ public class HoodieDatasetBulkInsertHelper {
functions.lit("").cast(DataTypes.StringType));
List<Column> originalFields =
- Arrays.stream(rowsWithMetaCols.schema().fields()).filter(field -> !field.name().contains("_hoodie_")).map(f -> new Column(f.name())).collect(Collectors.toList());
+ Arrays.stream(rowsWithMetaCols.schema().fields())
+ .filter(field -> !HoodieRecord.HOODIE_META_COLUMNS_WITH_OPERATION.contains(field.name()))
+ .map(f -> new Column(f.name())).collect(Collectors.toList());
List<Column> metaFields =
- Arrays.stream(rowsWithMetaCols.schema().fields()).filter(field -> field.name().contains("_hoodie_")).map(f -> new Column(f.name())).collect(Collectors.toList());
+ Arrays.stream(rowsWithMetaCols.schema().fields())
+ .filter(field -> HoodieRecord.HOODIE_META_COLUMNS_WITH_OPERATION.contains(field.name()))
+ .map(f -> new Column(f.name())).collect(Collectors.toList());
// reorder such that all meta columns are at the beginning followed by original columns
List<Column> allCols = new ArrayList<>();