You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by xu...@apache.org on 2022/07/25 04:41:47 UTC

[hudi] branch master updated: [MINOR] Fix typos in Spark client related classes (#6204)

This is an automated email from the ASF dual-hosted git repository.

xushiyan pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/master by this push:
     new 2a08a65f71 [MINOR] Fix typos in Spark client related classes (#6204)
2a08a65f71 is described below

commit 2a08a65f719b5c155dde85a0dc318af5033c31d5
Author: Vander <30...@users.noreply.github.com>
AuthorDate: Mon Jul 25 12:41:42 2022 +0800

    [MINOR] Fix typos in Spark client related classes (#6204)
---
 .../clustering/run/strategy/SingleSparkJobExecutionStrategy.java    | 2 +-
 .../org/apache/hudi/client/utils/SparkInternalSchemaConverter.java  | 4 ++--
 .../main/java/org/apache/hudi/client/utils/SparkValidatorUtils.java | 2 +-
 .../org/apache/hudi/common/table/log/block/HoodieAvroDataBlock.java | 6 +++---
 4 files changed, 7 insertions(+), 7 deletions(-)

diff --git a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/SingleSparkJobExecutionStrategy.java b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/SingleSparkJobExecutionStrategy.java
index 1158d0ada4..bb6d3df5f1 100644
--- a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/SingleSparkJobExecutionStrategy.java
+++ b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/clustering/run/strategy/SingleSparkJobExecutionStrategy.java
@@ -136,7 +136,7 @@ public abstract class SingleSparkJobExecutionStrategy<T extends HoodieRecordPayl
   /**
    * Execute clustering to write inputRecords into new files as defined by rules in strategy parameters.
    * The number of new file groups created is bounded by numOutputGroups.
-   * Note that commit is not done as part of strategy. commit is callers responsibility.
+   * Note that commit is not done as part of strategy. Commit is callers responsibility.
    */
   public abstract Iterator<List<WriteStatus>> performClusteringWithRecordsIterator(final Iterator<HoodieRecord<T>> records, final int numOutputGroups,
                                                                                    final String instantTime,
diff --git a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkInternalSchemaConverter.java b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkInternalSchemaConverter.java
index 8e086c2927..098870a60a 100644
--- a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkInternalSchemaConverter.java
+++ b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkInternalSchemaConverter.java
@@ -81,7 +81,7 @@ public class SparkInternalSchemaConverter {
   public static final String HOODIE_VALID_COMMITS_LIST = "hoodie.valid.commits.list";
 
   /**
-   * Converts a spark schema to an hudi internal schema. Fields without IDs are kept and assigned fallback IDs.
+   * Convert a spark schema to an hudi internal schema. Fields without IDs are kept and assigned fallback IDs.
    *
    * @param sparkSchema a spark schema
    * @return a matching internal schema for the provided spark schema
@@ -157,7 +157,7 @@ public class SparkInternalSchemaConverter {
   }
 
   /**
-   * Converts Spark schema to Hudi internal schema, and prune fields.
+   * Convert Spark schema to Hudi internal schema, and prune fields.
    * Fields without IDs are kept and assigned fallback IDs.
    *
    * @param sparkSchema a pruned spark schema
diff --git a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkValidatorUtils.java b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkValidatorUtils.java
index fd083f2c89..a6d03eae2b 100644
--- a/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkValidatorUtils.java
+++ b/hudi-client/hudi-spark-client/src/main/java/org/apache/hudi/client/utils/SparkValidatorUtils.java
@@ -50,7 +50,7 @@ import java.util.stream.Stream;
 import scala.collection.JavaConverters;
 
 /**
- * Spark validator utils to verify and run any precommit validators configured.
+ * Spark validator utils to verify and run any pre-commit validators configured.
  */
 public class SparkValidatorUtils {
   private static final Logger LOG = LogManager.getLogger(BaseSparkCommitActionExecutor.class);
diff --git a/hudi-common/src/main/java/org/apache/hudi/common/table/log/block/HoodieAvroDataBlock.java b/hudi-common/src/main/java/org/apache/hudi/common/table/log/block/HoodieAvroDataBlock.java
index 491c6700c9..9e74d14c04 100644
--- a/hudi-common/src/main/java/org/apache/hudi/common/table/log/block/HoodieAvroDataBlock.java
+++ b/hudi-common/src/main/java/org/apache/hudi/common/table/log/block/HoodieAvroDataBlock.java
@@ -308,7 +308,7 @@ public class HoodieAvroDataBlock extends HoodieDataBlock {
     ByteArrayOutputStream baos = new ByteArrayOutputStream();
     DataOutputStream output = new DataOutputStream(baos);
 
-    // 2. Compress and Write schema out
+    // 1. Compress and Write schema out
     byte[] schemaContent = compress(schema.toString());
     output.writeInt(schemaContent.length);
     output.write(schemaContent);
@@ -318,10 +318,10 @@ public class HoodieAvroDataBlock extends HoodieDataBlock {
       recordItr.forEachRemaining(records::add);
     }
 
-    // 3. Write total number of records
+    // 2. Write total number of records
     output.writeInt(records.size());
 
-    // 4. Write the records
+    // 3. Write the records
     Iterator<IndexedRecord> itr = records.iterator();
     while (itr.hasNext()) {
       IndexedRecord s = itr.next();