You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by bh...@apache.org on 2023/01/12 16:56:11 UTC

[hudi] branch asf-site updated: Update docs (#7639)

This is an automated email from the ASF dual-hosted git repository.

bhavanisudha pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new 87e30a06e7a Update docs (#7639)
87e30a06e7a is described below

commit 87e30a06e7a48f27ea3f2ba7f49a931b662e8c57
Author: nfarah86 <nf...@gmail.com>
AuthorDate: Thu Jan 12 08:56:02 2023 -0800

    Update docs (#7639)
    
    * updated deltastreamer docs and writing_data for class update
    
    * updated comparison blog
    
    Co-authored-by: nadine <nf...@nadines-MacBook-Pro.local>
---
 ...-vs-Delta-Lake-vs-Apache-Iceberg-Lakehouse-Feature-Comparison.mdx} | 0
 website/docs/hoodie_deltastreamer.md                                  | 3 ++-
 website/versioned_docs/version-0.10.0/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.10.1/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.11.0/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.11.1/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.12.0/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.12.1/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.12.2/writing_data.md                 | 4 ++--
 website/versioned_docs/version-0.5.3/writing_data.md                  | 4 ++--
 website/versioned_docs/version-0.6.0/writing_data.md                  | 4 ++--
 website/versioned_docs/version-0.7.0/writing_data.md                  | 4 ++--
 website/versioned_docs/version-0.8.0/writing_data.md                  | 4 ++--
 website/versioned_docs/version-0.9.0/writing_data.md                  | 4 ++--
 14 files changed, 26 insertions(+), 25 deletions(-)

diff --git a/website/blog/2022-08-18-Apache-Hudi-vs-Delta-Lake-vs-Apache-Iceberg-Lakehouse-Feature-Comparison.mdx b/website/blog/2023-01-11-Apache-Hudi-vs-Delta-Lake-vs-Apache-Iceberg-Lakehouse-Feature-Comparison.mdx
similarity index 100%
rename from website/blog/2022-08-18-Apache-Hudi-vs-Delta-Lake-vs-Apache-Iceberg-Lakehouse-Feature-Comparison.mdx
rename to website/blog/2023-01-11-Apache-Hudi-vs-Delta-Lake-vs-Apache-Iceberg-Lakehouse-Feature-Comparison.mdx
diff --git a/website/docs/hoodie_deltastreamer.md b/website/docs/hoodie_deltastreamer.md
index 42b98a2b90e..12f030fff43 100644
--- a/website/docs/hoodie_deltastreamer.md
+++ b/website/docs/hoodie_deltastreamer.md
@@ -158,13 +158,14 @@ and then ingest it as follows.
 
 In some cases, you may want to migrate your existing table into Hudi beforehand. Please refer to [migration guide](/docs/migration_guide).
 
+### Note on hudi utilities bundle usage for different spark versions
 From 0.11.0 release, we start to provide a new `hudi-utilities-slim-bundle` which aims to exclude dependencies that can
 cause conflicts and compatibility issues with different versions of Spark.  The `hudi-utilities-slim-bundle` should be
 used along with a Hudi Spark bundle corresponding the Spark version used to make utilities work with Spark, e.g.,
 `--packages org.apache.hudi:hudi-utilities-slim-bundle_2.12:0.12.1,org.apache.hudi:hudi-spark3.1-bundle_2.12:0.12.1`,
 if using `hudi-utilities-bundle` solely to run `HoodieDeltaStreamer` in Spark encounters compatibility issues.
 
-### MultiTableDeltaStreamer
+#### MultiTableDeltaStreamer
 
 `HoodieMultiTableDeltaStreamer`, a wrapper on top of `HoodieDeltaStreamer`, enables one to ingest multiple tables at a single go into hudi datasets. Currently it only supports sequential processing of tables to be ingested and COPY_ON_WRITE storage type. The command line options for `HoodieMultiTableDeltaStreamer` are pretty much similar to `HoodieDeltaStreamer` with the only exception that you are required to provide table wise configs in separate files in a dedicated config folder. The [...]
 
diff --git a/website/versioned_docs/version-0.10.0/writing_data.md b/website/versioned_docs/version-0.10.0/writing_data.md
index 24cc8252653..719813360c4 100644
--- a/website/versioned_docs/version-0.10.0/writing_data.md
+++ b/website/versioned_docs/version-0.10.0/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.10.1/writing_data.md b/website/versioned_docs/version-0.10.1/writing_data.md
index 24cc8252653..719813360c4 100644
--- a/website/versioned_docs/version-0.10.1/writing_data.md
+++ b/website/versioned_docs/version-0.10.1/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.11.0/writing_data.md b/website/versioned_docs/version-0.11.0/writing_data.md
index 792db27971d..701aa0f5d75 100644
--- a/website/versioned_docs/version-0.11.0/writing_data.md
+++ b/website/versioned_docs/version-0.11.0/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.11.1/writing_data.md b/website/versioned_docs/version-0.11.1/writing_data.md
index 792db27971d..701aa0f5d75 100644
--- a/website/versioned_docs/version-0.11.1/writing_data.md
+++ b/website/versioned_docs/version-0.11.1/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.12.0/writing_data.md b/website/versioned_docs/version-0.12.0/writing_data.md
index 59ebdcff674..840fe194072 100644
--- a/website/versioned_docs/version-0.12.0/writing_data.md
+++ b/website/versioned_docs/version-0.12.0/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.12.1/writing_data.md b/website/versioned_docs/version-0.12.1/writing_data.md
index 59ebdcff674..840fe194072 100644
--- a/website/versioned_docs/version-0.12.1/writing_data.md
+++ b/website/versioned_docs/version-0.12.1/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.12.2/writing_data.md b/website/versioned_docs/version-0.12.2/writing_data.md
index 51679a6cc88..f2393924f16 100644
--- a/website/versioned_docs/version-0.12.2/writing_data.md
+++ b/website/versioned_docs/version-0.12.2/writing_data.md
@@ -361,7 +361,7 @@ mode(Append).
 save(basePath);
 ```
 
-2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+2. Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
 
 This example will remove all the records from the table that exist in the DataSet `deleteDF`:
 ```scala
@@ -369,7 +369,7 @@ This example will remove all the records from the table that exist in the DataSe
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
 ```
 
 3. Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
diff --git a/website/versioned_docs/version-0.5.3/writing_data.md b/website/versioned_docs/version-0.5.3/writing_data.md
index d0ae3be5fc2..85aebf23c9d 100644
--- a/website/versioned_docs/version-0.5.3/writing_data.md
+++ b/website/versioned_docs/version-0.5.3/writing_data.md
@@ -221,14 +221,14 @@ For more info refer to [Delete support in Hudi](https://cwiki.apache.org/conflue
  - **Soft Deletes** : With soft deletes, user wants to retain the key but just null out the values for all other fields. 
  This can be simply achieved by ensuring the appropriate fields are nullable in the table schema and simply upserting the table after setting these fields to null.
  - **Hard Deletes** : A stronger form of delete is to physically remove any trace of the record from the table. This can be achieved by issuing an upsert with a custom payload implementation
- via either DataSource or DeltaStreamer which always returns Optional.Empty as the combined value. Hudi ships with a built-in `org.apache.hudi.EmptyHoodieRecordPayload` class that does exactly this.
+ via either DataSource or DeltaStreamer which always returns Optional.Empty as the combined value. Hudi ships with a built-in `org.apache.hudi.common.model.EmptyHoodieRecordPayload` class that does exactly this.
  
 ```java
  deleteDF // dataframe containing just records to be deleted
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
  
 ```
 
diff --git a/website/versioned_docs/version-0.6.0/writing_data.md b/website/versioned_docs/version-0.6.0/writing_data.md
index 8e503738b53..0a5b5933917 100644
--- a/website/versioned_docs/version-0.6.0/writing_data.md
+++ b/website/versioned_docs/version-0.6.0/writing_data.md
@@ -356,7 +356,7 @@ For more info refer to [Delete support in Hudi](https://cwiki.apache.org/conflue
 
    1) Using DataSource, set `OPERATION_OPT_KEY` to `DELETE_OPERATION_OPT_VAL`. This will remove all the records in the DataSet being submitted.
    
-   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
    
    3) Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
     
@@ -366,7 +366,7 @@ Example using hard delete method 2, remove all the records from the table that e
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
  
 ```
 
diff --git a/website/versioned_docs/version-0.7.0/writing_data.md b/website/versioned_docs/version-0.7.0/writing_data.md
index 7395b01ea2a..fb100aeb8f5 100644
--- a/website/versioned_docs/version-0.7.0/writing_data.md
+++ b/website/versioned_docs/version-0.7.0/writing_data.md
@@ -358,7 +358,7 @@ For more info refer to [Delete support in Hudi](https://cwiki.apache.org/conflue
 
    1) Using DataSource, set `OPERATION_OPT_KEY` to `DELETE_OPERATION_OPT_VAL`. This will remove all the records in the DataSet being submitted.
    
-   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
    
    3) Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
     
@@ -368,7 +368,7 @@ Example using hard delete method 2, remove all the records from the table that e
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
  
 ```
 
diff --git a/website/versioned_docs/version-0.8.0/writing_data.md b/website/versioned_docs/version-0.8.0/writing_data.md
index 7fd79f62ed5..fd2d5094f7a 100644
--- a/website/versioned_docs/version-0.8.0/writing_data.md
+++ b/website/versioned_docs/version-0.8.0/writing_data.md
@@ -392,7 +392,7 @@ For more info refer to [Delete support in Hudi](https://cwiki.apache.org/conflue
 
    1) Using DataSource, set `OPERATION_OPT_KEY` to `DELETE_OPERATION_OPT_VAL`. This will remove all the records in the DataSet being submitted.
    
-   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+   2) Using DataSource, set `PAYLOAD_CLASS_OPT_KEY` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
    
    3) Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
     
@@ -402,7 +402,7 @@ Example using hard delete method 2, remove all the records from the table that e
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(DataSourceWriteOptions.PAYLOAD_CLASS_OPT_KEY, "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
  
 ```
 
diff --git a/website/versioned_docs/version-0.9.0/writing_data.md b/website/versioned_docs/version-0.9.0/writing_data.md
index 3c406b268d4..7671593bacf 100644
--- a/website/versioned_docs/version-0.9.0/writing_data.md
+++ b/website/versioned_docs/version-0.9.0/writing_data.md
@@ -391,7 +391,7 @@ For more info refer to [Delete support in Hudi](https://cwiki.apache.org/conflue
 
    1) Using DataSource, set `OPERATION.key()` to `DELETE_OPERATION_OPT_VAL`. This will remove all the records in the DataSet being submitted.
    
-   2) Using DataSource, set `HoodieWriteConfig.WRITE_PAYLOAD_CLASS_NAME.key()` to `"org.apache.hudi.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
+   2) Using DataSource, set `HoodieWriteConfig.WRITE_PAYLOAD_CLASS_NAME.key()` to `"org.apache.hudi.common.model.EmptyHoodieRecordPayload"`. This will remove all the records in the DataSet being submitted. 
    
    3) Using DataSource or DeltaStreamer, add a column named `_hoodie_is_deleted` to DataSet. The value of this column must be set to `true` for all the records to be deleted and either `false` or left null for any records which are to be upserted.
     
@@ -401,7 +401,7 @@ Example using hard delete method 2, remove all the records from the table that e
    .write().format("org.apache.hudi")
    .option(...) // Add HUDI options like record-key, partition-path and others as needed for your setup
    // specify record_key, partition_key, precombine_fieldkey & usual params
-   .option(HoodieWriteConfig.WRITE_PAYLOAD_CLASS_NAME.key(), "org.apache.hudi.EmptyHoodieRecordPayload")
+   .option(HoodieWriteConfig.WRITE_PAYLOAD_CLASS_NAME.key(), "org.apache.hudi.common.model.EmptyHoodieRecordPayload")
  
 ```