You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by yi...@apache.org on 2022/07/07 06:06:46 UTC

[hudi] branch asf-site updated: [DOCS] Minor edits in the doc when mentioning HoodieDeltaStreamer (#6057)

This is an automated email from the ASF dual-hosted git repository.

yihua pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new edb1df7171 [DOCS] Minor edits in the doc when mentioning HoodieDeltaStreamer (#6057)
edb1df7171 is described below

commit edb1df7171a46e4b0c6bf5cf75811a3ae3e482eb
Author: Brandon <br...@users.noreply.github.com>
AuthorDate: Thu Jul 7 14:06:36 2022 +0800

    [DOCS] Minor edits in the doc when mentioning HoodieDeltaStreamer (#6057)
    
    Using inline code style when mentioning HoodieDeltaStreamer to make the doc consistent
---
 website/docs/hoodie_deltastreamer.md | 8 ++++----
 1 file changed, 4 insertions(+), 4 deletions(-)

diff --git a/website/docs/hoodie_deltastreamer.md b/website/docs/hoodie_deltastreamer.md
index 938127da31..607b6a989b 100644
--- a/website/docs/hoodie_deltastreamer.md
+++ b/website/docs/hoodie_deltastreamer.md
@@ -224,7 +224,7 @@ A deltastreamer job can then be triggered as follows:
 Read more in depth about concurrency control in the [concurrency control concepts](/docs/concurrency_control) section
 
 ## Checkpointing
-HoodieDeltaStreamer uses checkpoints to keep track of what data has been read already so it can resume without needing to reprocess all data.
+`HoodieDeltaStreamer` uses checkpoints to keep track of what data has been read already so it can resume without needing to reprocess all data.
 When using a Kafka source, the checkpoint is the [Kafka Offset](https://cwiki.apache.org/confluence/display/KAFKA/Offset+Management) 
 When using a DFS source, the checkpoint is the 'last modified' timestamp of the latest file read.
 Checkpoints are saved in the .hoodie commit file as `deltastreamer.checkpoint.key`.
@@ -316,7 +316,7 @@ For DFS sources the following behaviors are expected:
 - `HoodieDeltaStreamer` reads the files under the source base path (`hoodie.deltastreamer.source.dfs.root`) directly, and it won't use the partition paths under this base path as fields of the dataset. Detailed examples can be found [here](https://github.com/apache/hudi/issues/5485).
 
 ### Kafka
-Hudi can read directly from Kafka clusters. See more details on HoodieDeltaStreamer to learn how to setup streaming 
+Hudi can read directly from Kafka clusters. See more details on `HoodieDeltaStreamer` to learn how to setup streaming 
 ingestion with exactly once semantics, checkpointing, and plugin transformations. The following formats are supported 
 when reading data from Kafka:
 
@@ -335,7 +335,7 @@ to trigger/processing of new or changed data as soon as it is available on S3.
 3. Find the queue URL and Region to set these configurations:
    1. hoodie.deltastreamer.s3.source.queue.url=https://sqs.us-west-2.amazonaws.com/queue/url
    2. hoodie.deltastreamer.s3.source.queue.region=us-west-2 
-4. start the S3EventsSource and S3EventsHoodieIncrSource using the HoodieDeltaStreamer utility as shown in sample commands below:
+4. start the S3EventsSource and S3EventsHoodieIncrSource using the `HoodieDeltaStreamer` utility as shown in sample commands below:
 
 Insert code sample from this blog: https://hudi.apache.org/blog/2021/08/23/s3-events-source/#configuration-and-setup
 
@@ -524,7 +524,7 @@ There are 3 use cases for incremental query:
 | `read.end-commit` | `false` | the latest commit | -- |
 
 ## Kafka Connect Sink
-If you want to perform streaming ingestion into Hudi format similar to HoodieDeltaStreamer, but you don't want to depend on Spark,
+If you want to perform streaming ingestion into Hudi format similar to `HoodieDeltaStreamer`, but you don't want to depend on Spark,
 try out the new experimental release of Hudi Kafka Connect Sink. Read the [ReadMe](https://github.com/apache/hudi/tree/master/hudi-kafka-connect) 
 for full documentation.