You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by vi...@apache.org on 2019/05/31 05:46:45 UTC

[incubator-hudi] branch asf-site updated: Update hoodie-utilities to hoodie-utilities-bundle

This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new d1d74fe  Update hoodie-utilities to hoodie-utilities-bundle
d1d74fe is described below

commit d1d74fe54a20459505d739b3da40838a3112af93
Author: Vinoth Chandar <vi...@uber.com>
AuthorDate: Thu May 30 20:15:22 2019 -0700

    Update hoodie-utilities to hoodie-utilities-bundle
---
 docs/contributing.md | 2 +-
 docs/writing_data.md | 6 +++---
 2 files changed, 4 insertions(+), 4 deletions(-)

diff --git a/docs/contributing.md b/docs/contributing.md
index cf449f3..71673c8 100644
--- a/docs/contributing.md
+++ b/docs/contributing.md
@@ -93,7 +93,7 @@ Discussion about contributing code to Hudi happens on the [dev@ mailing list](co
   * `hoodie-integ-test` : Longer running integration test processes
   * `hoodie-spark` : Spark datasource for writing and reading Hudi datasets. Streaming sink.
   * `hoodie-utilities` : Houses tools like DeltaStreamer, SnapshotCopier
-  * `packaging` : Poms for building out bundles for easier drop in to Spark, Hive, Presto
+  * `packaging` : Poms for building out bundles for easier drop in to Spark, Hive, Presto, Utilities
   * `style`  : Code formatting, checkstyle files
 
 
diff --git a/docs/writing_data.md b/docs/writing_data.md
index 54a3801..c2d1df8 100644
--- a/docs/writing_data.md
+++ b/docs/writing_data.md
@@ -12,7 +12,7 @@ speeding up large Spark jobs via upserts using the [Hudi datasource](#datasource
 
 ## DeltaStreamer
 
-The `HoodieDeltaStreamer` utility (part of hoodie-utilities) provides the way to ingest from different sources such as DFS or Kafka, with the following capabilities.
+The `HoodieDeltaStreamer` utility (part of hoodie-utilities-bundle) provides the way to ingest from different sources such as DFS or Kafka, with the following capabilities.
 
  - Exactly once ingestion of new events from Kafka, [incremental imports](https://sqoop.apache.org/docs/1.4.2/SqoopUserGuide.html#_incremental_imports) from Sqoop or output of `HiveIncrementalPuller` or files under a DFS folder
  - Support json, avro or a custom record types for the incoming data
@@ -23,7 +23,7 @@ The `HoodieDeltaStreamer` utility (part of hoodie-utilities) provides the way to
 Command line options describe capabilities in more detail
 
 ```
-[hoodie]$ spark-submit --class com.uber.hoodie.utilities.deltastreamer.HoodieDeltaStreamer `ls hoodie-utilities/target/hoodie-utilities-*-SNAPSHOT.jar` --help
+[hoodie]$ spark-submit --class com.uber.hoodie.utilities.deltastreamer.HoodieDeltaStreamer `ls packaging/hoodie-utilities-bundle/target/hoodie-utilities-bundle-*.jar` --help
 Usage: <main class> [options]
   Options:
     --commit-on-errors
@@ -112,7 +112,7 @@ For e.g: once you have Confluent Kafka, Schema registry up & running, produce so
 and then ingest it as follows.
 
 ```
-[hoodie]$ spark-submit --class com.uber.hoodie.utilities.deltastreamer.HoodieDeltaStreamer `ls hoodie-utilities/target/hoodie-utilities-*-SNAPSHOT.jar` \
+[hoodie]$ spark-submit --class com.uber.hoodie.utilities.deltastreamer.HoodieDeltaStreamer `ls packaging/hoodie-utilities-bundle/target/hoodie-utilities-bundle-*.jar` \
   --props file://${PWD}/hoodie-utilities/src/test/resources/delta-streamer-config/kafka-source.properties \
   --schemaprovider-class com.uber.hoodie.utilities.schema.SchemaRegistryProvider \
   --source-class com.uber.hoodie.utilities.sources.AvroKafkaSource \