You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hudi.apache.org by vi...@apache.org on 2019/09/18 19:50:06 UTC

[incubator-hudi] branch asf-site updated: [HUDI-262] Update website to reflect name change in InputFormat class name

This is an automated email from the ASF dual-hosted git repository.

vinoth pushed a commit to branch asf-site
in repository https://gitbox.apache.org/repos/asf/incubator-hudi.git


The following commit(s) were added to refs/heads/asf-site by this push:
     new edc13c1  [HUDI-262] Update website to reflect name change in InputFormat class name
edc13c1 is described below

commit edc13c125f2bebb147fd21eb192b8a50ad568ec8
Author: Bhavani Sudha Saktheeswaran <bh...@uber.com>
AuthorDate: Wed Sep 18 12:32:40 2019 -0700

    [HUDI-262] Update website to reflect name change in InputFormat class name
---
 docs/README.md           | 2 +-
 docs/querying_data.cn.md | 6 +++---
 docs/querying_data.md    | 6 +++---
 3 files changed, 7 insertions(+), 7 deletions(-)

diff --git a/docs/README.md b/docs/README.md
index 4307a6a..74c78e1 100644
--- a/docs/README.md
+++ b/docs/README.md
@@ -5,7 +5,7 @@ This folder contains resources that build the [Apache Hudi website](https://hudi
 
 ### Building docs
 
-The site is based on a [Jekyll](https://jekyllrb.com/) theme hosted [here](idratherbewriting.com/documentation-theme-jekyll/) with detailed instructions.
+The site is based on a [Jekyll](https://jekyllrb.com/) theme hosted [here](https://idratherbewriting.com/documentation-theme-jekyll/) with detailed instructions.
 
 #### Docker
 
diff --git a/docs/querying_data.cn.md b/docs/querying_data.cn.md
index 3a6fd0f..1653b08 100644
--- a/docs/querying_data.cn.md
+++ b/docs/querying_data.cn.md
@@ -14,8 +14,8 @@ bundle has been provided, the dataset can be queried by popular query engines li
 Specifically, there are two Hive tables named off [table name](configurations.html#TABLE_NAME_OPT_KEY) passed during write. 
 For e.g, if `table name = hudi_tbl`, then we get  
 
- - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieInputFormat`, exposing purely columnar data.
- - `hudi_tbl_rt` realizes the real time view of the dataset  backed by `HoodieRealtimeInputFormat`, exposing merged view of base and log data.
+ - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieParquetInputFormat`, exposing purely columnar data.
+ - `hudi_tbl_rt` realizes the real time view of the dataset  backed by `HoodieParquetRealtimeInputFormat`, exposing merged view of base and log data.
 
 As discussed in the concepts section, the one key primitive needed for [incrementally processing](https://www.oreilly.com/ideas/ubers-case-for-incremental-processing-on-hadoop),
 is `incremental pulls` (to obtain a change stream/log from a dataset). Hudi datasets can be pulled incrementally, which means you can get ALL and ONLY the updated & new rows 
@@ -33,7 +33,7 @@ classes with its dependencies are available for query planning & execution.
 
 ### Read Optimized table {#hive-ro-view}
 In addition to setup above, for beeline cli access, the `hive.input.format` variable needs to be set to the  fully qualified path name of the 
-inputformat `org.apache.hudi.hadoop.HoodieInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set 
+inputformat `org.apache.hudi.hadoop.HoodieParquetInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set 
 to `org.apache.hadoop.hive.ql.io.HiveInputFormat`
 
 ### Real time table {#hive-rt-view}
diff --git a/docs/querying_data.md b/docs/querying_data.md
index 3a6fd0f..1653b08 100644
--- a/docs/querying_data.md
+++ b/docs/querying_data.md
@@ -14,8 +14,8 @@ bundle has been provided, the dataset can be queried by popular query engines li
 Specifically, there are two Hive tables named off [table name](configurations.html#TABLE_NAME_OPT_KEY) passed during write. 
 For e.g, if `table name = hudi_tbl`, then we get  
 
- - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieInputFormat`, exposing purely columnar data.
- - `hudi_tbl_rt` realizes the real time view of the dataset  backed by `HoodieRealtimeInputFormat`, exposing merged view of base and log data.
+ - `hudi_tbl` realizes the read optimized view of the dataset backed by `HoodieParquetInputFormat`, exposing purely columnar data.
+ - `hudi_tbl_rt` realizes the real time view of the dataset  backed by `HoodieParquetRealtimeInputFormat`, exposing merged view of base and log data.
 
 As discussed in the concepts section, the one key primitive needed for [incrementally processing](https://www.oreilly.com/ideas/ubers-case-for-incremental-processing-on-hadoop),
 is `incremental pulls` (to obtain a change stream/log from a dataset). Hudi datasets can be pulled incrementally, which means you can get ALL and ONLY the updated & new rows 
@@ -33,7 +33,7 @@ classes with its dependencies are available for query planning & execution.
 
 ### Read Optimized table {#hive-ro-view}
 In addition to setup above, for beeline cli access, the `hive.input.format` variable needs to be set to the  fully qualified path name of the 
-inputformat `org.apache.hudi.hadoop.HoodieInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set 
+inputformat `org.apache.hudi.hadoop.HoodieParquetInputFormat`. For Tez, additionally the `hive.tez.input.format` needs to be set 
 to `org.apache.hadoop.hive.ql.io.HiveInputFormat`
 
 ### Real time table {#hive-rt-view}