You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@carbondata.apache.org by ra...@apache.org on 2019/04/02 02:41:54 UTC

[carbondata] 34/41: [HOTFIX][DOC] Optimize quick-start-guide.md and dml-of-carbondata.md

This is an automated email from the ASF dual-hosted git repository.

ravipesala pushed a commit to branch branch-1.5
in repository https://gitbox.apache.org/repos/asf/carbondata.git

commit cde80f28618c3b187dc3c4a63dd53cb5c9ee5b0b
Author: Zhang Zhichao <44...@qq.com>
AuthorDate: Sun Mar 17 23:29:06 2019 +0800

    [HOTFIX][DOC] Optimize quick-start-guide.md and dml-of-carbondata.md
    
    add note for using Spark + Hive 1.1.X in 'quick-start-guide.md' file
    separate 'load data' and 'insert into' sql to avoid ambiguity
    remove 'LOAD DATA LOCAL' semantic, currently the 'LOCAL' is invalid.
    
    This closes #3151
---
 docs/dml-of-carbondata.md | 24 +++++++++++++++---------
 docs/quick-start-guide.md |  6 ++++--
 2 files changed, 19 insertions(+), 11 deletions(-)

diff --git a/docs/dml-of-carbondata.md b/docs/dml-of-carbondata.md
index f89c49a..6ec0520 100644
--- a/docs/dml-of-carbondata.md
+++ b/docs/dml-of-carbondata.md
@@ -35,10 +35,13 @@ CarbonData DML statements are documented here,which includes:
   This command is used to load csv files to carbondata, OPTIONS are not mandatory for data loading process. 
 
   ```
-  LOAD DATA [LOCAL] INPATH 'folder_path' 
+  LOAD DATA INPATH 'folder_path'
   INTO TABLE [db_name.]table_name 
   OPTIONS(property_name=property_value, ...)
   ```
+  **NOTE**:
+    * Use 'file://' prefix to indicate local input files path, but it just supports local mode.
+    * If run on cluster mode, please upload all input files to distributed file system, for example 'hdfs://' for hdfs.
 
   **Supported Properties:**
 
@@ -232,7 +235,7 @@ CarbonData DML statements are documented here,which includes:
    Example:
 
    ```
-   LOAD DATA local inpath '/opt/rawdata/data.csv' INTO table carbontable
+   LOAD DATA inpath '/opt/rawdata/data.csv' INTO table carbontable
    options('DELIMITER'=',', 'QUOTECHAR'='"','COMMENTCHAR'='#',
    'HEADER'='false',
    'FILEHEADER'='empno,empname,designation,doj,workgroupcategory,
@@ -350,17 +353,19 @@ CarbonData DML statements are documented here,which includes:
   This command allows you to load data using static partition.
 
   ```
-  LOAD DATA [LOCAL] INPATH 'folder_path' 
+  LOAD DATA INPATH 'folder_path'
   INTO TABLE [db_name.]table_name PARTITION (partition_spec) 
-  OPTIONS(property_name=property_value, ...)    
-  INSERT INTO INTO TABLE [db_name.]table_name PARTITION (partition_spec) <SELECT STATEMENT>
+  OPTIONS(property_name=property_value, ...)
+
+  INSERT INTO TABLE [db_name.]table_name PARTITION (partition_spec) <SELECT STATEMENT>
   ```
 
   Example:
   ```
-  LOAD DATA LOCAL INPATH '${env:HOME}/staticinput.csv'
+  LOAD DATA INPATH '${env:HOME}/staticinput.csv'
   INTO TABLE locationTable
-  PARTITION (country = 'US', state = 'CA')  
+  PARTITION (country = 'US', state = 'CA')
+
   INSERT INTO TABLE locationTable
   PARTITION (country = 'US', state = 'AL')
   SELECT <columns list excluding partition columns> FROM another_user
@@ -372,8 +377,9 @@ CarbonData DML statements are documented here,which includes:
 
   Example:
   ```
-  LOAD DATA LOCAL INPATH '${env:HOME}/staticinput.csv'
-  INTO TABLE locationTable          
+  LOAD DATA INPATH '${env:HOME}/staticinput.csv'
+  INTO TABLE locationTable
+
   INSERT INTO TABLE locationTable
   SELECT <columns list excluding partition columns> FROM another_user
   ```
diff --git a/docs/quick-start-guide.md b/docs/quick-start-guide.md
index 244a9ae..316fa26 100644
--- a/docs/quick-start-guide.md
+++ b/docs/quick-start-guide.md
@@ -241,7 +241,9 @@ mv carbondata.tar.gz carbonlib/
 --executor-cores 2
 ```
 
-**NOTE**: Make sure you have permissions for CarbonData JARs and files through which driver and executor will start.
+**NOTE**:
+ - Make sure you have permissions for CarbonData JARs and files through which driver and executor will start.
+ - If use Spark + Hive 1.1.X, it needs to add carbondata assembly jar and carbondata-hive jar into parameter 'spark.sql.hive.metastore.jars' in spark-default.conf file.
 
 
 
@@ -485,4 +487,4 @@ select * from carbon_table;
 
 **Note :** Create Tables and data loads should be done before executing queries as we can not create carbon table from this interface.
 
-```
\ No newline at end of file
+```