You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@carbondata.apache.org by ra...@apache.org on 2019/04/02 02:41:54 UTC
[carbondata] 34/41: [HOTFIX][DOC] Optimize quick-start-guide.md and
dml-of-carbondata.md
This is an automated email from the ASF dual-hosted git repository.
ravipesala pushed a commit to branch branch-1.5
in repository https://gitbox.apache.org/repos/asf/carbondata.git
commit cde80f28618c3b187dc3c4a63dd53cb5c9ee5b0b
Author: Zhang Zhichao <44...@qq.com>
AuthorDate: Sun Mar 17 23:29:06 2019 +0800
[HOTFIX][DOC] Optimize quick-start-guide.md and dml-of-carbondata.md
add note for using Spark + Hive 1.1.X in 'quick-start-guide.md' file
separate 'load data' and 'insert into' sql to avoid ambiguity
remove 'LOAD DATA LOCAL' semantic, currently the 'LOCAL' is invalid.
This closes #3151
---
docs/dml-of-carbondata.md | 24 +++++++++++++++---------
docs/quick-start-guide.md | 6 ++++--
2 files changed, 19 insertions(+), 11 deletions(-)
diff --git a/docs/dml-of-carbondata.md b/docs/dml-of-carbondata.md
index f89c49a..6ec0520 100644
--- a/docs/dml-of-carbondata.md
+++ b/docs/dml-of-carbondata.md
@@ -35,10 +35,13 @@ CarbonData DML statements are documented here,which includes:
This command is used to load csv files to carbondata, OPTIONS are not mandatory for data loading process.
```
- LOAD DATA [LOCAL] INPATH 'folder_path'
+ LOAD DATA INPATH 'folder_path'
INTO TABLE [db_name.]table_name
OPTIONS(property_name=property_value, ...)
```
+ **NOTE**:
+ * Use 'file://' prefix to indicate local input files path, but it just supports local mode.
+ * If run on cluster mode, please upload all input files to distributed file system, for example 'hdfs://' for hdfs.
**Supported Properties:**
@@ -232,7 +235,7 @@ CarbonData DML statements are documented here,which includes:
Example:
```
- LOAD DATA local inpath '/opt/rawdata/data.csv' INTO table carbontable
+ LOAD DATA inpath '/opt/rawdata/data.csv' INTO table carbontable
options('DELIMITER'=',', 'QUOTECHAR'='"','COMMENTCHAR'='#',
'HEADER'='false',
'FILEHEADER'='empno,empname,designation,doj,workgroupcategory,
@@ -350,17 +353,19 @@ CarbonData DML statements are documented here,which includes:
This command allows you to load data using static partition.
```
- LOAD DATA [LOCAL] INPATH 'folder_path'
+ LOAD DATA INPATH 'folder_path'
INTO TABLE [db_name.]table_name PARTITION (partition_spec)
- OPTIONS(property_name=property_value, ...)
- INSERT INTO INTO TABLE [db_name.]table_name PARTITION (partition_spec) <SELECT STATEMENT>
+ OPTIONS(property_name=property_value, ...)
+
+ INSERT INTO TABLE [db_name.]table_name PARTITION (partition_spec) <SELECT STATEMENT>
```
Example:
```
- LOAD DATA LOCAL INPATH '${env:HOME}/staticinput.csv'
+ LOAD DATA INPATH '${env:HOME}/staticinput.csv'
INTO TABLE locationTable
- PARTITION (country = 'US', state = 'CA')
+ PARTITION (country = 'US', state = 'CA')
+
INSERT INTO TABLE locationTable
PARTITION (country = 'US', state = 'AL')
SELECT <columns list excluding partition columns> FROM another_user
@@ -372,8 +377,9 @@ CarbonData DML statements are documented here,which includes:
Example:
```
- LOAD DATA LOCAL INPATH '${env:HOME}/staticinput.csv'
- INTO TABLE locationTable
+ LOAD DATA INPATH '${env:HOME}/staticinput.csv'
+ INTO TABLE locationTable
+
INSERT INTO TABLE locationTable
SELECT <columns list excluding partition columns> FROM another_user
```
diff --git a/docs/quick-start-guide.md b/docs/quick-start-guide.md
index 244a9ae..316fa26 100644
--- a/docs/quick-start-guide.md
+++ b/docs/quick-start-guide.md
@@ -241,7 +241,9 @@ mv carbondata.tar.gz carbonlib/
--executor-cores 2
```
-**NOTE**: Make sure you have permissions for CarbonData JARs and files through which driver and executor will start.
+**NOTE**:
+ - Make sure you have permissions for CarbonData JARs and files through which driver and executor will start.
+ - If use Spark + Hive 1.1.X, it needs to add carbondata assembly jar and carbondata-hive jar into parameter 'spark.sql.hive.metastore.jars' in spark-default.conf file.
@@ -485,4 +487,4 @@ select * from carbon_table;
**Note :** Create Tables and data loads should be done before executing queries as we can not create carbon table from this interface.
-```
\ No newline at end of file
+```