You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@doris.apache.org by mo...@apache.org on 2021/11/09 02:02:38 UTC
[incubator-doris] branch master updated: [Docs] Add hdfs outfile
example (#7052)
This is an automated email from the ASF dual-hosted git repository.
morningman pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-doris.git
The following commit(s) were added to refs/heads/master by this push:
new 5d946cc [Docs] Add hdfs outfile example (#7052)
5d946cc is described below
commit 5d946ccd5e39f801082c495884f96a93a8bdd490
Author: EmmyMiao87 <52...@qq.com>
AuthorDate: Tue Nov 9 10:02:28 2021 +0800
[Docs] Add hdfs outfile example (#7052)
---
docs/en/administrator-guide/outfile.md | 30 +++++++++++++++------
docs/zh-CN/administrator-guide/outfile.md | 44 +++++++++++++++++++++----------
2 files changed, 52 insertions(+), 22 deletions(-)
diff --git a/docs/en/administrator-guide/outfile.md b/docs/en/administrator-guide/outfile.md
index e2dd8c2..3b5dfe3 100644
--- a/docs/en/administrator-guide/outfile.md
+++ b/docs/en/administrator-guide/outfile.md
@@ -64,7 +64,7 @@ INTO OUTFILE "file_path"
Specify the relevant attributes. Currently it supports exporting through the Broker process, or through the S3, HDFS protocol.
+ Broker related attributes need to be prefixed with `broker.`. For details, please refer to [Broker Document](./broker.html).
- + HDFS protocal can directly execute HDFS protocal configuration.
+ + HDFS protocal can directly execute HDFS protocal configuration. hdfs.fs.defaultFS is used to fill in the namenode address and port. It is required.
+ S3 protocol can directly execute S3 protocol configuration.
```
@@ -141,11 +141,11 @@ Planning example for concurrent export:
1. Example 1
- Export simple query results to the file `hdfs:/path/to/result.txt`. Specify the export format as CSV. Use `my_broker` and set kerberos authentication information. Specify the column separator as `,` and the line delimiter as `\n`.
+ Export simple query results to the file `hdfs://path/to/result.txt`. Specify the export format as CSV. Use `my_broker` and set kerberos authentication information. Specify the column separator as `,` and the line delimiter as `\n`.
```
SELECT * FROM tbl
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
FORMAT AS CSV
PROPERTIES
(
@@ -165,11 +165,11 @@ Planning example for concurrent export:
2. Example 2
- Export simple query results to the file `hdfs:/path/to/result.parquet`. Specify the export format as PARQUET. Use `my_broker` and set kerberos authentication information.
+ Export simple query results to the file `hdfs://path/to/result.parquet`. Specify the export format as PARQUET. Use `my_broker` and set kerberos authentication information.
```
SELECT c1, c2, c3 FROM tbl
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
FORMAT AS PARQUET
PROPERTIES
(
@@ -185,7 +185,7 @@ Planning example for concurrent export:
3. Example 3
- Export the query result of the CTE statement to the file `hdfs:/path/to/result.txt`. The default export format is CSV. Use `my_broker` and set hdfs high availability information. Use the default column separators and line delimiter.
+ Export the query result of the CTE statement to the file `hdfs://path/to/result.txt`. The default export format is CSV. Use `my_broker` and set hdfs high availability information. Use the default column separators and line delimiter.
```
WITH
@@ -194,7 +194,7 @@ Planning example for concurrent export:
x2 AS
(SELECT k3 FROM tbl2)
SELEC k1 FROM x1 UNION SELECT k3 FROM x2
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
PROPERTIES
(
"broker.name" = "my_broker",
@@ -300,7 +300,21 @@ Planning example for concurrent export:
**But because the query statement has a top-level sorting node, even if the query is enabled for concurrently exported session variables, it cannot be exported concurrently.**
-7. Example 7
+8. Example 8
+
+ Use hdfs to export and export the simple query results to the file `hdfs://path/to/result.txt`. Specify the export format as csv.
+
+ ```
+ select * from tbl
+ into outfile "hdfs://path/to/result_"
+ format as csv
+ properties
+ (
+ "hdfs.fs.defaultfs" = "hdfs://namenode:port",
+ );
+ ```
+
+9. Example 9
Export simple query results to the file `hdfs://path/to/result.txt`. Specify the export format as CSV. Use HDFS protocal directly and set kerberos authentication information.
diff --git a/docs/zh-CN/administrator-guide/outfile.md b/docs/zh-CN/administrator-guide/outfile.md
index 762ce21..52b8345 100644
--- a/docs/zh-CN/administrator-guide/outfile.md
+++ b/docs/zh-CN/administrator-guide/outfile.md
@@ -65,7 +65,7 @@ INTO OUTFILE "file_path"
指定相关属性。目前支持通过 Broker 进程, 或通过 S3 协议进行导出。
+ Broker 相关属性需加前缀 `broker.`。具体参阅[Broker 文档](./broker.html)。
- + HDFS 相关属性需加前缀 `hdfs.`。
+ + HDFS 相关属性需加前缀 `hdfs.` 其中 hdfs.fs.defaultFS 用于填写 namenode 地址和端口。属于必填项。。
+ S3 协议则直接执行 S3 协议配置即可。
```
@@ -140,11 +140,11 @@ explain select xxx from xxx where xxx into outfile "s3://xxx" format as csv pro
1. 示例1
- 使用 broker 方式导出,将简单查询结果导出到文件 `hdfs:/path/to/result.txt`。指定导出格式为 CSV。使用 `my_broker` 并设置 kerberos 认证信息。指定列分隔符为 `,`,行分隔符为 `\n`。
+ 使用 broker 方式导出,将简单查询结果导出到文件 `hdfs://path/to/result.txt`。指定导出格式为 CSV。使用 `my_broker` 并设置 kerberos 认证信息。指定列分隔符为 `,`,行分隔符为 `\n`。
```
SELECT * FROM tbl
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
FORMAT AS CSV
PROPERTIES
(
@@ -164,11 +164,11 @@ explain select xxx from xxx where xxx into outfile "s3://xxx" format as csv pro
2. 示例2
- 将简单查询结果导出到文件 `hdfs:/path/to/result.parquet`。指定导出格式为 PARQUET。使用 `my_broker` 并设置 kerberos 认证信息。
+ 将简单查询结果导出到文件 `hdfs://path/to/result.parquet`。指定导出格式为 PARQUET。使用 `my_broker` 并设置 kerberos 认证信息。
```
SELECT c1, c2, c3 FROM tbl
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
FORMAT AS PARQUET
PROPERTIES
(
@@ -184,7 +184,7 @@ explain select xxx from xxx where xxx into outfile "s3://xxx" format as csv pro
3. 示例3
- 将 CTE 语句的查询结果导出到文件 `hdfs:/path/to/result.txt`。默认导出格式为 CSV。使用 `my_broker` 并设置 hdfs 高可用信息。使用默认的行列分隔符。
+ 将 CTE 语句的查询结果导出到文件 `hdfs://path/to/result.txt`。默认导出格式为 CSV。使用 `my_broker` 并设置 hdfs 高可用信息。使用默认的行列分隔符。
```
WITH
@@ -193,7 +193,7 @@ explain select xxx from xxx where xxx into outfile "s3://xxx" format as csv pro
x2 AS
(SELECT k3 FROM tbl2)
SELEC k1 FROM x1 UNION SELECT k3 FROM x2
- INTO OUTFILE "hdfs:/path/to/result_"
+ INTO OUTFILE "hdfs://path/to/result_"
PROPERTIES
(
"broker.name" = "my_broker",
@@ -301,20 +301,36 @@ explain select xxx from xxx where xxx into outfile "s3://xxx" format as csv pro
7. 示例7
- 使用 hdfs 方式导出,将简单查询结果导出到文件 `hdfs:/path/to/result.txt`。指定导出格式为 CSV。使用并设置 kerberos 认证信息。
+ 使用 hdfs 方式导出,将简单查询结果导出到文件 `hdfs://path/to/result.txt`。指定导出格式为 csv。
```
- SELECT * FROM tbl
- INTO OUTFILE "hdfs://path/to/result_"
- FORMAT AS CSV
- PROPERTIES
+ select * from tbl
+ into outfile "hdfs://path/to/result_"
+ format as csv
+ properties
+ (
+ "hdfs.fs.defaultfs" = "hdfs://namenode:port",
+ );
+ ```
+
+
+8. 示例8
+
+ 使用 hdfs 方式导出,将简单查询结果导出到文件 `hdfs://path/to/result.txt`。指定导出格式为 csv。使用并设置 kerberos 认证信息。
+
+ ```
+ select * from tbl
+ into outfile "hdfs://path/to/result_"
+ format as csv
+ properties
(
- "hdfs.fs.defaultFS" = "hdfs://namenode:port",
+ "hdfs.fs.defaultfs" = "hdfs://namenode:port",
"hdfs.hadoop.security.authentication" = "kerberos",
- "hdfs.kerberos_principal" = "doris@YOUR.COM",
+ "hdfs.kerberos_principal" = "doris@your.com",
"hdfs.kerberos_keytab" = "/home/doris/my.keytab"
);
```
+
## 返回结果
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@doris.apache.org
For additional commands, e-mail: commits-help@doris.apache.org