You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@linkis.apache.org by ca...@apache.org on 2023/06/21 04:43:29 UTC
[linkis-website] branch dev updated: Add datasource-generate-sql.md (#708)
This is an automated email from the ASF dual-hosted git repository.
casion pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/linkis-website.git
The following commit(s) were added to refs/heads/dev by this push:
new c9b0942fb4 Add datasource-generate-sql.md (#708)
c9b0942fb4 is described below
commit c9b0942fb473433dc29a8e213f4ec4fa927e4267
Author: ChengJie1053 <18...@163.com>
AuthorDate: Wed Jun 21 12:43:24 2023 +0800
Add datasource-generate-sql.md (#708)
* Add datasource-generate-sql.md
* Modified datasource-generate-sql.md
* Add datasource-generate-sql.md English version
---
docs/feature/datasource-generate-sql.md | 141 +++++++++++++++++++++
.../current/feature/datasource-generate-sql.md | 141 +++++++++++++++++++++
2 files changed, 282 insertions(+)
diff --git a/docs/feature/datasource-generate-sql.md b/docs/feature/datasource-generate-sql.md
new file mode 100644
index 0000000000..32178e36b5
--- /dev/null
+++ b/docs/feature/datasource-generate-sql.md
@@ -0,0 +1,141 @@
+---
+title: Generate SQL from the data source
+sidebar_position: 0.2
+---
+
+## 1. Background
+SparkSQL and JdbcSQL are generated based on data source information, including DDL, DML, and DQL
+
+## 2. Instructions for use
+### Generate SparkSQL
+Parameter Description:
+
+| parameter name | description | default value |
+|------------------------------|-------|-----|
+| `dataSourceName` | Data source name | - |
+| `system` | System name | - |
+| `database` | Database name | - |
+| `table` | Table name | - |
+
+Submit the task through RestFul, the request example is as follows.
+```json
+GET /api/rest_j/v1/metadataQuery/getSparkSql?dataSourceName=mysql&system=system&database=test&table=test
+```
+
+The following is an example of the response.
+```json
+{
+ "method": null,
+ "status": 0,
+ "message": "OK",
+ "data": {
+ "sparkSql": {
+ "ddl": "CREATE TEMPORARY TABLE test USING org.apache.spark.sql.jdbc OPTIONS ( url 'jdbc:mysql://localhost:3306/test', dbtable 'test', user 'root', password 'password')",
+ "dml": "INSERT INTO test SELECT * FROM ${resultTable}",
+ "dql": "SELECT id,name FROM test"
+ }
+ }
+}
+```
+Currently, jdbc, kafka, elasticsearch, and mongo data sources are supported. You can register spark table based on SparkSQLDdl for query
+
+### Generate JdbcSQL
+Parameter Description:
+
+| parameter name | description | default value |
+|------------------------------|-------|-----|
+| `dataSourceName` | Data source name | - |
+| `system` | System name | - |
+| `database` | Database name | - |
+| `table` | Table name | - |
+
+Submit the task through RestFul, the request example is as follows.
+```json
+GET /api/rest_j/v1/metadataQuery/getJdbcSql?dataSourceName=mysql&system=system&database=test&table=test
+```
+
+The following is an example of the response.
+```json
+{
+ "method": null,
+ "status": 0,
+ "message": "OK",
+ "data": {
+ "jdbcSql": {
+ "ddl": "CREATE TABLE `test` (\n\t `id` varchar(64) CHARACTER SET utf8mb4 COLLATE utf8mb4_0900_ai_ci NOT NULL COMMENT '列名是id',\n\t `name` varchar(255) CHARACTER SET utf8mb4 COLLATE utf8mb4_0900_ai_ci NOT NULL COMMENT '列名是name',\n\t PRIMARY KEY (`id`)\n\t) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_0900_ai_ci",
+ "dml": "INSERT INTO test SELECT * FROM ${resultTable}",
+ "dql": "SELECT id,name FROM test"
+ }
+ }
+}
+```
+Currently, jdbc data sources are supported, such as mysql, oracle, and postgres. JdbcSQLDdl can be used for front-end display
+
+## 3. Precautions
+1. You need to register the data source first
+
+## 4. Implementation principle
+### Generate SparkSQL implementation principles
+Define DDL_SQL_TEMPLATE to retrieve data source information for replacement
+```java
+ public static final String JDBC_DDL_SQL_TEMPLATE =
+ "CREATE TEMPORARY TABLE %s "
+ + "USING org.apache.spark.sql.jdbc "
+ + "OPTIONS ("
+ + " url '%s',"
+ + " dbtable '%s',"
+ + " user '%s',"
+ + " password '%s'"
+ + ")";
+```
+
+### Generate JdbcSQL implementation principles
+Concatenate DDL based on the table schema information
+```java
+ public String generateJdbcDdlSql(String database, String table) {
+ StringBuilder ddl = new StringBuilder();
+ ddl.append("CREATE TABLE ").append(String.format("%s.%s", database, table)).append(" (");
+
+ try {
+ List<MetaColumnInfo> columns = getColumns(database, table);
+ if (CollectionUtils.isNotEmpty(columns)) {
+ for (MetaColumnInfo column : columns) {
+ ddl.append("\n\t").append(column.getName()).append(" ").append(column.getType());
+ if (column.getLength() > 0) {
+ ddl.append("(").append(column.getLength()).append(")");
+ }
+ if (!column.isNullable()) {
+ ddl.append(" NOT NULL");
+ }
+ ddl.append(",");
+ }
+ String primaryKeys =
+ columns.stream()
+ .filter(MetaColumnInfo::isPrimaryKey)
+ .map(MetaColumnInfo::getName)
+ .collect(Collectors.joining(", "));
+ if (StringUtils.isNotBlank(primaryKeys)) {
+ ddl.append(String.format("\n\tPRIMARY KEY (%s),", primaryKeys));
+ }
+ ddl.deleteCharAt(ddl.length() - 1);
+ }
+ } catch (Exception e) {
+ LOG.warn("Fail to get Sql columns(获取字段列表失败)");
+ }
+
+ ddl.append("\n)");
+
+ return ddl.toString();
+ }
+```
+Some data sources support fetching DDL directly
+
+mysql
+```sql
+SHOW CREATE TABLE 'table'
+```
+
+oracle
+```sql
+SELECT DBMS_METADATA.GET_DDL('TABLE', 'table', 'database') AS DDL FROM DUAL
+```
\ No newline at end of file
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/feature/datasource-generate-sql.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/feature/datasource-generate-sql.md
new file mode 100644
index 0000000000..31ba0a2c37
--- /dev/null
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/feature/datasource-generate-sql.md
@@ -0,0 +1,141 @@
+---
+title: 根据数据源生成SQL
+sidebar_position: 0.2
+---
+
+## 1. 背景
+根据数据源信息生成SparkSQL和JdbcSQL,包含DDL、DML、DQL
+
+## 2. 使用说明
+### 生成SparkSQL
+参数说明:
+
+| 参数名 | 说明 | 默认值 |
+|------------------------------|-------|-----|
+| `dataSourceName` | 数据源名称 | - |
+| `system` | 系统名称 | - |
+| `database` | 数据库名称 | - |
+| `table` | 表名称 | - |
+
+通过 RestFul 的方式提交任务,请求示例如下。
+```json
+GET /api/rest_j/v1/metadataQuery/getSparkSql?dataSourceName=mysql&system=system&database=test&table=test
+```
+
+响应示例如下。
+```json
+{
+ "method": null,
+ "status": 0,
+ "message": "OK",
+ "data": {
+ "sparkSql": {
+ "ddl": "CREATE TEMPORARY TABLE test USING org.apache.spark.sql.jdbc OPTIONS ( url 'jdbc:mysql://localhost:3306/test', dbtable 'test', user 'root', password 'password')",
+ "dml": "INSERT INTO test SELECT * FROM ${resultTable}",
+ "dql": "SELECT id,name FROM test"
+ }
+ }
+}
+```
+目前支持jdbc、kafka、elasticsearch、mongo数据源,可以根据SparkSQLDdl注册spark table进行查询
+
+### 生成JdbcSQL
+参数说明:
+
+| 参数名 | 说明 | 默认值 |
+|------------------------------|-------|-----|
+| `dataSourceName` | 数据源名称 | - |
+| `system` | 系统名称 | - |
+| `database` | 数据库名称 | - |
+| `table` | 表名称 | - |
+
+通过 RestFul 的方式提交任务,请求示例如下。
+```json
+GET /api/rest_j/v1/metadataQuery/getJdbcSql?dataSourceName=mysql&system=system&database=test&table=test
+```
+
+响应示例如下。
+```json
+{
+ "method": null,
+ "status": 0,
+ "message": "OK",
+ "data": {
+ "jdbcSql": {
+ "ddl": "CREATE TABLE `test` (\n\t `id` varchar(64) CHARACTER SET utf8mb4 COLLATE utf8mb4_0900_ai_ci NOT NULL COMMENT '列名是id',\n\t `name` varchar(255) CHARACTER SET utf8mb4 COLLATE utf8mb4_0900_ai_ci NOT NULL COMMENT '列名是name',\n\t PRIMARY KEY (`id`)\n\t) ENGINE=InnoDB DEFAULT CHARSET=utf8mb4 COLLATE=utf8mb4_0900_ai_ci",
+ "dml": "INSERT INTO test SELECT * FROM ${resultTable}",
+ "dql": "SELECT id,name FROM test"
+ }
+ }
+}
+```
+目前支持jdbc数据源,如:mysql、oracle、postgres等,JdbcSQLDdl可以用于前端展示
+
+## 3. 注意事项
+1. 需要先注册数据源
+
+## 4. 实现原理
+### 生成SparkSQL实现原理
+定义DDL_SQL_TEMPLATE,获取数据源信息进行替换
+```java
+ public static final String JDBC_DDL_SQL_TEMPLATE =
+ "CREATE TEMPORARY TABLE %s "
+ + "USING org.apache.spark.sql.jdbc "
+ + "OPTIONS ("
+ + " url '%s',"
+ + " dbtable '%s',"
+ + " user '%s',"
+ + " password '%s'"
+ + ")";
+```
+
+### 生成JdbcSQL实现原理
+根据表schema信息拼接DDL
+```java
+ public String generateJdbcDdlSql(String database, String table) {
+ StringBuilder ddl = new StringBuilder();
+ ddl.append("CREATE TABLE ").append(String.format("%s.%s", database, table)).append(" (");
+
+ try {
+ List<MetaColumnInfo> columns = getColumns(database, table);
+ if (CollectionUtils.isNotEmpty(columns)) {
+ for (MetaColumnInfo column : columns) {
+ ddl.append("\n\t").append(column.getName()).append(" ").append(column.getType());
+ if (column.getLength() > 0) {
+ ddl.append("(").append(column.getLength()).append(")");
+ }
+ if (!column.isNullable()) {
+ ddl.append(" NOT NULL");
+ }
+ ddl.append(",");
+ }
+ String primaryKeys =
+ columns.stream()
+ .filter(MetaColumnInfo::isPrimaryKey)
+ .map(MetaColumnInfo::getName)
+ .collect(Collectors.joining(", "));
+ if (StringUtils.isNotBlank(primaryKeys)) {
+ ddl.append(String.format("\n\tPRIMARY KEY (%s),", primaryKeys));
+ }
+ ddl.deleteCharAt(ddl.length() - 1);
+ }
+ } catch (Exception e) {
+ LOG.warn("Fail to get Sql columns(获取字段列表失败)");
+ }
+
+ ddl.append("\n)");
+
+ return ddl.toString();
+ }
+```
+部分数据源支持直接获取DDL
+
+mysql
+```sql
+SHOW CREATE TABLE 'table'
+```
+
+oracle
+```sql
+SELECT DBMS_METADATA.GET_DDL('TABLE', 'table', 'database') AS DDL FROM DUAL
+```
\ No newline at end of file
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@linkis.apache.org
For additional commands, e-mail: commits-help@linkis.apache.org