You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@linkis.apache.org by ca...@apache.org on 2023/02/02 03:32:37 UTC
[linkis-website] branch dev updated: update ip format (#658)
This is an automated email from the ASF dual-hosted git repository.
casion pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/linkis-website.git
The following commit(s) were added to refs/heads/dev by this push:
new 1cd22dc090 update ip format (#658)
1cd22dc090 is described below
commit 1cd22dc0900a4a594edb3d014dc7e4ee485c7eef
Author: binbincheng <10...@users.noreply.github.com>
AuthorDate: Thu Feb 2 11:32:32 2023 +0800
update ip format (#658)
* update ip format
* update ip format
---
docs/deployment/integrated/sso-with-redis.md | 2 +-
docs/development/debug.md | 8 ++++----
docs/engine-usage/openlookeng.md | 2 +-
docs/user-guide/datasource-client.md | 4 ++--
.../current/deployment/integrated/sso-with-redis.md | 2 +-
.../docusaurus-plugin-content-docs/current/development/debug.md | 8 ++++----
.../current/user-guide/datasource-client.md | 4 ++--
.../version-1.1.0/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.1/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.2/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.3/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.2.0/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.3.0/user-guide/datasource-client.md | 4 ++--
.../version-1.3.0/user-guide/sso-with-redis.md | 2 +-
.../version-1.1.0/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.1/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.2/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.1.3/user-guide/linkis-datasource-client.md | 4 ++--
.../version-1.2.0/user-guide/linkis-datasource-client.md | 4 ++--
versioned_docs/version-1.3.0/user-guide/datasource-client.md | 4 ++--
versioned_docs/version-1.3.0/user-guide/sso-with-redis.md | 2 +-
21 files changed, 41 insertions(+), 41 deletions(-)
diff --git a/docs/deployment/integrated/sso-with-redis.md b/docs/deployment/integrated/sso-with-redis.md
index 8bf6eaa9af..cca146c9bb 100644
--- a/docs/deployment/integrated/sso-with-redis.md
+++ b/docs/deployment/integrated/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
# sentinel mode
linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
linkis.session.redis.password=test123
```
diff --git a/docs/development/debug.md b/docs/development/debug.md
index e7aa205ec6..39d25bd10e 100644
--- a/docs/development/debug.md
+++ b/docs/development/debug.md
@@ -440,24 +440,24 @@ Results of the:
"message": "OK",
"data": {
"taskID": 1,
- "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+ "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
}
}
````
Finally, check the running status of the task and get the running result set:
-GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress
+GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress
````json
{
- "method": "/api/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress",
+ "method": "/api/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress",
"status": 0,
"message": "OK",
"data": {
"progress": 1,
"progressInfo": [],
- "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+ "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
}
}
````
diff --git a/docs/engine-usage/openlookeng.md b/docs/engine-usage/openlookeng.md
index 6e1bd442e0..fd54e634d2 100644
--- a/docs/engine-usage/openlookeng.md
+++ b/docs/engine-usage/openlookeng.md
@@ -20,7 +20,7 @@ mv hetu-cli-1.5.0-executable.jar hetu-cli
chmod +x hetu-cli
# link service
-./hetu-cli --server 172.22.32.6:9090 --catalog tpcds --schema default
+./hetu-cli --server 127.0.0.1:9090 --catalog tpcds --schema default
# Execute query statement
lk:default> select d_date_sk, d_date_id, d_date, d_month_seq from tpcds.sf1.date_dim order by d_date limit 5;
diff --git a/docs/user-guide/datasource-client.md b/docs/user-guide/datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/docs/user-guide/datasource-client.md
+++ b/docs/user-guide/datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
index f0d728565e..aaf9548eb2 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
# 哨兵模式
linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
linkis.session.redis.password=test123
```
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
index 4760ad4305..42b7d1f14a 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
@@ -444,24 +444,24 @@ body参数:
"message": "OK",
"data": {
"taskID": 1,
- "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+ "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
}
}
```
最后检查任务运行状态和获取运行结果集:
-GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress
+GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress
```json
{
- "method": "/api/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress",
+ "method": "/api/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress",
"status": 0,
"message": "OK",
"data": {
"progress": 1,
"progressInfo": [],
- "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+ "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
}
}
```
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
index 1d81d95b9f..b0bcd75e95 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
index 3155ca4780..f4d9edb0d9 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
index f076ecd677..655f6841c5 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
# 哨兵模式
linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
linkis.session.redis.password=test123
```
diff --git a/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.3.0/user-guide/datasource-client.md b/versioned_docs/version-1.3.0/user-guide/datasource-client.md
index ae6673598e..eac03e612d 100644
--- a/versioned_docs/version-1.3.0/user-guide/datasource-client.md
+++ b/versioned_docs/version-1.3.0/user-guide/datasource-client.md
@@ -118,7 +118,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
- connParams.put("brokers","172.24.2.232:9092")
+ connParams.put("brokers","127.0.0.1:9092")
params.put("connectParams",connParams)
params.put("comment","kafka data source")
client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -128,7 +128,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
val params = new util.HashMap[String,Any]()
val connParams = new util.HashMap[String,Any]()
val elasticUrls = new util.ArrayList[String]()
- elasticUrls.add("http://172.24.2.231:9200")
+ elasticUrls.add("http://127.0.0.1:9200")
connParams.put("elasticUrls",elasticUrls)
params.put("connectParams",connParams)
params.put("comment","es data source")
diff --git a/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md b/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
index 63035883c4..5b525be72f 100644
--- a/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
+++ b/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
# Sentinel mode
linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
linkis.session.redis.password=test123
````
---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@linkis.apache.org
For additional commands, e-mail: commits-help@linkis.apache.org