You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@linkis.apache.org by ca...@apache.org on 2023/02/02 03:32:37 UTC

[linkis-website] branch dev updated: update ip format (#658)

This is an automated email from the ASF dual-hosted git repository.

casion pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/linkis-website.git


The following commit(s) were added to refs/heads/dev by this push:
     new 1cd22dc090 update ip format (#658)
1cd22dc090 is described below

commit 1cd22dc0900a4a594edb3d014dc7e4ee485c7eef
Author: binbincheng <10...@users.noreply.github.com>
AuthorDate: Thu Feb 2 11:32:32 2023 +0800

    update ip format (#658)
    
    * update ip format
    
    * update ip format
---
 docs/deployment/integrated/sso-with-redis.md                      | 2 +-
 docs/development/debug.md                                         | 8 ++++----
 docs/engine-usage/openlookeng.md                                  | 2 +-
 docs/user-guide/datasource-client.md                              | 4 ++--
 .../current/deployment/integrated/sso-with-redis.md               | 2 +-
 .../docusaurus-plugin-content-docs/current/development/debug.md   | 8 ++++----
 .../current/user-guide/datasource-client.md                       | 4 ++--
 .../version-1.1.0/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.1/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.2/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.3/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.2.0/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.3.0/user-guide/datasource-client.md                 | 4 ++--
 .../version-1.3.0/user-guide/sso-with-redis.md                    | 2 +-
 .../version-1.1.0/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.1/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.2/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.1.3/user-guide/linkis-datasource-client.md          | 4 ++--
 .../version-1.2.0/user-guide/linkis-datasource-client.md          | 4 ++--
 versioned_docs/version-1.3.0/user-guide/datasource-client.md      | 4 ++--
 versioned_docs/version-1.3.0/user-guide/sso-with-redis.md         | 2 +-
 21 files changed, 41 insertions(+), 41 deletions(-)

diff --git a/docs/deployment/integrated/sso-with-redis.md b/docs/deployment/integrated/sso-with-redis.md
index 8bf6eaa9af..cca146c9bb 100644
--- a/docs/deployment/integrated/sso-with-redis.md
+++ b/docs/deployment/integrated/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
 
 # sentinel mode
 linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
 linkis.session.redis.password=test123
 
 ```
diff --git a/docs/development/debug.md b/docs/development/debug.md
index e7aa205ec6..39d25bd10e 100644
--- a/docs/development/debug.md
+++ b/docs/development/debug.md
@@ -440,24 +440,24 @@ Results of the:
     "message": "OK",
     "data": {
         "taskID": 1,
-        "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+        "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
     }
 }
 ````
 
 Finally, check the running status of the task and get the running result set:
 
-GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress
+GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress
 
 ````json
 {
-    "method": "/api/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress",
+    "method": "/api/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress",
     "status": 0,
     "message": "OK",
     "data": {
         "progress": 1,
         "progressInfo": [],
-        "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+        "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
     }
 }
 ````
diff --git a/docs/engine-usage/openlookeng.md b/docs/engine-usage/openlookeng.md
index 6e1bd442e0..fd54e634d2 100644
--- a/docs/engine-usage/openlookeng.md
+++ b/docs/engine-usage/openlookeng.md
@@ -20,7 +20,7 @@ mv hetu-cli-1.5.0-executable.jar hetu-cli
 chmod +x hetu-cli
 
 # link service
-./hetu-cli --server 172.22.32.6:9090 --catalog tpcds --schema default
+./hetu-cli --server 127.0.0.1:9090 --catalog tpcds --schema default
 
 # Execute query statement
 lk:default> select d_date_sk, d_date_id, d_date, d_month_seq from tpcds.sf1.date_dim order by d_date limit 5;
diff --git a/docs/user-guide/datasource-client.md b/docs/user-guide/datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/docs/user-guide/datasource-client.md
+++ b/docs/user-guide/datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
index f0d728565e..aaf9548eb2 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/deployment/integrated/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
 
 # 哨兵模式
 linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
 linkis.session.redis.password=test123
 
 ```
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
index 4760ad4305..42b7d1f14a 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/development/debug.md
@@ -444,24 +444,24 @@ body参数:
     "message": "OK",
     "data": {
         "taskID": 1,
-        "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+        "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
     }
 }
 ```
 
 最后检查任务运行状态和获取运行结果集:
 
-GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress
+GET http://127.0.0.1:9001/api/rest_j/v1/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress
 
 ```json
 {
-    "method": "/api/entrance/exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0/progress",
+    "method": "/api/entrance/exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0/progress",
     "status": 0,
     "message": "OK",
     "data": {
         "progress": 1,
         "progressInfo": [],
-        "execID": "exec_id018017linkis-cg-entrance192.168.3.13:9104IDE_leojie_shell_0"
+        "execID": "exec_id018017linkis-cg-entrance127.0.0.1:9104IDE_leojie_shell_0"
     }
 }
 ```
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
index 1d81d95b9f..b0bcd75e95 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/current/user-guide/datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.0/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.1/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.2/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.1.3/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
index 105f177d6b..b3ad7f6596 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.2.0/user-guide/linkis-datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
index 3155ca4780..f4d9edb0d9 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/datasource-client.md
@@ -117,7 +117,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -127,7 +127,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
index f076ecd677..655f6841c5 100644
--- a/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
+++ b/i18n/zh-CN/docusaurus-plugin-content-docs/version-1.3.0/user-guide/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
 
 # 哨兵模式
 linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
 linkis.session.redis.password=test123
 
 ```
diff --git a/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.0/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.1/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.2/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.1.3/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md b/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
index 94a6576d69..c90dc16d17 100644
--- a/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
+++ b/versioned_docs/version-1.2.0/user-guide/linkis-datasource-client.md
@@ -119,7 +119,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -129,7 +129,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.3.0/user-guide/datasource-client.md b/versioned_docs/version-1.3.0/user-guide/datasource-client.md
index ae6673598e..eac03e612d 100644
--- a/versioned_docs/version-1.3.0/user-guide/datasource-client.md
+++ b/versioned_docs/version-1.3.0/user-guide/datasource-client.md
@@ -118,7 +118,7 @@ def testCreateDataSourceForEs(client:LinkisDataSourceRemoteClient): Unit ={
 def testUpdateDataSourceParameterForKafka(client:LinkisDataSourceRemoteClient): Unit ={
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
-    connParams.put("brokers","172.24.2.232:9092")
+    connParams.put("brokers","127.0.0.1:9092")
     params.put("connectParams",connParams)
     params.put("comment","kafka data source")
   client.updateDataSourceParameter(UpdateDataSourceParameterAction.builder().setUser("hadoop").setDataSourceId("7").addRequestPayloads(params).build())
@@ -128,7 +128,7 @@ def testUpdateDataSourceParameterForEs(client:LinkisDataSourceRemoteClient): Uni
     val params = new util.HashMap[String,Any]()
     val connParams = new util.HashMap[String,Any]()
     val elasticUrls = new util.ArrayList[String]()
-    elasticUrls.add("http://172.24.2.231:9200")
+    elasticUrls.add("http://127.0.0.1:9200")
     connParams.put("elasticUrls",elasticUrls)
     params.put("connectParams",connParams)
     params.put("comment","es data source")
diff --git a/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md b/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
index 63035883c4..5b525be72f 100644
--- a/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
+++ b/versioned_docs/version-1.3.0/user-guide/sso-with-redis.md
@@ -36,7 +36,7 @@ linkis.session.redis.password=test123
 
 # Sentinel mode
 linkis.session.redis.sentinel.master=sentinel-master-name
-linkis.session.redis.sentinel.nodes=192.168.1.1:6381,192.168.2.1:6381,192.168.3.1:6381
+linkis.session.redis.sentinel.nodes=127.0.1.1:6381,127.0.2.1:6381,127.0.3.1:6381
 linkis.session.redis.password=test123
 
 ````


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@linkis.apache.org
For additional commands, e-mail: commits-help@linkis.apache.org