You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@hugegraph.apache.org by ji...@apache.org on 2023/01/01 08:16:47 UTC

[incubator-hugegraph-doc] branch master updated: enhance validate doc (#171)

This is an automated email from the ASF dual-hosted git repository.

jin pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-hugegraph-doc.git


The following commit(s) were added to refs/heads/master by this push:
     new 89a0a1a6 enhance validate doc (#171)
89a0a1a6 is described below

commit 89a0a1a6a5341c157b8179e1888b9254317d03d3
Author: 青年 <10...@qq.com>
AuthorDate: Sun Jan 1 16:16:43 2023 +0800

    enhance validate doc (#171)
    
    * enhance validate doc
    
    * format a lot of tables
    
    * use java 11 as the 1st priority choose
    
    Co-authored-by: imbajin <ji...@apache.org>
---
 .asf.yaml                                          |   2 +-
 .github/workflows/hugo.yml                         |   4 +-
 content/cn/about/_index.html                       |   1 +
 .../changelog/hugegraph-0.12.0-release-notes.md    |   2 +-
 .../changelog/hugegraph-1.0.0-release-notes.md     |   2 +-
 content/cn/docs/config/config-computer.md          | 310 ++++++++++-----------
 content/cn/docs/config/config-option.md            |  20 +-
 .../contribution-guidelines/validate-release.md    |   7 +-
 content/cn/docs/download/download.md               |  12 +-
 content/cn/docs/guides/architectural.md            |   8 +-
 content/cn/docs/guides/faq.md                      |   2 +-
 content/cn/docs/language/hugegraph-example.md      |   4 +-
 content/cn/docs/language/hugegraph-gremlin.md      |  36 +--
 .../api-preformance/hugegraph-api-0.2.md           |  52 ++--
 .../api-preformance/hugegraph-api-0.4.4.md         |  28 +-
 .../hugegraph-api-0.5.6-Cassandra.md               |  30 +-
 .../api-preformance/hugegraph-api-0.5.6-RocksDB.md |  18 +-
 content/cn/docs/quickstart/hugegraph-client.md     |   4 +-
 content/cn/docs/quickstart/hugegraph-computer.md   |   2 +-
 content/cn/docs/quickstart/hugegraph-hubble.md     |  40 +--
 content/cn/docs/quickstart/hugegraph-loader.md     |  52 ++--
 content/cn/docs/quickstart/hugegraph-server.md     |   4 +-
 content/en/about/_index.html                       |   1 +
 .../changelog/hugegraph-0.11.2-release-notes.md    |   2 +-
 .../changelog/hugegraph-0.12.0-release-notes.md    |   4 +-
 .../changelog/hugegraph-1.0.0-release-notes.md     |   2 +-
 content/en/docs/clients/hugegraph-client.md        |   2 +-
 content/en/docs/config/config-computer.md          | 310 ++++++++++-----------
 content/en/docs/config/config-option.md            |  20 +-
 .../contribution-guidelines/validate-release.md    |  11 +-
 content/en/docs/download/download.md               |  18 +-
 content/en/docs/guides/faq.md                      |   2 +-
 content/en/docs/language/hugegraph-example.md      |   4 +-
 content/en/docs/language/hugegraph-gremlin.md      |  36 +--
 content/en/docs/quickstart/hugegraph-client.md     |   4 +-
 content/en/docs/quickstart/hugegraph-computer.md   |   2 +-
 content/en/docs/quickstart/hugegraph-loader.md     |  44 +--
 content/en/docs/quickstart/hugegraph-server.md     |   4 +-
 38 files changed, 557 insertions(+), 549 deletions(-)

diff --git a/.asf.yaml b/.asf.yaml
index 8fd7b996..58dfd5c4 100644
--- a/.asf.yaml
+++ b/.asf.yaml
@@ -42,7 +42,7 @@ github:
   protected_branches:
     master:
       required_status_checks:
-        # strict means "Require branches to be up to date before merging".
+        # strict means "Require branches to be up-to-date before merging".
         #strict: true
         contexts:
           - deploy
diff --git a/.github/workflows/hugo.yml b/.github/workflows/hugo.yml
index 3c323fcf..e2039793 100644
--- a/.github/workflows/hugo.yml
+++ b/.github/workflows/hugo.yml
@@ -38,7 +38,7 @@ jobs:
           fetch-depth: 0    # Fetch all history for .GitInfo and .Lastmod
 
       - name: Setup Node
-        uses: actions/setup-node@v2.4.0
+        uses: actions/setup-node@v3
         with:
           node-version: "16"
 
@@ -48,7 +48,7 @@ jobs:
           hugo-version: '0.102.3'
           extended: true
 
-      - uses: actions/cache@v2
+      - uses: actions/cache@v3
         with:
           path: /tmp/hugo_cache
           key: ${{ runner.os }}-hugomod-${{ hashFiles('**/go.sum') }}
diff --git a/content/cn/about/_index.html b/content/cn/about/_index.html
index 642ba464..d6b46c6c 100644
--- a/content/cn/about/_index.html
+++ b/content/cn/about/_index.html
@@ -1,6 +1,7 @@
 ---
 title: About HugeGraph
 linkTitle: About
+draft: true
 menu:
   main:
     weight: 10
diff --git a/content/cn/docs/changelog/hugegraph-0.12.0-release-notes.md b/content/cn/docs/changelog/hugegraph-0.12.0-release-notes.md
index 19a672cb..f88ee3d7 100644
--- a/content/cn/docs/changelog/hugegraph-0.12.0-release-notes.md
+++ b/content/cn/docs/changelog/hugegraph-0.12.0-release-notes.md
@@ -1,7 +1,7 @@
 ---
 title: "HugeGraph 0.12 Release Notes"
 linkTitle: "Release-0.12.0"
-weight: 1
+weight: 2
 ---
 
 ### API & Client
diff --git a/content/cn/docs/changelog/hugegraph-1.0.0-release-notes.md b/content/cn/docs/changelog/hugegraph-1.0.0-release-notes.md
index 2e1d24e4..bb28c096 100644
--- a/content/cn/docs/changelog/hugegraph-1.0.0-release-notes.md
+++ b/content/cn/docs/changelog/hugegraph-1.0.0-release-notes.md
@@ -1,7 +1,7 @@
 ---
 title: "HugeGraph 1.0.0 Release Notes"
 linkTitle: "Release-1.0.0"
-weight: 0
+weight: 1
 ---
 
 > TODO: update the summary soon, please check the detail in each repository first, thanks
diff --git a/content/cn/docs/config/config-computer.md b/content/cn/docs/config/config-computer.md
index 07779bfd..4ba628d5 100644
--- a/content/cn/docs/config/config-computer.md
+++ b/content/cn/docs/config/config-computer.md
@@ -6,172 +6,172 @@ weight: 5
 
 ### Computer Config Options
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|algorithm.message_class|org.apache.hugegraph.computer.core.config.Null|The class of message passed when compute vertex.|
-|algorithm.params_class|org.apache.hugegraph.computer.core.config.Null|The class used to transfer algorithms' parameters before algorithm been run.|
-|algorithm.result_class|org.apache.hugegraph.computer.core.config.Null|The class of vertex's value, the instance is used to store computation result for the vertex.|
-|allocator.max_vertices_per_thread|10000|Maximum number of vertices per thread processed in each memory allocator|
-|bsp.etcd_endpoints|http://localhost:2379|The end points to access etcd.|
-|bsp.log_interval|30000|The log interval(in ms) to print the log while waiting bsp event.|
-|bsp.max_super_step|10|The max super step of the algorithm.|
-|bsp.register_timeout|300000|The max timeout to wait for master and works to register.|
-|bsp.wait_master_timeout|86400000|The max timeout(in ms) to wait for master bsp event.|
-|bsp.wait_workers_timeout|86400000|The max timeout to wait for workers bsp event.|
-|hgkv.max_data_block_size|65536|The max byte size of hgkv-file data block.|
-|hgkv.max_file_size|2147483648|The max number of bytes in each hgkv-file.|
-|hgkv.max_merge_files|10|The max number of files to merge at one time.|
-|hgkv.temp_file_dir|/tmp/hgkv|This folder is used to store temporary files, temporary files will be generated during the file merging process.|
-|hugegraph.name|hugegraph|The graph name to load data and write results back.|
-|hugegraph.url|http://127.0.0.1:8080|The hugegraph url to load data and write results back.|
-|input.edge_direction|OUT|The data of the edge in which direction is loaded, when the value is BOTH, the edges in both OUT and IN direction will be loaded.|
-|input.edge_freq|MULTIPLE|The frequency of edges can exist between a pair of vertices, allowed values: [SINGLE, SINGLE_PER_LABEL, MULTIPLE]. SINGLE means that only one edge can exist between a pair of vertices, use sourceId + targetId to identify it; SINGLE_PER_LABEL means that each edge label can exist one edge between a pair of vertices, use sourceId + edgelabel + targetId to identify it; MULTIPLE means that many edge can exist between a pair of vertices, use sourceId + edgelabel + sor [...]
-|input.filter_class|org.apache.hugegraph.computer.core.input.filter.DefaultInputFilter|The class to create input-filter object, input-filter is used to Filter vertex edges according to user needs.|
-|input.loader_schema_path||The schema path of loader input, only takes effect when the input.source_type=loader is enabled|
-|input.loader_struct_path||The struct path of loader input, only takes effect when the input.source_type=loader is enabled|
-|input.max_edges_in_one_vertex|200|The maximum number of adjacent edges allowed to be attached to a vertex, the adjacent edges will be stored and transferred together as a batch unit.|
-|input.source_type|hugegraph-server|The source type to load input data, allowed values: ['hugegraph-server', 'hugegraph-loader'], the 'hugegraph-loader' means use hugegraph-loader load data from HDFS or file, if use 'hugegraph-loader' load data then please config 'input.loader_struct_path' and 'input.loader_schema_path'.|
-|input.split_fetch_timeout|300|The timeout in seconds to fetch input splits|
-|input.split_max_splits|10000000|The maximum number of input splits|
-|input.split_page_size|500|The page size for streamed load input split data|
-|input.split_size|1048576|The input split size in bytes|
-|job.id|local_0001|The job id on Yarn cluster or K8s cluster.|
-|job.partitions_count|1|The partitions count for computing one graph algorithm job.|
-|job.partitions_thread_nums|4|The number of threads for partition parallel compute.|
-|job.workers_count|1|The workers count for computing one graph algorithm job.|
-|master.computation_class|org.apache.hugegraph.computer.core.master.DefaultMasterComputation|Master-computation is computation that can determine whether to continue next superstep. It runs at the end of each superstep on master.|
-|output.batch_size|500|The batch size of output|
-|output.batch_threads|1|The threads number used to batch output|
-|output.hdfs_core_site_path||The hdfs core site path.|
-|output.hdfs_delimiter|,|The delimiter of hdfs output.|
-|output.hdfs_kerberos_enable|false|Is Kerberos authentication enabled for Hdfs.|
-|output.hdfs_kerberos_keytab||The Hdfs's key tab file for kerberos authentication.|
-|output.hdfs_kerberos_principal||The Hdfs's principal for kerberos authentication.|
-|output.hdfs_krb5_conf|/etc/krb5.conf|Kerberos configuration file.|
-|output.hdfs_merge_partitions|true|Whether merge output files of multiple partitions.|
-|output.hdfs_path_prefix|/hugegraph-computer/results|The directory of hdfs output result.|
-|output.hdfs_replication|3|The replication number of hdfs.|
-|output.hdfs_site_path||The hdfs site path.|
-|output.hdfs_url|hdfs://127.0.0.1:9000|The hdfs url of output.|
-|output.hdfs_user|hadoop|The hdfs user of output.|
-|output.output_class|org.apache.hugegraph.computer.core.output.LogOutput|The class to output the computation result of each vertex. Be called after iteration computation.|
-|output.result_name|value|The value is assigned dynamically by #name() of instance created by WORKER_COMPUTATION_CLASS.|
-|output.result_write_type|OLAP_COMMON|The result write-type to output to hugegraph, allowed values are: [OLAP_COMMON, OLAP_SECONDARY, OLAP_RANGE].|
-|output.retry_interval|10|The retry interval when output failed|
-|output.retry_times|3|The retry times when output failed|
-|output.single_threads|1|The threads number used to single output|
-|output.thread_pool_shutdown_timeout|60|The timeout seconds of output threads pool shutdown|
-|output.with_adjacent_edges|false|Output the adjacent edges of the vertex or not|
-|output.with_edge_properties|false|Output the properties of the edge or not|
-|output.with_vertex_properties|false|Output the properties of the vertex or not|
-|sort.thread_nums|4|The number of threads performing internal sorting.|
-|transport.client_connect_timeout|3000|The timeout(in ms) of client connect to server.|
-|transport.client_threads|4|The number of transport threads for client.|
-|transport.close_timeout|10000|The timeout(in ms) of close server or close client.|
-|transport.finish_session_timeout|0|The timeout(in ms) to finish session, 0 means using (transport.sync_request_timeout * transport.max_pending_requests).|
-|transport.heartbeat_interval|20000|The minimum interval(in ms) between heartbeats on client side.|
-|transport.io_mode|AUTO|The network IO Mode, either 'NIO', 'EPOLL', 'AUTO', the 'AUTO' means selecting the property mode automatically.|
-|transport.max_pending_requests|8|The max number of client unreceived ack, it will trigger the sending unavailable if the number of unreceived ack >= max_pending_requests.|
-|transport.max_syn_backlog|511|The capacity of SYN queue on server side, 0 means using system default value.|
-|transport.max_timeout_heartbeat_count|120|The maximum times of timeout heartbeat on client side, if the number of timeouts waiting for heartbeat response continuously > max_heartbeat_timeouts the channel will be closed from client side.|
-|transport.min_ack_interval|200|The minimum interval(in ms) of server reply ack.|
-|transport.min_pending_requests|6|The minimum number of client unreceived ack, it will trigger the sending available if the number of unreceived ack < min_pending_requests.|
-|transport.network_retries|3|The number of retry attempts for network communication,if network unstable.|
-|transport.provider_class|org.apache.hugegraph.computer.core.network.netty.NettyTransportProvider|The transport provider, currently only supports Netty.|
-|transport.receive_buffer_size|0|The size of socket receive-buffer in bytes, 0 means using system default value.|
-|transport.recv_file_mode|true|Whether enable receive buffer-file mode, it will receive buffer write file from socket by zero-copy if enable.|
-|transport.send_buffer_size|0|The size of socket send-buffer in bytes, 0 means using system default value.|
-|transport.server_host|127.0.0.1|The server hostname or ip to listen on to transfer data.|
-|transport.server_idle_timeout|360000|The max timeout(in ms) of server idle.|
-|transport.server_port|0|The server port to listen on to transfer data. The system will assign a random port if it's set to 0.|
-|transport.server_threads|4|The number of transport threads for server.|
-|transport.sync_request_timeout|10000|The timeout(in ms) to wait response after sending sync-request.|
-|transport.tcp_keep_alive|true|Whether enable TCP keep-alive.|
-|transport.transport_epoll_lt|false|Whether enable EPOLL level-trigger.|
-|transport.write_buffer_high_mark|67108864|The high water mark for write buffer in bytes, it will trigger the sending unavailable if the number of queued bytes > write_buffer_high_mark.|
-|transport.write_buffer_low_mark|33554432|The low water mark for write buffer in bytes, it will trigger the sending available if the number of queued bytes < write_buffer_low_mark.org.apache.hugegraph.config.OptionChecker$$Lambda$97/0x00000008001c8440@776a6d9b|
-|transport.write_socket_timeout|3000|The timeout(in ms) to write data to socket buffer.|
-|valuefile.max_segment_size|1073741824|The max number of bytes in each segment of value-file.|
-|worker.combiner_class|org.apache.hugegraph.computer.core.config.Null|Combiner can combine messages into one value for a vertex, for example page-rank algorithm can combine messages of a vertex to a sum value.|
-|worker.computation_class|org.apache.hugegraph.computer.core.config.Null|The class to create worker-computation object, worker-computation is used to compute each vertex in each superstep.|
-|worker.data_dirs|[jobs]|The directories separated by ',' that received vertices and messages can persist into.|
-|worker.edge_properties_combiner_class|org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner|The combiner can combine several properties of the same edge into one properties at inputstep.|
-|worker.partitioner|org.apache.hugegraph.computer.core.graph.partition.HashPartitioner|The partitioner that decides which partition a vertex should be in, and which worker a partition should be in.|
-|worker.received_buffers_bytes_limit|104857600|The limit bytes of buffers of received data, the total size of all buffers can't excess this limit. If received buffers reach this limit, they will be merged into a file.|
-|worker.vertex_properties_combiner_class|org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner|The combiner can combine several properties of the same vertex into one properties at inputstep.|
-|worker.wait_finish_messages_timeout|86400000|The max timeout(in ms) message-handler wait for finish-message of all workers.|
-|worker.wait_sort_timeout|600000|The max timeout(in ms) message-handler wait for sort-thread to sort one batch of buffers.|
-|worker.write_buffer_capacity|52428800|The initial size of write buffer that used to store vertex or message.|
-|worker.write_buffer_threshold|52428800|The threshold of write buffer, exceeding it will trigger sorting, the write buffer is used to store vertex or message.|
+| config option                           | default value                                                                                   | description                                                                                                                                                                                                                                                                                                                                                    [...]
+|-----------------------------------------|-------------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- [...]
+| algorithm.message_class                 | org.apache.hugegraph.computer.core.config.Null                                                  | The class of message passed when compute vertex.                                                                                                                                                                                                                                                                                                               [...]
+| algorithm.params_class                  | org.apache.hugegraph.computer.core.config.Null                                                  | The class used to transfer algorithms' parameters before algorithm been run.                                                                                                                                                                                                                                                                                   [...]
+| algorithm.result_class                  | org.apache.hugegraph.computer.core.config.Null                                                  | The class of vertex's value, the instance is used to store computation result for the vertex.                                                                                                                                                                                                                                                                  [...]
+| allocator.max_vertices_per_thread       | 10000                                                                                           | Maximum number of vertices per thread processed in each memory allocator                                                                                                                                                                                                                                                                                       [...]
+| bsp.etcd_endpoints                      | http://localhost:2379                                                                           | The end points to access etcd.                                                                                                                                                                                                                                                                                                                                 [...]
+| bsp.log_interval                        | 30000                                                                                           | The log interval(in ms) to print the log while waiting bsp event.                                                                                                                                                                                                                                                                                              [...]
+| bsp.max_super_step                      | 10                                                                                              | The max super step of the algorithm.                                                                                                                                                                                                                                                                                                                           [...]
+| bsp.register_timeout                    | 300000                                                                                          | The max timeout to wait for master and works to register.                                                                                                                                                                                                                                                                                                      [...]
+| bsp.wait_master_timeout                 | 86400000                                                                                        | The max timeout(in ms) to wait for master bsp event.                                                                                                                                                                                                                                                                                                           [...]
+| bsp.wait_workers_timeout                | 86400000                                                                                        | The max timeout to wait for workers bsp event.                                                                                                                                                                                                                                                                                                                 [...]
+| hgkv.max_data_block_size                | 65536                                                                                           | The max byte size of hgkv-file data block.                                                                                                                                                                                                                                                                                                                     [...]
+| hgkv.max_file_size                      | 2147483648                                                                                      | The max number of bytes in each hgkv-file.                                                                                                                                                                                                                                                                                                                     [...]
+| hgkv.max_merge_files                    | 10                                                                                              | The max number of files to merge at one time.                                                                                                                                                                                                                                                                                                                  [...]
+| hgkv.temp_file_dir                      | /tmp/hgkv                                                                                       | This folder is used to store temporary files, temporary files will be generated during the file merging process.                                                                                                                                                                                                                                               [...]
+| hugegraph.name                          | hugegraph                                                                                       | The graph name to load data and write results back.                                                                                                                                                                                                                                                                                                            [...]
+| hugegraph.url                           | http://127.0.0.1:8080                                                                           | The hugegraph url to load data and write results back.                                                                                                                                                                                                                                                                                                         [...]
+| input.edge_direction                    | OUT                                                                                             | The data of the edge in which direction is loaded, when the value is BOTH, the edges in both OUT and IN direction will be loaded.                                                                                                                                                                                                                              [...]
+| input.edge_freq                         | MULTIPLE                                                                                        | The frequency of edges can exist between a pair of vertices, allowed values: [SINGLE, SINGLE_PER_LABEL, MULTIPLE]. SINGLE means that only one edge can exist between a pair of vertices, use sourceId + targetId to identify it; SINGLE_PER_LABEL means that each edge label can exist one edge between a pair of vertices, use sourceId + edgelabel + targetI [...]
+| input.filter_class                      | org.apache.hugegraph.computer.core.input.filter.DefaultInputFilter                              | The class to create input-filter object, input-filter is used to Filter vertex edges according to user needs.                                                                                                                                                                                                                                                  [...]
+| input.loader_schema_path                || The schema path of loader input, only takes effect when the input.source_type=loader is enabled |
+| input.loader_struct_path                || The struct path of loader input, only takes effect when the input.source_type=loader is enabled |
+| input.max_edges_in_one_vertex           | 200                                                                                             | The maximum number of adjacent edges allowed to be attached to a vertex, the adjacent edges will be stored and transferred together as a batch unit.                                                                                                                                                                                                           [...]
+| input.source_type                       | hugegraph-server                                                                                | The source type to load input data, allowed values: ['hugegraph-server', 'hugegraph-loader'], the 'hugegraph-loader' means use hugegraph-loader load data from HDFS or file, if use 'hugegraph-loader' load data then please config 'input.loader_struct_path' and 'input.loader_schema_path'.                                                                 [...]
+| input.split_fetch_timeout               | 300                                                                                             | The timeout in seconds to fetch input splits                                                                                                                                                                                                                                                                                                                   [...]
+| input.split_max_splits                  | 10000000                                                                                        | The maximum number of input splits                                                                                                                                                                                                                                                                                                                             [...]
+| input.split_page_size                   | 500                                                                                             | The page size for streamed load input split data                                                                                                                                                                                                                                                                                                               [...]
+| input.split_size                        | 1048576                                                                                         | The input split size in bytes                                                                                                                                                                                                                                                                                                                                  [...]
+| job.id                                  | local_0001                                                                                      | The job id on Yarn cluster or K8s cluster.                                                                                                                                                                                                                                                                                                                     [...]
+| job.partitions_count                    | 1                                                                                               | The partitions count for computing one graph algorithm job.                                                                                                                                                                                                                                                                                                    [...]
+| job.partitions_thread_nums              | 4                                                                                               | The number of threads for partition parallel compute.                                                                                                                                                                                                                                                                                                          [...]
+| job.workers_count                       | 1                                                                                               | The workers count for computing one graph algorithm job.                                                                                                                                                                                                                                                                                                       [...]
+| master.computation_class                | org.apache.hugegraph.computer.core.master.DefaultMasterComputation                              | Master-computation is computation that can determine whether to continue next superstep. It runs at the end of each superstep on master.                                                                                                                                                                                                                       [...]
+| output.batch_size                       | 500                                                                                             | The batch size of output                                                                                                                                                                                                                                                                                                                                       [...]
+| output.batch_threads                    | 1                                                                                               | The threads number used to batch output                                                                                                                                                                                                                                                                                                                        [...]
+| output.hdfs_core_site_path              || The hdfs core site path.                                                                        |
+| output.hdfs_delimiter                   | ,                                                                                               | The delimiter of hdfs output.                                                                                                                                                                                                                                                                                                                                  [...]
+| output.hdfs_kerberos_enable             | false                                                                                           | Is Kerberos authentication enabled for Hdfs.                                                                                                                                                                                                                                                                                                                   [...]
+| output.hdfs_kerberos_keytab             || The Hdfs's key tab file for kerberos authentication.                                            |
+| output.hdfs_kerberos_principal          || The Hdfs's principal for kerberos authentication.                                               |
+| output.hdfs_krb5_conf                   | /etc/krb5.conf                                                                                  | Kerberos configuration file.                                                                                                                                                                                                                                                                                                                                   [...]
+| output.hdfs_merge_partitions            | true                                                                                            | Whether merge output files of multiple partitions.                                                                                                                                                                                                                                                                                                             [...]
+| output.hdfs_path_prefix                 | /hugegraph-computer/results                                                                     | The directory of hdfs output result.                                                                                                                                                                                                                                                                                                                           [...]
+| output.hdfs_replication                 | 3                                                                                               | The replication number of hdfs.                                                                                                                                                                                                                                                                                                                                [...]
+| output.hdfs_site_path                   || The hdfs site path.                                                                             |
+| output.hdfs_url                         | hdfs://127.0.0.1:9000                                                                           | The hdfs url of output.                                                                                                                                                                                                                                                                                                                                        [...]
+| output.hdfs_user                        | hadoop                                                                                          | The hdfs user of output.                                                                                                                                                                                                                                                                                                                                       [...]
+| output.output_class                     | org.apache.hugegraph.computer.core.output.LogOutput                                             | The class to output the computation result of each vertex. Be called after iteration computation.                                                                                                                                                                                                                                                              [...]
+| output.result_name                      | value                                                                                           | The value is assigned dynamically by #name() of instance created by WORKER_COMPUTATION_CLASS.                                                                                                                                                                                                                                                                  [...]
+| output.result_write_type                | OLAP_COMMON                                                                                     | The result write-type to output to hugegraph, allowed values are: [OLAP_COMMON, OLAP_SECONDARY, OLAP_RANGE].                                                                                                                                                                                                                                                   [...]
+| output.retry_interval                   | 10                                                                                              | The retry interval when output failed                                                                                                                                                                                                                                                                                                                          [...]
+| output.retry_times                      | 3                                                                                               | The retry times when output failed                                                                                                                                                                                                                                                                                                                             [...]
+| output.single_threads                   | 1                                                                                               | The threads number used to single output                                                                                                                                                                                                                                                                                                                       [...]
+| output.thread_pool_shutdown_timeout     | 60                                                                                              | The timeout seconds of output threads pool shutdown                                                                                                                                                                                                                                                                                                            [...]
+| output.with_adjacent_edges              | false                                                                                           | Output the adjacent edges of the vertex or not                                                                                                                                                                                                                                                                                                                 [...]
+| output.with_edge_properties             | false                                                                                           | Output the properties of the edge or not                                                                                                                                                                                                                                                                                                                       [...]
+| output.with_vertex_properties           | false                                                                                           | Output the properties of the vertex or not                                                                                                                                                                                                                                                                                                                     [...]
+| sort.thread_nums                        | 4                                                                                               | The number of threads performing internal sorting.                                                                                                                                                                                                                                                                                                             [...]
+| transport.client_connect_timeout        | 3000                                                                                            | The timeout(in ms) of client connect to server.                                                                                                                                                                                                                                                                                                                [...]
+| transport.client_threads                | 4                                                                                               | The number of transport threads for client.                                                                                                                                                                                                                                                                                                                    [...]
+| transport.close_timeout                 | 10000                                                                                           | The timeout(in ms) of close server or close client.                                                                                                                                                                                                                                                                                                            [...]
+| transport.finish_session_timeout        | 0                                                                                               | The timeout(in ms) to finish session, 0 means using (transport.sync_request_timeout * transport.max_pending_requests).                                                                                                                                                                                                                                         [...]
+| transport.heartbeat_interval            | 20000                                                                                           | The minimum interval(in ms) between heartbeats on client side.                                                                                                                                                                                                                                                                                                 [...]
+| transport.io_mode                       | AUTO                                                                                            | The network IO Mode, either 'NIO', 'EPOLL', 'AUTO', the 'AUTO' means selecting the property mode automatically.                                                                                                                                                                                                                                                [...]
+| transport.max_pending_requests          | 8                                                                                               | The max number of client unreceived ack, it will trigger the sending unavailable if the number of unreceived ack >= max_pending_requests.                                                                                                                                                                                                                      [...]
+| transport.max_syn_backlog               | 511                                                                                             | The capacity of SYN queue on server side, 0 means using system default value.                                                                                                                                                                                                                                                                                  [...]
+| transport.max_timeout_heartbeat_count   | 120                                                                                             | The maximum times of timeout heartbeat on client side, if the number of timeouts waiting for heartbeat response continuously > max_heartbeat_timeouts the channel will be closed from client side.                                                                                                                                                             [...]
+| transport.min_ack_interval              | 200                                                                                             | The minimum interval(in ms) of server reply ack.                                                                                                                                                                                                                                                                                                               [...]
+| transport.min_pending_requests          | 6                                                                                               | The minimum number of client unreceived ack, it will trigger the sending available if the number of unreceived ack < min_pending_requests.                                                                                                                                                                                                                     [...]
+| transport.network_retries               | 3                                                                                               | The number of retry attempts for network communication,if network unstable.                                                                                                                                                                                                                                                                                    [...]
+| transport.provider_class                | org.apache.hugegraph.computer.core.network.netty.NettyTransportProvider                         | The transport provider, currently only supports Netty.                                                                                                                                                                                                                                                                                                         [...]
+| transport.receive_buffer_size           | 0                                                                                               | The size of socket receive-buffer in bytes, 0 means using system default value.                                                                                                                                                                                                                                                                                [...]
+| transport.recv_file_mode                | true                                                                                            | Whether enable receive buffer-file mode, it will receive buffer write file from socket by zero-copy if enable.                                                                                                                                                                                                                                                 [...]
+| transport.send_buffer_size              | 0                                                                                               | The size of socket send-buffer in bytes, 0 means using system default value.                                                                                                                                                                                                                                                                                   [...]
+| transport.server_host                   | 127.0.0.1                                                                                       | The server hostname or ip to listen on to transfer data.                                                                                                                                                                                                                                                                                                       [...]
+| transport.server_idle_timeout           | 360000                                                                                          | The max timeout(in ms) of server idle.                                                                                                                                                                                                                                                                                                                         [...]
+| transport.server_port                   | 0                                                                                               | The server port to listen on to transfer data. The system will assign a random port if it's set to 0.                                                                                                                                                                                                                                                          [...]
+| transport.server_threads                | 4                                                                                               | The number of transport threads for server.                                                                                                                                                                                                                                                                                                                    [...]
+| transport.sync_request_timeout          | 10000                                                                                           | The timeout(in ms) to wait response after sending sync-request.                                                                                                                                                                                                                                                                                                [...]
+| transport.tcp_keep_alive                | true                                                                                            | Whether enable TCP keep-alive.                                                                                                                                                                                                                                                                                                                                 [...]
+| transport.transport_epoll_lt            | false                                                                                           | Whether enable EPOLL level-trigger.                                                                                                                                                                                                                                                                                                                            [...]
+| transport.write_buffer_high_mark        | 67108864                                                                                        | The high water mark for write buffer in bytes, it will trigger the sending unavailable if the number of queued bytes > write_buffer_high_mark.                                                                                                                                                                                                                 [...]
+| transport.write_buffer_low_mark         | 33554432                                                                                        | The low water mark for write buffer in bytes, it will trigger the sending available if the number of queued bytes < write_buffer_low_mark.org.apache.hugegraph.config.OptionChecker$$Lambda$97/0x00000008001c8440@776a6d9b                                                                                                                                     [...]
+| transport.write_socket_timeout          | 3000                                                                                            | The timeout(in ms) to write data to socket buffer.                                                                                                                                                                                                                                                                                                             [...]
+| valuefile.max_segment_size              | 1073741824                                                                                      | The max number of bytes in each segment of value-file.                                                                                                                                                                                                                                                                                                         [...]
+| worker.combiner_class                   | org.apache.hugegraph.computer.core.config.Null                                                  | Combiner can combine messages into one value for a vertex, for example page-rank algorithm can combine messages of a vertex to a sum value.                                                                                                                                                                                                                    [...]
+| worker.computation_class                | org.apache.hugegraph.computer.core.config.Null                                                  | The class to create worker-computation object, worker-computation is used to compute each vertex in each superstep.                                                                                                                                                                                                                                            [...]
+| worker.data_dirs                        | [jobs]                                                                                          | The directories separated by ',' that received vertices and messages can persist into.                                                                                                                                                                                                                                                                         [...]
+| worker.edge_properties_combiner_class   | org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner                         | The combiner can combine several properties of the same edge into one properties at inputstep.                                                                                                                                                                                                                                                                 [...]
+| worker.partitioner                      | org.apache.hugegraph.computer.core.graph.partition.HashPartitioner                              | The partitioner that decides which partition a vertex should be in, and which worker a partition should be in.                                                                                                                                                                                                                                                 [...]
+| worker.received_buffers_bytes_limit     | 104857600                                                                                       | The limit bytes of buffers of received data, the total size of all buffers can't excess this limit. If received buffers reach this limit, they will be merged into a file.                                                                                                                                                                                     [...]
+| worker.vertex_properties_combiner_class | org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner                         | The combiner can combine several properties of the same vertex into one properties at inputstep.                                                                                                                                                                                                                                                               [...]
+| worker.wait_finish_messages_timeout     | 86400000                                                                                        | The max timeout(in ms) message-handler wait for finish-message of all workers.                                                                                                                                                                                                                                                                                 [...]
+| worker.wait_sort_timeout                | 600000                                                                                          | The max timeout(in ms) message-handler wait for sort-thread to sort one batch of buffers.                                                                                                                                                                                                                                                                      [...]
+| worker.write_buffer_capacity            | 52428800                                                                                        | The initial size of write buffer that used to store vertex or message.                                                                                                                                                                                                                                                                                         [...]
+| worker.write_buffer_threshold           | 52428800                                                                                        | The threshold of write buffer, exceeding it will trigger sorting, the write buffer is used to store vertex or message.                                                                                                                                                                                                                                         [...]
 
 ### K8s Operator Config Options
 
-> NOTE: Option needs to be converted through environment variable settings, e.g k8s.internal_etcd_url => INTERNAL_ETCD_URL
+> NOTE: Option needs to be converted through environment variable settings, e.g. k8s.internal_etcd_url => INTERNAL_ETCD_URL
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|k8s.auto_destroy_pod|true|Whether to automatically destroy all pods when the job is completed or failed.|
-|k8s.close_reconciler_timeout|120|The max timeout(in ms) to close reconciler.|
-|k8s.internal_etcd_url|http://127.0.0.1:2379|The internal etcd url for operator system.|
-|k8s.max_reconcile_retry|3|The max retry times of reconcile.|
-|k8s.probe_backlog|50|The maximum backlog for serving health probes.|
-|k8s.probe_port|9892|The value is the port that the controller bind to for serving health probes.|
-|k8s.ready_check_internal|1000|The time interval(ms) of check ready.|
-|k8s.ready_timeout|30000|The max timeout(in ms) of check ready.|
-|k8s.reconciler_count|10|The max number of reconciler thread.|
-|k8s.resync_period|600000|The minimum frequency at which watched resources are reconciled.|
-|k8s.timezone|Asia/Shanghai|The timezone of computer job and operator.|
-|k8s.watch_namespace|hugegraph-computer-system|The value is watch custom resources in the namespace, ignore other namespaces, the '*' means is all namespaces will be watched.|
+| config option                | default value             | description                                                                                                                     |
+|------------------------------|---------------------------|---------------------------------------------------------------------------------------------------------------------------------|
+| k8s.auto_destroy_pod         | true                      | Whether to automatically destroy all pods when the job is completed or failed.                                                  |
+| k8s.close_reconciler_timeout | 120                       | The max timeout(in ms) to close reconciler.                                                                                     |
+| k8s.internal_etcd_url        | http://127.0.0.1:2379     | The internal etcd url for operator system.                                                                                      |
+| k8s.max_reconcile_retry      | 3                         | The max retry times of reconcile.                                                                                               |
+| k8s.probe_backlog            | 50                        | The maximum backlog for serving health probes.                                                                                  |
+| k8s.probe_port               | 9892                      | The value is the port that the controller bind to for serving health probes.                                                    |
+| k8s.ready_check_internal     | 1000                      | The time interval(ms) of check ready.                                                                                           |
+| k8s.ready_timeout            | 30000                     | The max timeout(in ms) of check ready.                                                                                          |
+| k8s.reconciler_count         | 10                        | The max number of reconciler thread.                                                                                            |
+| k8s.resync_period            | 600000                    | The minimum frequency at which watched resources are reconciled.                                                                |
+| k8s.timezone                 | Asia/Shanghai             | The timezone of computer job and operator.                                                                                      |
+| k8s.watch_namespace          | hugegraph-computer-system | The value is watch custom resources in the namespace, ignore other namespaces, the '*' means is all namespaces will be watched. |
 
 ### HugeGraph-Computer CRD
 
 > CRD: https://github.com/apache/hugegraph-computer/blob/master/computer-k8s-operator/manifest/hugegraph-computer-crd.v1.yaml
 
-| spec            | default value           | description                                                  | required |
-| --------------- | ----------------------- | ------------------------------------------------------------ | -------- |
-| algorithmName   |                         | The name of algorithm.                                       | true     |
-| jobId           |                         | The job id.                                                  | true     |
-| image           |                         | The image of algorithm.                                      | true     |
-| computerConf    |                         | The map of computer config options.                          | true     |
-| workerInstances |                         | The number of worker instances, it will instead the 'job.workers_count' option. | true     |
-| pullPolicy      | Always                  | The pull-policy of image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy | false    |
-| pullSecrets     |                         | The pull-secrets of Image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod | false    |
-| masterCpu       |                         | The cpu limit of master, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu) | false    |
-| workerCpu       |                         | The cpu limit of worker, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu) | false    |
+| spec            | default value           | description                                                                                                                                                                                                                                                                                      | required |
+|-----------------|-------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------|
+| algorithmName   |                         | The name of algorithm.                                                                                                                                                                                                                                                                           | true     |
+| jobId           |                         | The job id.                                                                                                                                                                                                                                                                                      | true     |
+| image           |                         | The image of algorithm.                                                                                                                                                                                                                                                                          | true     |
+| computerConf    |                         | The map of computer config options.                                                                                                                                                                                                                                                              | true     |
+| workerInstances |                         | The number of worker instances, it will instead the 'job.workers_count' option.                                                                                                                                                                                                                  | true     |
+| pullPolicy      | Always                  | The pull-policy of image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy                                                                                                                                                                       | false    |
+| pullSecrets     |                         | The pull-secrets of Image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod                                                                                                                                                   | false    |
+| masterCpu       |                         | The cpu limit of master, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu)               | false    |
+| workerCpu       |                         | The cpu limit of worker, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu)               | false    |
 | masterMemory    |                         | The memory limit of master, the unit can be one of Ei、Pi、Ti、Gi、Mi、Ki detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory) | false    |
 | workerMemory    |                         | The memory limit of worker, the unit can be one of Ei、Pi、Ti、Gi、Mi、Ki detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory) | false    |
-| log4jXml        |                         | The content of log4j.xml for computer job.                   | false    |
-| jarFile         |                         | The jar path of computer algorithm.                          | false    |
-| remoteJarUri    |                         | The remote jar uri of computer algorithm, it will overlay algorithm image. | false    |
-| jvmOptions      |                         | The java startup parameters of computer job.                 | false    |
-| envVars         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-interdependent-environment-variables/ | false    |
-| envFrom         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-environment-variable-container/ | false    |
-| masterCommand   | bin/start-computer.sh   | The run command of master, equivalent to 'Entrypoint' field of Docker. | false    |
-| masterArgs      | ["-r master", "-d k8s"] | The run args of master, equivalent to 'Cmd' field of Docker. | false    |
-| workerCommand   | bin/start-computer.sh   | The run command of worker, equivalent to 'Entrypoint' field of Docker. | false    |
-| workerArgs      | ["-r worker", "-d k8s"] | The run args of worker, equivalent to 'Cmd' field of Docker. | false    |
-| volumes         |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/ | false    |
-| volumeMounts    |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/ | false    |
-| secretPaths     |                         | The map of k8s-secret name and mount path.                   | false    |
-| configMapPaths  |                         | The map of k8s-configmap name and mount path.                | false    |
-| podTemplateSpec |                         | Please refer to: https://kubernetes.io/docs/reference/kubernetes-api/workload-resources/pod-template-v1/#PodTemplateSpec | false    |
-| securityContext |                         | Please refer to: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ | false    |
+| log4jXml        |                         | The content of log4j.xml for computer job.                                                                                                                                                                                                                                                       | false    |
+| jarFile         |                         | The jar path of computer algorithm.                                                                                                                                                                                                                                                              | false    |
+| remoteJarUri    |                         | The remote jar uri of computer algorithm, it will overlay algorithm image.                                                                                                                                                                                                                       | false    |
+| jvmOptions      |                         | The java startup parameters of computer job.                                                                                                                                                                                                                                                     | false    |
+| envVars         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-interdependent-environment-variables/                                                                                                                                                                           | false    |
+| envFrom         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-environment-variable-container/                                                                                                                                                                                 | false    |
+| masterCommand   | bin/start-computer.sh   | The run command of master, equivalent to 'Entrypoint' field of Docker.                                                                                                                                                                                                                           | false    |
+| masterArgs      | ["-r master", "-d k8s"] | The run args of master, equivalent to 'Cmd' field of Docker.                                                                                                                                                                                                                                     | false    |
+| workerCommand   | bin/start-computer.sh   | The run command of worker, equivalent to 'Entrypoint' field of Docker.                                                                                                                                                                                                                           | false    |
+| workerArgs      | ["-r worker", "-d k8s"] | The run args of worker, equivalent to 'Cmd' field of Docker.                                                                                                                                                                                                                                     | false    |
+| volumes         |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/                                                                                                                                                                                                                            | false    |
+| volumeMounts    |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/                                                                                                                                                                                                                            | false    |
+| secretPaths     |                         | The map of k8s-secret name and mount path.                                                                                                                                                                                                                                                       | false    |
+| configMapPaths  |                         | The map of k8s-configmap name and mount path.                                                                                                                                                                                                                                                    | false    |
+| podTemplateSpec |                         | Please refer to: https://kubernetes.io/docs/reference/kubernetes-api/workload-resources/pod-template-v1/#PodTemplateSpec                                                                                                                                                                         | false    |
+| securityContext |                         | Please refer to: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/                                                                                                                                                                                                      | false    |
 
 ### KubeDriver Config Options
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|k8s.build_image_bash_path||The path of command used to build image.|
-|k8s.enable_internal_algorithm|true|Whether enable internal algorithm.|
-|k8s.framework_image_url|hugegraph/hugegraph-computer:latest|The image url of computer framework.|
-|k8s.image_repository_password||The password for login image repository.|
-|k8s.image_repository_registry||The address for login image repository.|
-|k8s.image_repository_url|hugegraph/hugegraph-computer|The url of image repository.|
-|k8s.image_repository_username||The username for login image repository.|
-|k8s.internal_algorithm|[pageRank]|The name list of all internal algorithm.|
-|k8s.internal_algorithm_image_url|hugegraph/hugegraph-computer:latest|The image url of internal algorithm.|
-|k8s.jar_file_dir|/cache/jars/|The directory where the algorithm jar to upload location.|
-|k8s.kube_config|~/.kube/config|The path of k8s config file.|
-|k8s.log4j_xml_path||The log4j.xml path for computer job.|
-|k8s.namespace|hugegraph-computer-system|The namespace of hugegraph-computer system.|
-|k8s.pull_secret_names|[]|The names of pull-secret for pulling image.|
+| config option                    | default value                            | description                                               |
+|----------------------------------|------------------------------------------|-----------------------------------------------------------|
+| k8s.build_image_bash_path        || The path of command used to build image. |
+| k8s.enable_internal_algorithm    | true                                     | Whether enable internal algorithm.                        |
+| k8s.framework_image_url          | hugegraph/hugegraph-computer:latest      | The image url of computer framework.                      |
+| k8s.image_repository_password    || The password for login image repository. |
+| k8s.image_repository_registry    || The address for login image repository.  |
+| k8s.image_repository_url         | hugegraph/hugegraph-computer             | The url of image repository.                              |
+| k8s.image_repository_username    || The username for login image repository. |
+| k8s.internal_algorithm           | [pageRank]                               | The name list of all internal algorithm.                  |
+| k8s.internal_algorithm_image_url | hugegraph/hugegraph-computer:latest      | The image url of internal algorithm.                      |
+| k8s.jar_file_dir                 | /cache/jars/                             | The directory where the algorithm jar to upload location. |
+| k8s.kube_config                  | ~/.kube/config                           | The path of k8s config file.                              |
+| k8s.log4j_xml_path               || The log4j.xml path for computer job.     |
+| k8s.namespace                    | hugegraph-computer-system                | The namespace of hugegraph-computer system.               |
+| k8s.pull_secret_names            | []                                       | The names of pull-secret for pulling image.               |
diff --git a/content/cn/docs/config/config-option.md b/content/cn/docs/config/config-option.md
index dae587be..36570651 100644
--- a/content/cn/docs/config/config-option.md
+++ b/content/cn/docs/config/config-option.md
@@ -8,7 +8,7 @@ weight: 2
 
 对应配置文件`gremlin-server.yaml`
 
-| config option           | default value                                                                                                | description                                                                      |
+| config option           | default value                                                                                                | description                                                                     |
 |-------------------------|--------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------|
 | host                    | 127.0.0.1                                                                                                    | The host or ip of Gremlin Server.                                               |
 | port                    | 8182                                                                                                         | The listening port of Gremlin Server.                                           |
@@ -21,7 +21,7 @@ weight: 2
 
 对应配置文件`rest-server.properties`
 
-| config option                      | default value                                    | description                                                                                                                                                                                                    |
+| config option                      | default value                                    | description                                                                                                                                                                                                   |
 |------------------------------------|--------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 | graphs                             | [hugegraph:conf/hugegraph.properties]            | The map of graphs' name and config file.                                                                                                                                                                      |
 | server.id                          | server-1                                         | The id of rest server, used for license verification.                                                                                                                                                         |
@@ -41,7 +41,7 @@ weight: 2
 | batch.max_vertices_per_batch       | 500                                              | The maximum number of vertices submitted per batch.                                                                                                                                                           |
 | batch.max_write_ratio              | 50                                               | The maximum thread ratio for batch writing, only take effect if the batch.max_write_threads is 0.                                                                                                             |
 | batch.max_write_threads            | 0                                                | The maximum threads for batch writing, if the value is 0, the actual value will be set to batch.max_write_ratio * restserver.max_worker_threads.                                                              |
-| auth.authenticator                 |                                                  | The class path of authenticator implementation. e.g., com.baidu.hugegraph.auth.StandardAuthenticator, or com.baidu.hugegraph.auth.ConfigAuthenticator.                                                          |
+| auth.authenticator                 |                                                  | The class path of authenticator implementation. e.g., com.baidu.hugegraph.auth.StandardAuthenticator, or com.baidu.hugegraph.auth.ConfigAuthenticator.                                                        |
 | auth.admin_token                   | 162f7848-0b6d-4faf-b557-3a0797869c55             | Token for administrator operations, only for com.baidu.hugegraph.auth.ConfigAuthenticator.                                                                                                                    |
 | auth.graph_store                   | hugegraph                                        | The name of graph used to store authentication information, like users, only for com.baidu.hugegraph.auth.StandardAuthenticator.                                                                              |
 | auth.user_tokens                   | [hugegraph:9fd95c9c-711b-415b-b85f-d4df46ba5c31] | The map of user tokens with name and password, only for com.baidu.hugegraph.auth.ConfigAuthenticator.                                                                                                         |
@@ -57,7 +57,7 @@ weight: 2
 
 基本配置项及后端配置项对应配置文件:{graph-name}.properties,如`hugegraph.properties`
 
-| config option                         | default value                                | description                                                                                                                                                                                                                                                                                                                                                                                                         [...]
+| config option                         | default value                                | description                                                                                                                                                                                                                                                                                                                                                                                                         [...]
 |---------------------------------------|----------------------------------------------|-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- [...]
 | gremlin.graph	                        | com.baidu.hugegraph.HugeFactory              | Gremlin entrance to create graph.                                                                                                                                                                                                                                                                                                                                                                                   [...]
 | backend                               | rocksdb                                      | The data store type, available values are [memory, rocksdb, cassandra, scylladb, hbase, mysql].                                                                                                                                                                                                                                                                                                                     [...]
@@ -149,7 +149,7 @@ weight: 2
 
 ### Cassandra 后端配置项
 
-| config option                  | default value  | description                                                                                                                                     |
+| config option                  | default value  | description                                                                                                                                    |
 |--------------------------------|----------------|------------------------------------------------------------------------------------------------------------------------------------------------|
 | backend                        |                | Must be set to `cassandra`.                                                                                                                    |
 | serializer                     |                | Must be set to `cassandra`.                                                                                                                    |
@@ -167,7 +167,7 @@ weight: 2
 
 ### ScyllaDB 后端配置项
 
-| config option | default value | description                 |
+| config option | default value | description                |
 |---------------|---------------|----------------------------|
 | backend       |               | Must be set to `scylladb`. |
 | serializer    |               | Must be set to `scylladb`. |
@@ -176,7 +176,7 @@ weight: 2
 
 ### RocksDB 后端配置项
 
-| config option                                   | default value                                                                                                                        | description                                                                                                                                                                                                                                                                                                       [...]
+| config option                                   | default value                                                                                                                        | description                                                                                                                                                                                                                                                                                                       [...]
 |-------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [...]
 | backend                                         |                                                                                                                                      | Must be set to `rocksdb`.                                                                                                                                                                                                                                                                                         [...]
 | serializer                                      |                                                                                                                                      | Must be set to `binary`.                                                                                                                                                                                                                                                                                          [...]
@@ -230,7 +230,7 @@ weight: 2
 
 ### HBase 后端配置项
 
-| config option             | default value                  | description                                                               |
+| config option             | default value                  | description                                                              |
 |---------------------------|--------------------------------|--------------------------------------------------------------------------|
 | backend                   |                                | Must be set to `hbase`.                                                  |
 | serializer                |                                | Must be set to `hbase`.                                                  |
@@ -251,7 +251,7 @@ weight: 2
 
 ### MySQL & PostgreSQL 后端配置项
 
-| config option                    | default value               | description                                                                          |
+| config option                    | default value               | description                                                                         |
 |----------------------------------|-----------------------------|-------------------------------------------------------------------------------------|
 | backend                          |                             | Must be set to `mysql`.                                                             |
 | serializer                       |                             | Must be set to `mysql`.                                                             |
@@ -267,7 +267,7 @@ weight: 2
 
 ### PostgreSQL 后端配置项
 
-| config option | default value | description                   |
+| config option | default value | description                  |
 |---------------|---------------|------------------------------|
 | backend       |               | Must be set to `postgresql`. |
 | serializer    |               | Must be set to `postgresql`. |
diff --git a/content/cn/docs/contribution-guidelines/validate-release.md b/content/cn/docs/contribution-guidelines/validate-release.md
index 0bda3ec0..3ffd52d5 100644
--- a/content/cn/docs/contribution-guidelines/validate-release.md
+++ b/content/cn/docs/contribution-guidelines/validate-release.md
@@ -99,8 +99,11 @@ for i in *.tar.gz; do echo $i; gpg --verify $i.asc $i ; done
 8. 最后, 确保源码可以正常/正确编译 (然后看看测试和规范)
 
 ```bash
-# 同时也可以检查一下代码风格是否符合规范, 不符合的可以放下一次调整
-mvn clean package -Dmaven.test.skip=true -Dcheckstyle.skip=false
+# 请优先使用/切换到 java 11 版本进行后序的编译和运行操作
+# java --version
+
+# 尝试在 Unix 环境下编译测试是否正常
+mvn clean package -Dmaven.test.skip=true -Dcheckstyle.skip=true
 ```
 
 ##### B. 二进制包
diff --git a/content/cn/docs/download/download.md b/content/cn/docs/download/download.md
index 22811f38..395348c2 100644
--- a/content/cn/docs/download/download.md
+++ b/content/cn/docs/download/download.md
@@ -8,21 +8,21 @@ weight: 2
 
 The latest HugeGraph: **1.0.0**, released on _2022-12-31_.
 
-| components          | description                                      | download                                                                                                                     |
-| ------------------- | ------------------------------------------------ | ---------------------------------------------------------------------------------------------------------------------------- |
-| HugeGraph-Server    | HugeGraph 的主程序                               | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)           |
+| components          | description                  | download                                                                                                                     |
+|---------------------|------------------------------|------------------------------------------------------------------------------------------------------------------------------|
+| HugeGraph-Server    | HugeGraph 的主程序               | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)           |
 | HugeGraph-Toolchain | 数据导入/导出/备份, Web 可视化图形界面等工具合集 | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-toolchain-incubating-1.0.0.tar.gz) |
 
 ### Binary Versions mapping
 
 | Version | Release Date | server                                                                                                                                                                                                                                                                                                                                                                     | toolchain                                                                                            [...]
-| ------- | ------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------- [...]
+|---------|--------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------- [...]
 | 1.0.0   | 2022-12-31   | [[Binary](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)] [[Sign](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz.asc)] [[SHA512](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz.sha512)] | [[Binary](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-toolchai [...]
 
 ### Source Versions mapping
 
 | Version | Release Date | server                                                                                                                                                                                                                                                                                                                                                                                 | toolchain                                                                                [...]
-| ------- | ------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------- [...]
+|---------|--------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------- [...]
 | 1.0.0   | 2022-12-31   | [[Source](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz)] [[Sign](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz.asc)] [[SHA512](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz.sha512)] | [[Source](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegr [...]
 
 ---
@@ -30,7 +30,7 @@ The latest HugeGraph: **1.0.0**, released on _2022-12-31_.
 ### 旧版本下载地址 (Outdated Versions mapping)
 
 | server                                                                                             | client                                                                                 | loader                                                                                                           | hubble                                                                                                             | common                                                               [...]
-| -------------------------------------------------------------------------------------------------- | -------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------ | -------------------------------------------------------------------- [...]
+|----------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------- [...]
 | [0.12.0](https://github.com/hugegraph/hugegraph/releases/download/v0.12.0/hugegraph-0.12.0.tar.gz) | [2.0.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/2.0.1) | [0.12.0](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.12.0/hugegraph-loader-0.12.0.tar.gz) | [1.6.0](https://github.com/hugegraph/hugegraph-hubble/releases/download/v1.6.0/hugegraph-hubble-1.6.0.tar.gz)      | [2.0.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugeg [...]
 | [0.11.2](https://github.com/hugegraph/hugegraph/releases/download/v0.11.2/hugegraph-0.11.2.tar.gz) | [1.9.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/1.9.1) | [0.11.1](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.11.1/hugegraph-loader-0.11.1.tar.gz) | [1.5.0](https://github.com/hugegraph/hugegraph-hubble/releases/download/v1.5.0/hugegraph-hubble-1.5.0.tar.gz)      | [1.8.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugeg [...]
 | [0.10.4](https://github.com/hugegraph/hugegraph/releases/download/v0.10.4/hugegraph-0.10.4.tar.gz) | [1.8.0](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/1.8.0) | [0.10.1](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.10.1/hugegraph-loader-0.10.1.tar.gz) | [0.10.0](https://github.com/hugegraph/hugegraph-studio/releases/download/v0.10.0/hugegraph-studio-0.10.0.tar.gz)   | [1.6.16](https://mvnrepository.com/artifact/com.baidu.hugegraph/huge [...]
diff --git a/content/cn/docs/guides/architectural.md b/content/cn/docs/guides/architectural.md
index f989dd30..6e7663d1 100644
--- a/content/cn/docs/guides/architectural.md
+++ b/content/cn/docs/guides/architectural.md
@@ -8,18 +8,18 @@ weight: 1
 
 作为一款通用的图数据库产品,HugeGraph需具备图数据的基本功能,如下图所示。HugeGraph包括三个层次的功能,分别是存储层、计算层和用户接口层。 HugeGraph支持OLTP和OLAP两种图计算类型,其中OLTP实现了[Apache TinkerPop3](https://tinkerpop.apache.org)框架,并支持[Gremlin](https://tinkerpop.apache.org/gremlin.html)查询语言。 OLAP计算是基于SparkGraphX实现。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/design/architectural-overview.png" alt="image">
-</center>
+</div>
 
 
 ### 2 组件
 
 HugeGraph的主要功能分为HugeCore、ApiServer、HugeGraph-Client、HugeGraph-Loader和HugeGraph-Studio等组件构成,各组件之间的通信关系如下图所示。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/design/architectural-component.png" alt="image">
-</center>
+</div>
 
 
 - HugeCore :HugeGraph的核心模块,TinkerPop的接口主要在该模块中实现。HugeCore的功能涵盖包括OLTP和OLAP两个部分。
diff --git a/content/cn/docs/guides/faq.md b/content/cn/docs/guides/faq.md
index 9503ea8e..907f4fb7 100644
--- a/content/cn/docs/guides/faq.md
+++ b/content/cn/docs/guides/faq.md
@@ -10,7 +10,7 @@ weight: 5
 
 - 启动服务时提示:`xxx (core dumped) xxx`
 
-  请检查JDK版本是否为1.8  
+  请检查JDK版本是否为 Java11 (至少是Java8)
 
 - 启动服务成功了,但是操作图时有类似于"无法连接到后端或连接未打开"的提示
 
diff --git a/content/cn/docs/language/hugegraph-example.md b/content/cn/docs/language/hugegraph-example.md
index 050b8cad..7ad5bdab 100644
--- a/content/cn/docs/language/hugegraph-example.md
+++ b/content/cn/docs/language/hugegraph-example.md
@@ -26,9 +26,9 @@ HugeGraph相对于TitanDB而言,其主要特点如下:
 
 本示例通过Property Graph Model图数据模型来描述希腊神话中各人物角色的关系(也被成为人物关系图谱),具体关系详见下图。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/graph-of-gods.png" alt="image">
-</center>
+</div>
 
 
 其中,圆形节点代表实体(Vertex),箭头代表关系(Edge),方框的内容为属性。
diff --git a/content/cn/docs/language/hugegraph-gremlin.md b/content/cn/docs/language/hugegraph-gremlin.md
index 47b9ac68..9eda438f 100644
--- a/content/cn/docs/language/hugegraph-gremlin.md
+++ b/content/cn/docs/language/hugegraph-gremlin.md
@@ -23,7 +23,7 @@ HugeGraph实现了TinkerPop框架,但是并没有实现TinkerPop所有的特
 | Computer             | Determines if the {@code Graph} implementation supports {@link GraphComputer} based processing                                                                                                                                                                                                                                                                                                                                                                | false   |
 | Transactions         | Determines if the {@code Graph} implementations supports transactions.                                                                                                                                                                                                                                                                                                                                                                                        | true    |
 | Persistence          | Determines if the {@code Graph} implementation supports persisting it's contents natively to disk.This feature does not refer to every graph's ability to write to disk via the Gremlin IO packages(.e.g. GraphML), unless the graph natively persists to disk via those options somehow. For example,TinkerGraph does not support this feature as it is a pure in-sideEffects graph.                                                                         | true    |
-| ThreadedTransactions | Determines if the {@code Graph} implementation supports threaded transactions which allow a transaction be executed across multiple threads via {@link Transaction#createThreadedTx()}.                                                                                                                                                                                                                                                                     | false   |
+| ThreadedTransactions | Determines if the {@code Graph} implementation supports threaded transactions which allow a transaction be executed across multiple threads via {@link Transaction#createThreadedTx()}.                                                                                                                                                                                                                                                                       | false   |
 | ConcurrentAccess     | Determines if the {@code Graph} implementation supports more than one connection to the same instance at the same time. For example, Neo4j embedded does not support this feature because concurrent access to the same database files by multiple instances is not possible. However, Neo4j HA could support this feature as each new {@code Graph} instance coordinates with the Neo4j cluster allowing multiple instances to operate on the same database. | false   |
 
 ### Vertex Features
@@ -61,25 +61,25 @@ HugeGraph实现了TinkerPop框架,但是并没有实现TinkerPop所有的特
 
 ### Data Type Features
 
-| Name               | Description                                                                                                                                                                                                                                                         | Support |
-|--------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------|
-| BooleanValues      |                                                                                                                                                                                                                                                                     | true    |
-| ByteValues         |                                                                                                                                                                                                                                                                     | true    |
-| DoubleValues       |                                                                                                                                                                                                                                                                     | true    |
-| FloatValues        |                                                                                                                                                                                                                                                                     | true    |
-| IntegerValues      |                                                                                                                                                                                                                                                                     | true    |
-| LongValues         |                                                                                                                                                                                                                                                                     | true    |
+| Name               | Description                                                                                                                                                                                                                                                          | Support |
+|--------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------|
+| BooleanValues      |                                                                                                                                                                                                                                                                      | true    |
+| ByteValues         |                                                                                                                                                                                                                                                                      | true    |
+| DoubleValues       |                                                                                                                                                                                                                                                                      | true    |
+| FloatValues        |                                                                                                                                                                                                                                                                      | true    |
+| IntegerValues      |                                                                                                                                                                                                                                                                      | true    |
+| LongValues         |                                                                                                                                                                                                                                                                      | true    |
 | MapValues          | Supports setting of a {@code Map} value. The assumption is that the {@code Map} can contain arbitrary serializable values that may or may not be defined as a feature itself                                                                                         | false   |
 | MixedListValues    | Supports setting of a {@code List} value. The assumption is that the {@code List} can contain arbitrary serializable values that may or may not be defined as a feature itself. As this{@code List} is "mixed" it does not need to contain objects of the same type. | false   |
-| BooleanArrayValues |                                                                                                                                                                                                                                                                     | false   |
-| ByteArrayValues    |                                                                                                                                                                                                                                                                     | true    |
-| DoubleArrayValues  |                                                                                                                                                                                                                                                                     | false   |
-| FloatArrayValues   |                                                                                                                                                                                                                                                                     | false   |
-| IntegerArrayValues |                                                                                                                                                                                                                                                                     | false   |
-| LongArrayValues    |                                                                                                                                                                                                                                                                     | false   |
-| SerializableValues |                                                                                                                                                                                                                                                                     | false   |
-| StringArrayValues  |                                                                                                                                                                                                                                                                     | false   |
-| StringValues       |                                                                                                                                                                                                                                                                     | true    |
+| BooleanArrayValues |                                                                                                                                                                                                                                                                      | false   |
+| ByteArrayValues    |                                                                                                                                                                                                                                                                      | true    |
+| DoubleArrayValues  |                                                                                                                                                                                                                                                                      | false   |
+| FloatArrayValues   |                                                                                                                                                                                                                                                                      | false   |
+| IntegerArrayValues |                                                                                                                                                                                                                                                                      | false   |
+| LongArrayValues    |                                                                                                                                                                                                                                                                      | false   |
+| SerializableValues |                                                                                                                                                                                                                                                                      | false   |
+| StringArrayValues  |                                                                                                                                                                                                                                                                      | false   |
+| StringValues       |                                                                                                                                                                                                                                                                      | true    |
 | UniformListValues  | Supports setting of a {@code List} value. The assumption is that the {@code List} can contain arbitrary serializable values that may or may not be defined as a feature itself. As this{@code List} is "uniform" it must contain objects of the same type.           | false   |
 
 ### Gremlin的步骤
diff --git a/content/cn/docs/performance/api-preformance/hugegraph-api-0.2.md b/content/cn/docs/performance/api-preformance/hugegraph-api-0.2.md
index cd4730e2..0019f708 100644
--- a/content/cn/docs/performance/api-preformance/hugegraph-api-0.2.md
+++ b/content/cn/docs/performance/api-preformance/hugegraph-api-0.2.md
@@ -10,9 +10,9 @@ weight: 4
 
 起压和被压机器配置相同,基本参数如下:
 
-CPU                                          | Memory | 网卡
--------------------------------------------- | ------ | --------
-24 Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz | 61G    | 1000Mbps
+| CPU                                          | Memory | 网卡       |
+|----------------------------------------------|--------|----------|
+| 24 Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz | 61G    | 1000Mbps |
 
 测试工具:apache-Jmeter-2.5.1
 
@@ -38,7 +38,7 @@ CPU                                          | Memory | 网卡
 - Min -- 最小响应时间
 - Max -- 最大响应时间
 - Error -- 出错率
-- Troughput -- 吞吐量Â
+- Throughput -- 吞吐量Â
 - KB/sec -- 以流量做衡量的吞吐量
 
 _注:时间的单位均为ms_
@@ -47,11 +47,11 @@ _注:时间的单位均为ms_
 
 #### 2.1 schema
 
-Label         | Samples | Average | Median | 90%Line | Min | Max | Error% | Throughput | KB/sec
-------------- | ------- | ------- | ------ | ------- | --- | --- | ------ | ---------- | ------
-property_keys | 331000  | 1       | 1      | 2       | 0   | 172 | 0.00%  | 920.7/sec  | 178.1
-vertex_labels | 331000  | 1       | 2      | 2       | 1   | 126 | 0.00%  | 920.7/sec  | 193.4
-edge_labels   | 331000  | 2       | 2      | 3       | 1   | 158 | 0.00%  | 920.7/sec  | 242.8
+| Label         | Samples | Average | Median | 90%Line | Min | Max | Error% | Throughput | KB/sec |
+|---------------|---------|---------|--------|---------|-----|-----|--------|------------|--------|
+| property_keys | 331000  | 1       | 1      | 2       | 0   | 172 | 0.00%  | 920.7/sec  | 178.1  |
+| vertex_labels | 331000  | 1       | 2      | 2       | 1   | 126 | 0.00%  | 920.7/sec  | 193.4  |
+| edge_labels   | 331000  | 2       | 2      | 3       | 1   | 158 | 0.00%  | 920.7/sec  | 242.8  |
 
 结论:schema的接口,在1000并发持续5分钟的压力下,平均响应时间1-2ms,无压力
 
@@ -68,10 +68,10 @@ edge_labels   | 331000  | 2       | 2      | 3       | 1   | 158 | 0.00%  | 920.
 
 ###### 性能指标
 
-Label                  | Samples | Average | Median | 90%Line | Min | Max | Error% | Throughput | KB/sec
----------------------- | ------- | ------- | ------ | ------- | --- | --- | ------ | ---------- | ------
-single_insert_vertices | 331000  | 0       | 1      | 1       | 0   | 21  | 0.00%  | 920.7/sec  | 234.4
-single_insert_edges    | 331000  | 2       | 2      | 3       | 1   | 53  | 0.00%  | 920.7/sec  | 309.1
+| Label                  | Samples | Average | Median | 90%Line | Min | Max | Error% | Throughput | KB/sec |
+|------------------------|---------|---------|--------|---------|-----|-----|--------|------------|--------|
+| single_insert_vertices | 331000  | 0       | 1      | 1       | 0   | 21  | 0.00%  | 920.7/sec  | 234.4  |
+| single_insert_edges    | 331000  | 2       | 2      | 3       | 1   | 53  | 0.00%  | 920.7/sec  | 309.1  |
 
 ###### 结论
 
@@ -89,15 +89,15 @@ single_insert_edges    | 331000  | 2       | 2      | 3       | 1   | 53  | 0.00
 
 ###### 性能指标
 
-Concurrency  | Samples | Average | Median | 90%Line | Min | Max  | Error% | Throughput | KB/sec
------------- | ------- | ------- | ------ | ------- | --- | ---- | ------ | ---------- | ------
-2000(vertex) | 661916  | 1       | 1      | 1       | 0   | 3012 | 0.00%  | 1842.9/sec | 469.1
-4000(vertex) | 1316124 | 13      | 1      | 14      | 0   | 9023 | 0.00%  | 3673.1/sec | 935.0
-5000(vertex) | 1468121 | 1010    | 1135   | 1227    | 0   | 9223 | 0.06%  | 4095.6/sec | 1046.0
-7000(vertex) | 1378454 | 1617    | 1708   | 1886    | 0   | 9361 | 0.08%  | 3860.3/sec | 987.1
-2000(edge)   | 629399  | 953     | 1043   | 1113    | 1   | 9001 | 0.00%  | 1750.3/sec | 587.6
-3000(edge)   | 648364  | 2258    | 2404   | 2500    | 2   | 9001 | 0.00%  | 1810.7/sec | 607.9
-4000(edge)   | 649904  | 1992    | 2112   | 2211    | 1   | 9001 | 0.06%  | 1812.5/sec | 608.5
+| Concurrency  | Samples | Average | Median | 90%Line | Min | Max  | Error% | Throughput | KB/sec |
+|--------------|---------|---------|--------|---------|-----|------|--------|------------|--------|
+| 2000(vertex) | 661916  | 1       | 1      | 1       | 0   | 3012 | 0.00%  | 1842.9/sec | 469.1  |
+| 4000(vertex) | 1316124 | 13      | 1      | 14      | 0   | 9023 | 0.00%  | 3673.1/sec | 935.0  |
+| 5000(vertex) | 1468121 | 1010    | 1135   | 1227    | 0   | 9223 | 0.06%  | 4095.6/sec | 1046.0 |
+| 7000(vertex) | 1378454 | 1617    | 1708   | 1886    | 0   | 9361 | 0.08%  | 3860.3/sec | 987.1  |
+| 2000(edge)   | 629399  | 953     | 1043   | 1113    | 1   | 9001 | 0.00%  | 1750.3/sec | 587.6  |
+| 3000(edge)   | 648364  | 2258    | 2404   | 2500    | 2   | 9001 | 0.00%  | 1810.7/sec | 607.9  |
+| 4000(edge)   | 649904  | 1992    | 2112   | 2211    | 1   | 9001 | 0.06%  | 1812.5/sec | 608.5  |
 
 ###### 结论
 
@@ -123,10 +123,10 @@ Concurrency  | Samples | Average | Median | 90%Line | Min | Max  | Error% | Thro
 
 ###### 性能指标
 
-Label                 | Samples | Average | Median | 90%Line | Min | Max   | Error% | Throughput | KB/sec
---------------------- | ------- | ------- | ------ | ------- | --- | ----- | ------ | ---------- | ------
-batch_insert_vertices | 37162   | 8959    | 9595   | 9704    | 17  | 9852  | 0.00%  | 103.4/sec  | 393.3
-batch_insert_edges    | 10800   | 31849   | 34544  | 35132   | 435 | 35747 | 0.00%  | 28.8/sec   | 814.9
+| Label                 | Samples | Average | Median | 90%Line | Min | Max   | Error% | Throughput | KB/sec |
+|-----------------------|---------|---------|--------|---------|-----|-------|--------|------------|--------|
+| batch_insert_vertices | 37162   | 8959    | 9595   | 9704    | 17  | 9852  | 0.00%  | 103.4/sec  | 393.3  |
+| batch_insert_edges    | 10800   | 31849   | 34544  | 35132   | 435 | 35747 | 0.00%  | 28.8/sec   | 814.9  |
 
 ###### 结论
 
diff --git a/content/cn/docs/performance/api-preformance/hugegraph-api-0.4.4.md b/content/cn/docs/performance/api-preformance/hugegraph-api-0.4.4.md
index a2742bc3..7e8c0c4c 100644
--- a/content/cn/docs/performance/api-preformance/hugegraph-api-0.4.4.md
+++ b/content/cn/docs/performance/api-preformance/hugegraph-api-0.4.4.md
@@ -8,10 +8,10 @@ weight: 3
 
 被压机器信息
 
-机器编号                              | CPU                                          | Memory | 网卡        | 磁盘
---------------------------------- | -------------------------------------------- | ------ | --------- | ------------------
-1 | 24 Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz | 61G    | 1000Mbps  | 1.4T HDD
-2 | 48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD
+| 机器编号 | CPU                                          | Memory | 网卡        | 磁盘                 |
+|------|----------------------------------------------|--------|-----------|--------------------|
+| 1    | 24 Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz | 61G    | 1000Mbps  | 1.4T HDD           |
+| 2    | 48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD |
 
 - 起压力机器信息:与编号 1 机器同配置
 - 测试工具:apache-Jmeter-2.5.1
@@ -59,9 +59,9 @@ weight: 3
 
 ###### 顶点和边的最大插入速度(高性能服务器,使用SSD存储RocksDB数据):
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.4.4/best.png" alt="image">
-</center>
+</div>
 
 
 ###### 结论:
@@ -71,9 +71,9 @@ weight: 3
 
 **1\. CPU和内存对插入性能的影响(服务器都使用HDD存储RocksDB数据,批量插入)**
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.4.4/cpu-memory.png" alt="image">
-</center>
+</div>
 
 
 ###### 结论:
@@ -84,9 +84,9 @@ weight: 3
 
 **2\. SSD和HDD对插入性能的影响(高性能服务器,批量插入)**
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.4.4/ssd.png" alt="image">
-</center>
+</div>
 
 
 ###### 结论:
@@ -96,9 +96,9 @@ weight: 3
 
 **3\. 不同并发线程数对插入性能的影响(普通服务器,使用HDD存储RocksDB数据)**
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.4.4/threads-batch.png" alt="image">
-</center>
+</div>
 
 
 ###### 结论:
@@ -119,9 +119,9 @@ weight: 3
 - 持续时间:5min
 - 服务异常标志:错误率大于0.00%
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.4.4/threads-single.png" alt="image">
-</center>
+</div>
 
 ###### 结论:
 
diff --git a/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-Cassandra.md b/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-Cassandra.md
index 13f5083e..1a847102 100644
--- a/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-Cassandra.md
+++ b/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-Cassandra.md
@@ -8,9 +8,9 @@ weight: 2
 
 被压机器信息
 
-CPU                                          | Memory | 网卡      | 磁盘
--------------------------------------------- | ------ | --------- | ------------------
-48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD
+| CPU                                          | Memory | 网卡        | 磁盘                 |
+|----------------------------------------------|--------|-----------|--------------------|
+| 48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD |
 
 - 起压力机器信息:与被压机器同配置
 - 测试工具:apache-Jmeter-2.5.1
@@ -57,9 +57,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 顶点的最大插入速度:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/vertex_batch.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -68,9 +68,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 边的最大插入速度
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/edge_batch.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -92,9 +92,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 顶点的单条插入
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/vertex_single.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -103,9 +103,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 边的单条插入
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/edge_single.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -127,9 +127,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 顶点的按id查询
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/vertex_id_query.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -138,9 +138,9 @@ CPU                                          | Memory | 网卡      | 磁盘
 
 ###### 边的按id查询
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/cassandra/edge_id_query.png" alt="image">
-</center>
+</div>
 
 ####### 结论:
 
diff --git a/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-RocksDB.md b/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-RocksDB.md
index a59a0970..0e88cb89 100644
--- a/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-RocksDB.md
+++ b/content/cn/docs/performance/api-preformance/hugegraph-api-0.5.6-RocksDB.md
@@ -8,9 +8,9 @@ weight: 1
 
 被压机器信息
 
-CPU                                          | Memory | 网卡        | 磁盘
--------------------------------------------- | ------ | --------- | ------------------
-48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD
+| CPU                                          | Memory | 网卡        | 磁盘                 |
+|----------------------------------------------|--------|-----------|--------------------|
+| 48 Intel(R) Xeon(R) CPU E5-2650 v4 @ 2.20GHz | 128G   | 10000Mbps | 750GB SSD,2.7T HDD |
 
 - 起压力机器信息:与被压机器同配置
 - 测试工具:apache-Jmeter-2.5.1
@@ -57,9 +57,9 @@ CPU                                          | Memory | 网卡        | 磁盘
 
 ###### 顶点的最大插入速度:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/rocksdb/vertex_batch.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
@@ -68,9 +68,9 @@ CPU                                          | Memory | 网卡        | 磁盘
 
 ###### 边的最大插入速度
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/rocksdb/edge_batch.png" alt="image">
-</center>
+</div>
 
 ####### 结论:
 
@@ -91,9 +91,9 @@ CPU                                          | Memory | 网卡        | 磁盘
 
 ###### 顶点的单条插入
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/API-perf/v0.5.6/rocksdb/vertex_single.png" alt="image">
-</center>
+</div>
 
 
 ####### 结论:
diff --git a/content/cn/docs/quickstart/hugegraph-client.md b/content/cn/docs/quickstart/hugegraph-client.md
index 5d2125fb..cb4c4bf2 100644
--- a/content/cn/docs/quickstart/hugegraph-client.md
+++ b/content/cn/docs/quickstart/hugegraph-client.md
@@ -10,8 +10,8 @@ HugeGraph-Client向HugeGraph-Server发出HTTP请求,获取并解析Server的
 
 ### 2 环境要求
 
-- jdk1.8
-- maven-3.3.9
+- java 11 (兼容 java 8)
+- maven 3.5+
 
 ### 3 使用流程
 
diff --git a/content/cn/docs/quickstart/hugegraph-computer.md b/content/cn/docs/quickstart/hugegraph-computer.md
index d3f4597f..02233fc3 100644
--- a/content/cn/docs/quickstart/hugegraph-computer.md
+++ b/content/cn/docs/quickstart/hugegraph-computer.md
@@ -22,7 +22,7 @@ weight: 7
 
 ### 2.1 在本地运行 PageRank 算法
 
-> 要使用 HugeGraph-Computer 运行算法,您需要安装 64 位 JREJDK 11 或更高版本。
+> 要使用 HugeGraph-Computer 运行算法,您需要安装 64 位 Java 11 或更高版本。
 >
 > 还需要首先部署 HugeGraph-Server 和 [Etcd](https://etcd.io/docs/v3.5/quickstart/).
 
diff --git a/content/cn/docs/quickstart/hugegraph-hubble.md b/content/cn/docs/quickstart/hugegraph-hubble.md
index bab2a038..1aad4800 100644
--- a/content/cn/docs/quickstart/hugegraph-hubble.md
+++ b/content/cn/docs/quickstart/hugegraph-hubble.md
@@ -36,9 +36,9 @@ HugeGraph是一款面向分析型,支持批量操作的图数据库系统,
 
 平台的模块使用流程如下:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/2平台使用流程.png" alt="image">
-</center>
+</div>
 
 
 ### 3	平台使用说明
@@ -46,42 +46,42 @@ HugeGraph是一款面向分析型,支持批量操作的图数据库系统,
 ##### 3.1.1	图创建
 图管理模块下,点击【创建图】,通过填写图ID、图名称、主机名、端口号、用户名、密码的信息,实现多图的连接。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/311图创建.png" alt="image">
-</center>
+</div>
 
 
 创建图填写内容如下:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/311图创建2.png" alt="image">
-</center>
+</div>
 
 
 ##### 3.1.2	图访问
 实现图空间的信息访问,进入后,可进行图的多维查询分析、元数据管理、数据导入、算法分析等操作。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/312图访问.png" alt="image">
-</center>
+</div>
 
 
 ##### 3.1.3	图管理
 1. 用户通过对图的概览、搜索以及单图的信息编辑与删除,实现图的统一管理。
 2. 搜索范围:可对图名称和ID进行搜索。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/313图管理.png" alt="image">
-</center>
+</div>
 
 
 #### 3.2	元数据建模(列表+图模式)
 ##### 3.2.1	模块入口
 左侧导航处:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/321元数据入口.png" alt="image">
-</center>
+</div>
 
 
 ##### 3.2.2	属性类型
@@ -91,16 +91,16 @@ HugeGraph是一款面向分析型,支持批量操作的图数据库系统,
 
 列表模式:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/3221属性创建.png" alt="image">
-</center>
+</div>
 
 
 图模式:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/3221属性创建2.png" alt="image">
-</center>
+</div>
 
 
 ###### 3.2.2.2	复用
@@ -109,16 +109,16 @@ HugeGraph是一款面向分析型,支持批量操作的图数据库系统,
 
 选择复用项:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/3222属性复用.png" alt="image">
-</center>
+</div>
 
 
 校验复用项:
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/images-hubble/3222属性复用2.png" alt="image">
-</center>
+</div>
 
 
 ###### 3.2.2.3	管理
diff --git a/content/cn/docs/quickstart/hugegraph-loader.md b/content/cn/docs/quickstart/hugegraph-loader.md
index 4c07edae..d6fe8470 100644
--- a/content/cn/docs/quickstart/hugegraph-loader.md
+++ b/content/cn/docs/quickstart/hugegraph-loader.md
@@ -47,16 +47,16 @@ $ git clone https://github.com/hugegraph/hugegraph-loader.git
 由于Oracle ojdbc license的限制,需要手动安装ojdbc到本地maven仓库。
 访问[Oracle jdbc 下载](https://www.oracle.com/database/technologies/appdev/jdbc-downloads.html) 页面。选择Oracle Database 12c Release 2 (12.2.0.1) drivers,如下图所示。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/oracle-download.png" alt="image">
-</center>
+</div>
 
 
 打开链接后,选择“ojdbc8.jar”, 如下图所示。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/ojdbc8.png" alt="image">
-</center>
+</div>
 
 
  把ojdbc8安装到本地maven仓库,进入``ojdbc8.jar``所在目录,执行以下命令。
@@ -87,10 +87,10 @@ mvn clean package -DskipTests
 比如想创建一个拥有两类顶点及两类边的图,顶点是"人"和"软件",边是"人认识人"和"人创造软件",并且这些顶点和边都带有一些属性,比如顶点"人"有:"姓名"、"年龄"等属性,
 "软件"有:"名字"、"售卖价格"等属性;边"认识"有: "日期"属性等。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/demo-graph-model.png" alt="image">
   <p>示例图模型</p>
-</center>
+</div>
 
 
 在设计好了图模型之后,我们可以用`groovy`编写出`schema`的定义,并保存至文件中,这里命名为`schema.groovy`。
@@ -535,7 +535,7 @@ Office,388
 }
 ```
 
-映射文件 1.0 版本是以顶点和边为中心,设置输入源;而 2.0 版本是以输入源为中心,设置顶点和边映射。有些输入源(比如一个文件)既能生成顶点,也能生成边,如果用 1.0 版的格式写,就需要在 vertex 和 egde 映射块中各写一次 input 块,这两次的 input 块是完全一样的;而 2.0 版本只需要写一次 input。所以 2.0 版相比于 1.0 版,能省掉一些 input 的重复书写。
+映射文件 1.0 版本是以顶点和边为中心,设置输入源;而 2.0 版本是以输入源为中心,设置顶点和边映射。有些输入源(比如一个文件)既能生成顶点,也能生成边,如果用 1.0 版的格式写,就需要在 vertex 和 edge 映射块中各写一次 input 块,这两次的 input 块是完全一样的;而 2.0 版本只需要写一次 input。所以 2.0 版相比于 1.0 版,能省掉一些 input 的重复书写。
 
 在 hugegraph-loader-{version} 的 bin 目录下,有一个脚本工具 `mapping-convert.sh` 能直接将 1.0 版本的映射文件转换为 2.0 版本的,使用方式如下:
 
@@ -596,41 +596,41 @@ bin/mapping-convert.sh struct.json
 
 **MYSQL**
 
-| 节点 | 固定值或常见值 |
-| --- | --- | 
-| vendor | MYSQL |
-| driver | com.mysql.cj.jdbc.Driver |
-| url | jdbc:mysql://127.0.0.1:3306 |
+| 节点     | 固定值或常见值                     |
+|--------|-----------------------------| 
+| vendor | MYSQL                       |
+| driver | com.mysql.cj.jdbc.Driver    |
+| url    | jdbc:mysql://127.0.0.1:3306 |
 
 schema: 可空,若填写必须与database的值一样
 
 **POSTGRESQL**
 
-| 节点 | 固定值或常见值 |
-| --- | --- | 
-| vendor | POSTGRESQL |
-| driver | org.postgresql.Driver |
-| url | jdbc:postgresql://127.0.0.1:5432 |
+| 节点     | 固定值或常见值                          |
+|--------|----------------------------------| 
+| vendor | POSTGRESQL                       |
+| driver | org.postgresql.Driver            |
+| url    | jdbc:postgresql://127.0.0.1:5432 |
 
 schema: 可空,默认值为“public”
 
 **ORACLE**
 
-| 节点 | 固定值或常见值 |
-| --- | --- | 
-| vendor | ORACLE |
-| driver | oracle.jdbc.driver.OracleDriver |
-| url | jdbc:oracle:thin:@127.0.0.1:1521 |
+| 节点     | 固定值或常见值                          |
+|--------|----------------------------------| 
+| vendor | ORACLE                           |
+| driver | oracle.jdbc.driver.OracleDriver  |
+| url    | jdbc:oracle:thin:@127.0.0.1:1521 |
 
 schema: 可空,默认值与用户名相同
 
 **SQLSERVER**
 
-| 节点 | 固定值或常见值 |
-| --- | --- | 
-| vendor | SQLSERVER |
+| 节点     | 固定值或常见值                                      |
+|--------|----------------------------------------------| 
+| vendor | SQLSERVER                                    |
 | driver | com.microsoft.sqlserver.jdbc.SQLServerDriver |
-| url | jdbc:sqlserver://127.0.0.1:1433 |
+| url    | jdbc:sqlserver://127.0.0.1:1433              |
 
 schema: 必填
 
diff --git a/content/cn/docs/quickstart/hugegraph-server.md b/content/cn/docs/quickstart/hugegraph-server.md
index 28158bf8..66d5601f 100644
--- a/content/cn/docs/quickstart/hugegraph-server.md
+++ b/content/cn/docs/quickstart/hugegraph-server.md
@@ -14,9 +14,9 @@ Core模块是Tinkerpop接口的实现,Backend模块用于管理数据存储,
 
 ### 2 依赖
 
-#### 2.1 安装JDK-1.8
+#### 2.1 安装 Java 11 (JDK 11)
 
-HugeGraph-Server 基于jdk-1.8开发,代码用到了较多jdk-1.8中的类和方法,请用户自行安装配置。
+请优先考虑在 Java11 的环境上启动 `HugeGraph-Server`, 目前同时保留对 Java8 的兼容 
 
 **在往下阅读之前务必执行`java -version`命令查看jdk版本**
 
diff --git a/content/en/about/_index.html b/content/en/about/_index.html
index 141fcd81..3e7e8773 100644
--- a/content/en/about/_index.html
+++ b/content/en/about/_index.html
@@ -1,6 +1,7 @@
 ---
 title: About Apache HugeGraph
 linkTitle: About
+draft: true
 menu:
   main:
     weight: 10
diff --git a/content/en/docs/changelog/hugegraph-0.11.2-release-notes.md b/content/en/docs/changelog/hugegraph-0.11.2-release-notes.md
index 17de550a..9802d28c 100644
--- a/content/en/docs/changelog/hugegraph-0.11.2-release-notes.md
+++ b/content/en/docs/changelog/hugegraph-0.11.2-release-notes.md
@@ -17,7 +17,7 @@ weight: 2
 - 支持共同邻居,Jaccard 相似度,全部最短路径,带权最短路径和单源最短路径5种遍历算法(hugegraph #936,hugegraph-client #80)
 - 支持用户认证和细粒度权限控制(hugegraph #749,hugegraph #985,hugegraph-client #81)
 - 支持遍历 API 的顶点计数功能(hugegraph #995,hugegraph-client #84)
-- 支持 HTTPS 协议(hugegrap #1036,hugegraph-client #85)
+- 支持 HTTPS 协议(hugegraph #1036,hugegraph-client #85)
 - 支持创建索引时控制是否重建索引(hugegraph #1106,hugegraph-client #91)
 - 支持定制的 kout/kneighbor,多点最短路径,最相似 Jaccard 点和模板路径5种遍历算法(hugegraph #1174,hugegraph-client #100,hugegraph-client #106)
 
diff --git a/content/en/docs/changelog/hugegraph-0.12.0-release-notes.md b/content/en/docs/changelog/hugegraph-0.12.0-release-notes.md
index 19a672cb..08bdd099 100644
--- a/content/en/docs/changelog/hugegraph-0.12.0-release-notes.md
+++ b/content/en/docs/changelog/hugegraph-0.12.0-release-notes.md
@@ -1,7 +1,7 @@
 ---
 title: "HugeGraph 0.12 Release Notes"
 linkTitle: "Release-0.12.0"
-weight: 1
+weight: 2
 ---
 
 ### API & Client
@@ -205,7 +205,7 @@ weight: 1
 - 修复 trust store file 路径问题(hugegraph-tools #75)
 - 支持权限信息的备份恢复(hugegraph-tools #76)
 - 支持无参数的 Printer 打印(hugegraph-tools #79)
-- 修复 MacOS free_memory 计算问题(hugegraph-tools #82)
+- 修复 macOS free_memory 计算问题(hugegraph-tools #82)
 - 支持备份恢复时指定线程数hugegraph-tools #83)
 - 支持动态创建图、克隆图、删除图等命令(hugegraph-tools #95)
 
diff --git a/content/en/docs/changelog/hugegraph-1.0.0-release-notes.md b/content/en/docs/changelog/hugegraph-1.0.0-release-notes.md
index 8a7e16e0..4d857825 100644
--- a/content/en/docs/changelog/hugegraph-1.0.0-release-notes.md
+++ b/content/en/docs/changelog/hugegraph-1.0.0-release-notes.md
@@ -1,7 +1,7 @@
 ---
 title: "HugeGraph 1.0.0 Release Notes"
 linkTitle: "Release-1.0.0"
-weight: 0
+weight: 1
 ---
 
 > TODO: update the summary soon, please check the detail in each repository first, thanks
diff --git a/content/en/docs/clients/hugegraph-client.md b/content/en/docs/clients/hugegraph-client.md
index 7525c399..18a83a0f 100644
--- a/content/en/docs/clients/hugegraph-client.md
+++ b/content/en/docs/clients/hugegraph-client.md
@@ -139,7 +139,7 @@ The constraint information that VertexLabel allows to define include: name, idSt
 |--------------------------|-------|----------|
 | vertexLabel(String name) | name  | y        |
 
-- idStrategy: Each VertexLabel can choose its own ID strategy. There are currently three strategies to choose from, namely Automatic (automatically generated), Customize (user input) and PrimaryKey (primary attribute key). Among them, Automatic uses the Snowflake algorithm to generate Id, Customize requires the user to pass in the Id of string or number type, and PrimaryKey allows the user to select several properties of VertexLabel as the basis for differentiation. HugeGraph will be spl [...]
+- idStrategy: Each VertexLabel can choose its own ID strategy. There are currently three strategies to choose from, namely Automatic (automatically generated), Customize (user input) and PrimaryKey (primary attribute key). Among them, Automatic uses the Snowflake algorithm to generate ID, Customize requires the user to pass in the ID of string or number type, and PrimaryKey allows the user to select several properties of VertexLabel as the basis for differentiation. HugeGraph will be spl [...]
 
 | interface            | idStrategy       | description                                             |
 |----------------------|------------------|---------------------------------------------------------|
diff --git a/content/en/docs/config/config-computer.md b/content/en/docs/config/config-computer.md
index b6c60202..08f804c1 100644
--- a/content/en/docs/config/config-computer.md
+++ b/content/en/docs/config/config-computer.md
@@ -6,172 +6,172 @@ weight: 5
 
 ### Computer Config Options
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|algorithm.message_class|org.apache.hugegraph.computer.core.config.Null|The class of message passed when compute vertex.|
-|algorithm.params_class|org.apache.hugegraph.computer.core.config.Null|The class used to transfer algorithms' parameters before algorithm been run.|
-|algorithm.result_class|org.apache.hugegraph.computer.core.config.Null|The class of vertex's value, the instance is used to store computation result for the vertex.|
-|allocator.max_vertices_per_thread|10000|Maximum number of vertices per thread processed in each memory allocator|
-|bsp.etcd_endpoints|http://localhost:2379|The end points to access etcd.|
-|bsp.log_interval|30000|The log interval(in ms) to print the log while waiting bsp event.|
-|bsp.max_super_step|10|The max super step of the algorithm.|
-|bsp.register_timeout|300000|The max timeout to wait for master and works to register.|
-|bsp.wait_master_timeout|86400000|The max timeout(in ms) to wait for master bsp event.|
-|bsp.wait_workers_timeout|86400000|The max timeout to wait for workers bsp event.|
-|hgkv.max_data_block_size|65536|The max byte size of hgkv-file data block.|
-|hgkv.max_file_size|2147483648|The max number of bytes in each hgkv-file.|
-|hgkv.max_merge_files|10|The max number of files to merge at one time.|
-|hgkv.temp_file_dir|/tmp/hgkv|This folder is used to store temporary files, temporary files will be generated during the file merging process.|
-|hugegraph.name|hugegraph|The graph name to load data and write results back.|
-|hugegraph.url|http://127.0.0.1:8080|The hugegraph url to load data and write results back.|
-|input.edge_direction|OUT|The data of the edge in which direction is loaded, when the value is BOTH, the edges in both OUT and IN direction will be loaded.|
-|input.edge_freq|MULTIPLE|The frequency of edges can exist between a pair of vertices, allowed values: [SINGLE, SINGLE_PER_LABEL, MULTIPLE]. SINGLE means that only one edge can exist between a pair of vertices, use sourceId + targetId to identify it; SINGLE_PER_LABEL means that each edge label can exist one edge between a pair of vertices, use sourceId + edgelabel + targetId to identify it; MULTIPLE means that many edge can exist between a pair of vertices, use sourceId + edgelabel + sor [...]
-|input.filter_class|org.apache.hugegraph.computer.core.input.filter.DefaultInputFilter|The class to create input-filter object, input-filter is used to Filter vertex edges according to user needs.|
-|input.loader_schema_path||The schema path of loader input, only takes effect when the input.source_type=loader is enabled|
-|input.loader_struct_path||The struct path of loader input, only takes effect when the input.source_type=loader is enabled|
-|input.max_edges_in_one_vertex|200|The maximum number of adjacent edges allowed to be attached to a vertex, the adjacent edges will be stored and transferred together as a batch unit.|
-|input.source_type|hugegraph-server|The source type to load input data, allowed values: ['hugegraph-server', 'hugegraph-loader'], the 'hugegraph-loader' means use hugegraph-loader load data from HDFS or file, if use 'hugegraph-loader' load data then please config 'input.loader_struct_path' and 'input.loader_schema_path'.|
-|input.split_fetch_timeout|300|The timeout in seconds to fetch input splits|
-|input.split_max_splits|10000000|The maximum number of input splits|
-|input.split_page_size|500|The page size for streamed load input split data|
-|input.split_size|1048576|The input split size in bytes|
-|job.id|local_0001|The job id on Yarn cluster or K8s cluster.|
-|job.partitions_count|1|The partitions count for computing one graph algorithm job.|
-|job.partitions_thread_nums|4|The number of threads for partition parallel compute.|
-|job.workers_count|1|The workers count for computing one graph algorithm job.|
-|master.computation_class|org.apache.hugegraph.computer.core.master.DefaultMasterComputation|Master-computation is computation that can determine whether to continue next superstep. It runs at the end of each superstep on master.|
-|output.batch_size|500|The batch size of output|
-|output.batch_threads|1|The threads number used to batch output|
-|output.hdfs_core_site_path||The hdfs core site path.|
-|output.hdfs_delimiter|,|The delimiter of hdfs output.|
-|output.hdfs_kerberos_enable|false|Is Kerberos authentication enabled for Hdfs.|
-|output.hdfs_kerberos_keytab||The Hdfs's key tab file for kerberos authentication.|
-|output.hdfs_kerberos_principal||The Hdfs's principal for kerberos authentication.|
-|output.hdfs_krb5_conf|/etc/krb5.conf|Kerberos configuration file.|
-|output.hdfs_merge_partitions|true|Whether merge output files of multiple partitions.|
-|output.hdfs_path_prefix|/hugegraph-computer/results|The directory of hdfs output result.|
-|output.hdfs_replication|3|The replication number of hdfs.|
-|output.hdfs_site_path||The hdfs site path.|
-|output.hdfs_url|hdfs://127.0.0.1:9000|The hdfs url of output.|
-|output.hdfs_user|hadoop|The hdfs user of output.|
-|output.output_class|org.apache.hugegraph.computer.core.output.LogOutput|The class to output the computation result of each vertex. Be called after iteration computation.|
-|output.result_name|value|The value is assigned dynamically by #name() of instance created by WORKER_COMPUTATION_CLASS.|
-|output.result_write_type|OLAP_COMMON|The result write-type to output to hugegraph, allowed values are: [OLAP_COMMON, OLAP_SECONDARY, OLAP_RANGE].|
-|output.retry_interval|10|The retry interval when output failed|
-|output.retry_times|3|The retry times when output failed|
-|output.single_threads|1|The threads number used to single output|
-|output.thread_pool_shutdown_timeout|60|The timeout seconds of output threads pool shutdown|
-|output.with_adjacent_edges|false|Output the adjacent edges of the vertex or not|
-|output.with_edge_properties|false|Output the properties of the edge or not|
-|output.with_vertex_properties|false|Output the properties of the vertex or not|
-|sort.thread_nums|4|The number of threads performing internal sorting.|
-|transport.client_connect_timeout|3000|The timeout(in ms) of client connect to server.|
-|transport.client_threads|4|The number of transport threads for client.|
-|transport.close_timeout|10000|The timeout(in ms) of close server or close client.|
-|transport.finish_session_timeout|0|The timeout(in ms) to finish session, 0 means using (transport.sync_request_timeout * transport.max_pending_requests).|
-|transport.heartbeat_interval|20000|The minimum interval(in ms) between heartbeats on client side.|
-|transport.io_mode|AUTO|The network IO Mode, either 'NIO', 'EPOLL', 'AUTO', the 'AUTO' means selecting the property mode automatically.|
-|transport.max_pending_requests|8|The max number of client unreceived ack, it will trigger the sending unavailable if the number of unreceived ack >= max_pending_requests.|
-|transport.max_syn_backlog|511|The capacity of SYN queue on server side, 0 means using system default value.|
-|transport.max_timeout_heartbeat_count|120|The maximum times of timeout heartbeat on client side, if the number of timeouts waiting for heartbeat response continuously > max_heartbeat_timeouts the channel will be closed from client side.|
-|transport.min_ack_interval|200|The minimum interval(in ms) of server reply ack.|
-|transport.min_pending_requests|6|The minimum number of client unreceived ack, it will trigger the sending available if the number of unreceived ack < min_pending_requests.|
-|transport.network_retries|3|The number of retry attempts for network communication,if network unstable.|
-|transport.provider_class|org.apache.hugegraph.computer.core.network.netty.NettyTransportProvider|The transport provider, currently only supports Netty.|
-|transport.receive_buffer_size|0|The size of socket receive-buffer in bytes, 0 means using system default value.|
-|transport.recv_file_mode|true|Whether enable receive buffer-file mode, it will receive buffer write file from socket by zero-copy if enable.|
-|transport.send_buffer_size|0|The size of socket send-buffer in bytes, 0 means using system default value.|
-|transport.server_host|127.0.0.1|The server hostname or ip to listen on to transfer data.|
-|transport.server_idle_timeout|360000|The max timeout(in ms) of server idle.|
-|transport.server_port|0|The server port to listen on to transfer data. The system will assign a random port if it's set to 0.|
-|transport.server_threads|4|The number of transport threads for server.|
-|transport.sync_request_timeout|10000|The timeout(in ms) to wait response after sending sync-request.|
-|transport.tcp_keep_alive|true|Whether enable TCP keep-alive.|
-|transport.transport_epoll_lt|false|Whether enable EPOLL level-trigger.|
-|transport.write_buffer_high_mark|67108864|The high water mark for write buffer in bytes, it will trigger the sending unavailable if the number of queued bytes > write_buffer_high_mark.|
-|transport.write_buffer_low_mark|33554432|The low water mark for write buffer in bytes, it will trigger the sending available if the number of queued bytes < write_buffer_low_mark.org.apache.hugegraph.config.OptionChecker$$Lambda$97/0x00000008001c8440@776a6d9b|
-|transport.write_socket_timeout|3000|The timeout(in ms) to write data to socket buffer.|
-|valuefile.max_segment_size|1073741824|The max number of bytes in each segment of value-file.|
-|worker.combiner_class|org.apache.hugegraph.computer.core.config.Null|Combiner can combine messages into one value for a vertex, for example page-rank algorithm can combine messages of a vertex to a sum value.|
-|worker.computation_class|org.apache.hugegraph.computer.core.config.Null|The class to create worker-computation object, worker-computation is used to compute each vertex in each superstep.|
-|worker.data_dirs|[jobs]|The directories separated by ',' that received vertices and messages can persist into.|
-|worker.edge_properties_combiner_class|org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner|The combiner can combine several properties of the same edge into one properties at inputstep.|
-|worker.partitioner|org.apache.hugegraph.computer.core.graph.partition.HashPartitioner|The partitioner that decides which partition a vertex should be in, and which worker a partition should be in.|
-|worker.received_buffers_bytes_limit|104857600|The limit bytes of buffers of received data, the total size of all buffers can't excess this limit. If received buffers reach this limit, they will be merged into a file.|
-|worker.vertex_properties_combiner_class|org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner|The combiner can combine several properties of the same vertex into one properties at inputstep.|
-|worker.wait_finish_messages_timeout|86400000|The max timeout(in ms) message-handler wait for finish-message of all workers.|
-|worker.wait_sort_timeout|600000|The max timeout(in ms) message-handler wait for sort-thread to sort one batch of buffers.|
-|worker.write_buffer_capacity|52428800|The initial size of write buffer that used to store vertex or message.|
-|worker.write_buffer_threshold|52428800|The threshold of write buffer, exceeding it will trigger sorting, the write buffer is used to store vertex or message.|
+| config option                           | default value                                                                                   | description                                                                                                                                                                                                                                                                                                                                                    [...]
+|-----------------------------------------|-------------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- [...]
+| algorithm.message_class                 | org.apache.hugegraph.computer.core.config.Null                                                  | The class of message passed when compute vertex.                                                                                                                                                                                                                                                                                                               [...]
+| algorithm.params_class                  | org.apache.hugegraph.computer.core.config.Null                                                  | The class used to transfer algorithms' parameters before algorithm been run.                                                                                                                                                                                                                                                                                   [...]
+| algorithm.result_class                  | org.apache.hugegraph.computer.core.config.Null                                                  | The class of vertex's value, the instance is used to store computation result for the vertex.                                                                                                                                                                                                                                                                  [...]
+| allocator.max_vertices_per_thread       | 10000                                                                                           | Maximum number of vertices per thread processed in each memory allocator                                                                                                                                                                                                                                                                                       [...]
+| bsp.etcd_endpoints                      | http://localhost:2379                                                                           | The end points to access etcd.                                                                                                                                                                                                                                                                                                                                 [...]
+| bsp.log_interval                        | 30000                                                                                           | The log interval(in ms) to print the log while waiting bsp event.                                                                                                                                                                                                                                                                                              [...]
+| bsp.max_super_step                      | 10                                                                                              | The max super step of the algorithm.                                                                                                                                                                                                                                                                                                                           [...]
+| bsp.register_timeout                    | 300000                                                                                          | The max timeout to wait for master and works to register.                                                                                                                                                                                                                                                                                                      [...]
+| bsp.wait_master_timeout                 | 86400000                                                                                        | The max timeout(in ms) to wait for master bsp event.                                                                                                                                                                                                                                                                                                           [...]
+| bsp.wait_workers_timeout                | 86400000                                                                                        | The max timeout to wait for workers bsp event.                                                                                                                                                                                                                                                                                                                 [...]
+| hgkv.max_data_block_size                | 65536                                                                                           | The max byte size of hgkv-file data block.                                                                                                                                                                                                                                                                                                                     [...]
+| hgkv.max_file_size                      | 2147483648                                                                                      | The max number of bytes in each hgkv-file.                                                                                                                                                                                                                                                                                                                     [...]
+| hgkv.max_merge_files                    | 10                                                                                              | The max number of files to merge at one time.                                                                                                                                                                                                                                                                                                                  [...]
+| hgkv.temp_file_dir                      | /tmp/hgkv                                                                                       | This folder is used to store temporary files, temporary files will be generated during the file merging process.                                                                                                                                                                                                                                               [...]
+| hugegraph.name                          | hugegraph                                                                                       | The graph name to load data and write results back.                                                                                                                                                                                                                                                                                                            [...]
+| hugegraph.url                           | http://127.0.0.1:8080                                                                           | The hugegraph url to load data and write results back.                                                                                                                                                                                                                                                                                                         [...]
+| input.edge_direction                    | OUT                                                                                             | The data of the edge in which direction is loaded, when the value is BOTH, the edges in both OUT and IN direction will be loaded.                                                                                                                                                                                                                              [...]
+| input.edge_freq                         | MULTIPLE                                                                                        | The frequency of edges can exist between a pair of vertices, allowed values: [SINGLE, SINGLE_PER_LABEL, MULTIPLE]. SINGLE means that only one edge can exist between a pair of vertices, use sourceId + targetId to identify it; SINGLE_PER_LABEL means that each edge label can exist one edge between a pair of vertices, use sourceId + edgelabel + targetI [...]
+| input.filter_class                      | org.apache.hugegraph.computer.core.input.filter.DefaultInputFilter                              | The class to create input-filter object, input-filter is used to Filter vertex edges according to user needs.                                                                                                                                                                                                                                                  [...]
+| input.loader_schema_path                || The schema path of loader input, only takes effect when the input.source_type=loader is enabled |
+| input.loader_struct_path                || The struct path of loader input, only takes effect when the input.source_type=loader is enabled |
+| input.max_edges_in_one_vertex           | 200                                                                                             | The maximum number of adjacent edges allowed to be attached to a vertex, the adjacent edges will be stored and transferred together as a batch unit.                                                                                                                                                                                                           [...]
+| input.source_type                       | hugegraph-server                                                                                | The source type to load input data, allowed values: ['hugegraph-server', 'hugegraph-loader'], the 'hugegraph-loader' means use hugegraph-loader load data from HDFS or file, if use 'hugegraph-loader' load data then please config 'input.loader_struct_path' and 'input.loader_schema_path'.                                                                 [...]
+| input.split_fetch_timeout               | 300                                                                                             | The timeout in seconds to fetch input splits                                                                                                                                                                                                                                                                                                                   [...]
+| input.split_max_splits                  | 10000000                                                                                        | The maximum number of input splits                                                                                                                                                                                                                                                                                                                             [...]
+| input.split_page_size                   | 500                                                                                             | The page size for streamed load input split data                                                                                                                                                                                                                                                                                                               [...]
+| input.split_size                        | 1048576                                                                                         | The input split size in bytes                                                                                                                                                                                                                                                                                                                                  [...]
+| job.id                                  | local_0001                                                                                      | The job id on Yarn cluster or K8s cluster.                                                                                                                                                                                                                                                                                                                     [...]
+| job.partitions_count                    | 1                                                                                               | The partitions count for computing one graph algorithm job.                                                                                                                                                                                                                                                                                                    [...]
+| job.partitions_thread_nums              | 4                                                                                               | The number of threads for partition parallel compute.                                                                                                                                                                                                                                                                                                          [...]
+| job.workers_count                       | 1                                                                                               | The workers count for computing one graph algorithm job.                                                                                                                                                                                                                                                                                                       [...]
+| master.computation_class                | org.apache.hugegraph.computer.core.master.DefaultMasterComputation                              | Master-computation is computation that can determine whether to continue next superstep. It runs at the end of each superstep on master.                                                                                                                                                                                                                       [...]
+| output.batch_size                       | 500                                                                                             | The batch size of output                                                                                                                                                                                                                                                                                                                                       [...]
+| output.batch_threads                    | 1                                                                                               | The threads number used to batch output                                                                                                                                                                                                                                                                                                                        [...]
+| output.hdfs_core_site_path              || The hdfs core site path.                                                                        |
+| output.hdfs_delimiter                   | ,                                                                                               | The delimiter of hdfs output.                                                                                                                                                                                                                                                                                                                                  [...]
+| output.hdfs_kerberos_enable             | false                                                                                           | Is Kerberos authentication enabled for Hdfs.                                                                                                                                                                                                                                                                                                                   [...]
+| output.hdfs_kerberos_keytab             || The Hdfs's key tab file for kerberos authentication.                                            |
+| output.hdfs_kerberos_principal          || The Hdfs's principal for kerberos authentication.                                               |
+| output.hdfs_krb5_conf                   | /etc/krb5.conf                                                                                  | Kerberos configuration file.                                                                                                                                                                                                                                                                                                                                   [...]
+| output.hdfs_merge_partitions            | true                                                                                            | Whether merge output files of multiple partitions.                                                                                                                                                                                                                                                                                                             [...]
+| output.hdfs_path_prefix                 | /hugegraph-computer/results                                                                     | The directory of hdfs output result.                                                                                                                                                                                                                                                                                                                           [...]
+| output.hdfs_replication                 | 3                                                                                               | The replication number of hdfs.                                                                                                                                                                                                                                                                                                                                [...]
+| output.hdfs_site_path                   || The hdfs site path.                                                                             |
+| output.hdfs_url                         | hdfs://127.0.0.1:9000                                                                           | The hdfs url of output.                                                                                                                                                                                                                                                                                                                                        [...]
+| output.hdfs_user                        | hadoop                                                                                          | The hdfs user of output.                                                                                                                                                                                                                                                                                                                                       [...]
+| output.output_class                     | org.apache.hugegraph.computer.core.output.LogOutput                                             | The class to output the computation result of each vertex. Be called after iteration computation.                                                                                                                                                                                                                                                              [...]
+| output.result_name                      | value                                                                                           | The value is assigned dynamically by #name() of instance created by WORKER_COMPUTATION_CLASS.                                                                                                                                                                                                                                                                  [...]
+| output.result_write_type                | OLAP_COMMON                                                                                     | The result write-type to output to hugegraph, allowed values are: [OLAP_COMMON, OLAP_SECONDARY, OLAP_RANGE].                                                                                                                                                                                                                                                   [...]
+| output.retry_interval                   | 10                                                                                              | The retry interval when output failed                                                                                                                                                                                                                                                                                                                          [...]
+| output.retry_times                      | 3                                                                                               | The retry times when output failed                                                                                                                                                                                                                                                                                                                             [...]
+| output.single_threads                   | 1                                                                                               | The threads number used to single output                                                                                                                                                                                                                                                                                                                       [...]
+| output.thread_pool_shutdown_timeout     | 60                                                                                              | The timeout seconds of output threads pool shutdown                                                                                                                                                                                                                                                                                                            [...]
+| output.with_adjacent_edges              | false                                                                                           | Output the adjacent edges of the vertex or not                                                                                                                                                                                                                                                                                                                 [...]
+| output.with_edge_properties             | false                                                                                           | Output the properties of the edge or not                                                                                                                                                                                                                                                                                                                       [...]
+| output.with_vertex_properties           | false                                                                                           | Output the properties of the vertex or not                                                                                                                                                                                                                                                                                                                     [...]
+| sort.thread_nums                        | 4                                                                                               | The number of threads performing internal sorting.                                                                                                                                                                                                                                                                                                             [...]
+| transport.client_connect_timeout        | 3000                                                                                            | The timeout(in ms) of client connect to server.                                                                                                                                                                                                                                                                                                                [...]
+| transport.client_threads                | 4                                                                                               | The number of transport threads for client.                                                                                                                                                                                                                                                                                                                    [...]
+| transport.close_timeout                 | 10000                                                                                           | The timeout(in ms) of close server or close client.                                                                                                                                                                                                                                                                                                            [...]
+| transport.finish_session_timeout        | 0                                                                                               | The timeout(in ms) to finish session, 0 means using (transport.sync_request_timeout * transport.max_pending_requests).                                                                                                                                                                                                                                         [...]
+| transport.heartbeat_interval            | 20000                                                                                           | The minimum interval(in ms) between heartbeats on client side.                                                                                                                                                                                                                                                                                                 [...]
+| transport.io_mode                       | AUTO                                                                                            | The network IO Mode, either 'NIO', 'EPOLL', 'AUTO', the 'AUTO' means selecting the property mode automatically.                                                                                                                                                                                                                                                [...]
+| transport.max_pending_requests          | 8                                                                                               | The max number of client unreceived ack, it will trigger the sending unavailable if the number of unreceived ack >= max_pending_requests.                                                                                                                                                                                                                      [...]
+| transport.max_syn_backlog               | 511                                                                                             | The capacity of SYN queue on server side, 0 means using system default value.                                                                                                                                                                                                                                                                                  [...]
+| transport.max_timeout_heartbeat_count   | 120                                                                                             | The maximum times of timeout heartbeat on client side, if the number of timeouts waiting for heartbeat response continuously > max_heartbeat_timeouts the channel will be closed from client side.                                                                                                                                                             [...]
+| transport.min_ack_interval              | 200                                                                                             | The minimum interval(in ms) of server reply ack.                                                                                                                                                                                                                                                                                                               [...]
+| transport.min_pending_requests          | 6                                                                                               | The minimum number of client unreceived ack, it will trigger the sending available if the number of unreceived ack < min_pending_requests.                                                                                                                                                                                                                     [...]
+| transport.network_retries               | 3                                                                                               | The number of retry attempts for network communication,if network unstable.                                                                                                                                                                                                                                                                                    [...]
+| transport.provider_class                | org.apache.hugegraph.computer.core.network.netty.NettyTransportProvider                         | The transport provider, currently only supports Netty.                                                                                                                                                                                                                                                                                                         [...]
+| transport.receive_buffer_size           | 0                                                                                               | The size of socket receive-buffer in bytes, 0 means using system default value.                                                                                                                                                                                                                                                                                [...]
+| transport.recv_file_mode                | true                                                                                            | Whether enable receive buffer-file mode, it will receive buffer write file from socket by zero-copy if enable.                                                                                                                                                                                                                                                 [...]
+| transport.send_buffer_size              | 0                                                                                               | The size of socket send-buffer in bytes, 0 means using system default value.                                                                                                                                                                                                                                                                                   [...]
+| transport.server_host                   | 127.0.0.1                                                                                       | The server hostname or ip to listen on to transfer data.                                                                                                                                                                                                                                                                                                       [...]
+| transport.server_idle_timeout           | 360000                                                                                          | The max timeout(in ms) of server idle.                                                                                                                                                                                                                                                                                                                         [...]
+| transport.server_port                   | 0                                                                                               | The server port to listen on to transfer data. The system will assign a random port if it's set to 0.                                                                                                                                                                                                                                                          [...]
+| transport.server_threads                | 4                                                                                               | The number of transport threads for server.                                                                                                                                                                                                                                                                                                                    [...]
+| transport.sync_request_timeout          | 10000                                                                                           | The timeout(in ms) to wait response after sending sync-request.                                                                                                                                                                                                                                                                                                [...]
+| transport.tcp_keep_alive                | true                                                                                            | Whether enable TCP keep-alive.                                                                                                                                                                                                                                                                                                                                 [...]
+| transport.transport_epoll_lt            | false                                                                                           | Whether enable EPOLL level-trigger.                                                                                                                                                                                                                                                                                                                            [...]
+| transport.write_buffer_high_mark        | 67108864                                                                                        | The high water mark for write buffer in bytes, it will trigger the sending unavailable if the number of queued bytes > write_buffer_high_mark.                                                                                                                                                                                                                 [...]
+| transport.write_buffer_low_mark         | 33554432                                                                                        | The low water mark for write buffer in bytes, it will trigger the sending available if the number of queued bytes < write_buffer_low_mark.org.apache.hugegraph.config.OptionChecker$$Lambda$97/0x00000008001c8440@776a6d9b                                                                                                                                     [...]
+| transport.write_socket_timeout          | 3000                                                                                            | The timeout(in ms) to write data to socket buffer.                                                                                                                                                                                                                                                                                                             [...]
+| valuefile.max_segment_size              | 1073741824                                                                                      | The max number of bytes in each segment of value-file.                                                                                                                                                                                                                                                                                                         [...]
+| worker.combiner_class                   | org.apache.hugegraph.computer.core.config.Null                                                  | Combiner can combine messages into one value for a vertex, for example page-rank algorithm can combine messages of a vertex to a sum value.                                                                                                                                                                                                                    [...]
+| worker.computation_class                | org.apache.hugegraph.computer.core.config.Null                                                  | The class to create worker-computation object, worker-computation is used to compute each vertex in each superstep.                                                                                                                                                                                                                                            [...]
+| worker.data_dirs                        | [jobs]                                                                                          | The directories separated by ',' that received vertices and messages can persist into.                                                                                                                                                                                                                                                                         [...]
+| worker.edge_properties_combiner_class   | org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner                         | The combiner can combine several properties of the same edge into one properties at inputstep.                                                                                                                                                                                                                                                                 [...]
+| worker.partitioner                      | org.apache.hugegraph.computer.core.graph.partition.HashPartitioner                              | The partitioner that decides which partition a vertex should be in, and which worker a partition should be in.                                                                                                                                                                                                                                                 [...]
+| worker.received_buffers_bytes_limit     | 104857600                                                                                       | The limit bytes of buffers of received data, the total size of all buffers can't excess this limit. If received buffers reach this limit, they will be merged into a file.                                                                                                                                                                                     [...]
+| worker.vertex_properties_combiner_class | org.apache.hugegraph.computer.core.combiner.OverwritePropertiesCombiner                         | The combiner can combine several properties of the same vertex into one properties at inputstep.                                                                                                                                                                                                                                                               [...]
+| worker.wait_finish_messages_timeout     | 86400000                                                                                        | The max timeout(in ms) message-handler wait for finish-message of all workers.                                                                                                                                                                                                                                                                                 [...]
+| worker.wait_sort_timeout                | 600000                                                                                          | The max timeout(in ms) message-handler wait for sort-thread to sort one batch of buffers.                                                                                                                                                                                                                                                                      [...]
+| worker.write_buffer_capacity            | 52428800                                                                                        | The initial size of write buffer that used to store vertex or message.                                                                                                                                                                                                                                                                                         [...]
+| worker.write_buffer_threshold           | 52428800                                                                                        | The threshold of write buffer, exceeding it will trigger sorting, the write buffer is used to store vertex or message.                                                                                                                                                                                                                                         [...]
 
 ### K8s Operator Config Options
 
-> NOTE: Option needs to be converted through environment variable settings, e.g k8s.internal_etcd_url => INTERNAL_ETCD_URL
+> NOTE: Option needs to be converted through environment variable settings, e.g. k8s.internal_etcd_url => INTERNAL_ETCD_URL
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|k8s.auto_destroy_pod|true|Whether to automatically destroy all pods when the job is completed or failed.|
-|k8s.close_reconciler_timeout|120|The max timeout(in ms) to close reconciler.|
-|k8s.internal_etcd_url|http://127.0.0.1:2379|The internal etcd url for operator system.|
-|k8s.max_reconcile_retry|3|The max retry times of reconcile.|
-|k8s.probe_backlog|50|The maximum backlog for serving health probes.|
-|k8s.probe_port|9892|The value is the port that the controller bind to for serving health probes.|
-|k8s.ready_check_internal|1000|The time interval(ms) of check ready.|
-|k8s.ready_timeout|30000|The max timeout(in ms) of check ready.|
-|k8s.reconciler_count|10|The max number of reconciler thread.|
-|k8s.resync_period|600000|The minimum frequency at which watched resources are reconciled.|
-|k8s.timezone|Asia/Shanghai|The timezone of computer job and operator.|
-|k8s.watch_namespace|hugegraph-computer-system|The value is watch custom resources in the namespace, ignore other namespaces, the '*' means is all namespaces will be watched.|
+| config option                | default value             | description                                                                                                                     |
+|------------------------------|---------------------------|---------------------------------------------------------------------------------------------------------------------------------|
+| k8s.auto_destroy_pod         | true                      | Whether to automatically destroy all pods when the job is completed or failed.                                                  |
+| k8s.close_reconciler_timeout | 120                       | The max timeout(in ms) to close reconciler.                                                                                     |
+| k8s.internal_etcd_url        | http://127.0.0.1:2379     | The internal etcd url for operator system.                                                                                      |
+| k8s.max_reconcile_retry      | 3                         | The max retry times of reconcile.                                                                                               |
+| k8s.probe_backlog            | 50                        | The maximum backlog for serving health probes.                                                                                  |
+| k8s.probe_port               | 9892                      | The value is the port that the controller bind to for serving health probes.                                                    |
+| k8s.ready_check_internal     | 1000                      | The time interval(ms) of check ready.                                                                                           |
+| k8s.ready_timeout            | 30000                     | The max timeout(in ms) of check ready.                                                                                          |
+| k8s.reconciler_count         | 10                        | The max number of reconciler thread.                                                                                            |
+| k8s.resync_period            | 600000                    | The minimum frequency at which watched resources are reconciled.                                                                |
+| k8s.timezone                 | Asia/Shanghai             | The timezone of computer job and operator.                                                                                      |
+| k8s.watch_namespace          | hugegraph-computer-system | The value is watch custom resources in the namespace, ignore other namespaces, the '*' means is all namespaces will be watched. |
 
 ### HugeGraph-Computer CRD
 
 > CRD: https://github.com/apache/hugegraph-computer/blob/master/computer-k8s-operator/manifest/hugegraph-computer-crd.v1.yaml
 
-| spec            | default value           | description                                                  | required |
-| --------------- | ----------------------- | ------------------------------------------------------------ | -------- |
-| algorithmName   |                         | The name of algorithm.                                       | true     |
-| jobId           |                         | The job id.                                                  | true     |
-| image           |                         | The image of algorithm.                                      | true     |
-| computerConf    |                         | The map of computer config options.                          | true     |
-| workerInstances |                         | The number of worker instances, it will instead the 'job.workers_count' option. | true     |
-| pullPolicy      | Always                  | The pull-policy of image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy | false    |
-| pullSecrets     |                         | The pull-secrets of Image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod | false    |
-| masterCpu       |                         | The cpu limit of master, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu) | false    |
-| workerCpu       |                         | The cpu limit of worker, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu) | false    |
+| spec            | default value           | description                                                                                                                                                                                                                                                                                      | required |
+|-----------------|-------------------------|--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------|
+| algorithmName   |                         | The name of algorithm.                                                                                                                                                                                                                                                                           | true     |
+| jobId           |                         | The job id.                                                                                                                                                                                                                                                                                      | true     |
+| image           |                         | The image of algorithm.                                                                                                                                                                                                                                                                          | true     |
+| computerConf    |                         | The map of computer config options.                                                                                                                                                                                                                                                              | true     |
+| workerInstances |                         | The number of worker instances, it will instead the 'job.workers_count' option.                                                                                                                                                                                                                  | true     |
+| pullPolicy      | Always                  | The pull-policy of image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#image-pull-policy                                                                                                                                                                       | false    |
+| pullSecrets     |                         | The pull-secrets of Image, detail please refer to: https://kubernetes.io/docs/concepts/containers/images/#specifying-imagepullsecrets-on-a-pod                                                                                                                                                   | false    |
+| masterCpu       |                         | The cpu limit of master, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu)               | false    |
+| workerCpu       |                         | The cpu limit of worker, the unit can be 'm' or without unit detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-cpu)               | false    |
 | masterMemory    |                         | The memory limit of master, the unit can be one of Ei、Pi、Ti、Gi、Mi、Ki detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory) | false    |
 | workerMemory    |                         | The memory limit of worker, the unit can be one of Ei、Pi、Ti、Gi、Mi、Ki detail please refer to:[https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory](https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/#meaning-of-memory) | false    |
-| log4jXml        |                         | The content of log4j.xml for computer job.                   | false    |
-| jarFile         |                         | The jar path of computer algorithm.                          | false    |
-| remoteJarUri    |                         | The remote jar uri of computer algorithm, it will overlay algorithm image. | false    |
-| jvmOptions      |                         | The java startup parameters of computer job.                 | false    |
-| envVars         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-interdependent-environment-variables/ | false    |
-| envFrom         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-environment-variable-container/ | false    |
-| masterCommand   | bin/start-computer.sh   | The run command of master, equivalent to 'Entrypoint' field of Docker. | false    |
-| masterArgs      | ["-r master", "-d k8s"] | The run args of master, equivalent to 'Cmd' field of Docker. | false    |
-| workerCommand   | bin/start-computer.sh   | The run command of worker, equivalent to 'Entrypoint' field of Docker. | false    |
-| workerArgs      | ["-r worker", "-d k8s"] | The run args of worker, equivalent to 'Cmd' field of Docker. | false    |
-| volumes         |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/ | false    |
-| volumeMounts    |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/ | false    |
-| secretPaths     |                         | The map of k8s-secret name and mount path.                   | false    |
-| configMapPaths  |                         | The map of k8s-configmap name and mount path.                | false    |
-| podTemplateSpec |                         | Please refer to: https://kubernetes.io/docs/reference/kubernetes-api/workload-resources/pod-template-v1/#PodTemplateSpec | false    |
-| securityContext |                         | Please refer to: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/ | false    |
+| log4jXml        |                         | The content of log4j.xml for computer job.                                                                                                                                                                                                                                                       | false    |
+| jarFile         |                         | The jar path of computer algorithm.                                                                                                                                                                                                                                                              | false    |
+| remoteJarUri    |                         | The remote jar uri of computer algorithm, it will overlay algorithm image.                                                                                                                                                                                                                       | false    |
+| jvmOptions      |                         | The java startup parameters of computer job.                                                                                                                                                                                                                                                     | false    |
+| envVars         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-interdependent-environment-variables/                                                                                                                                                                           | false    |
+| envFrom         |                         | please refer to: https://kubernetes.io/docs/tasks/inject-data-application/define-environment-variable-container/                                                                                                                                                                                 | false    |
+| masterCommand   | bin/start-computer.sh   | The run command of master, equivalent to 'Entrypoint' field of Docker.                                                                                                                                                                                                                           | false    |
+| masterArgs      | ["-r master", "-d k8s"] | The run args of master, equivalent to 'Cmd' field of Docker.                                                                                                                                                                                                                                     | false    |
+| workerCommand   | bin/start-computer.sh   | The run command of worker, equivalent to 'Entrypoint' field of Docker.                                                                                                                                                                                                                           | false    |
+| workerArgs      | ["-r worker", "-d k8s"] | The run args of worker, equivalent to 'Cmd' field of Docker.                                                                                                                                                                                                                                     | false    |
+| volumes         |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/                                                                                                                                                                                                                            | false    |
+| volumeMounts    |                         | Please refer to: https://kubernetes.io/docs/concepts/storage/volumes/                                                                                                                                                                                                                            | false    |
+| secretPaths     |                         | The map of k8s-secret name and mount path.                                                                                                                                                                                                                                                       | false    |
+| configMapPaths  |                         | The map of k8s-configmap name and mount path.                                                                                                                                                                                                                                                    | false    |
+| podTemplateSpec |                         | Please refer to: https://kubernetes.io/docs/reference/kubernetes-api/workload-resources/pod-template-v1/#PodTemplateSpec                                                                                                                                                                         | false    |
+| securityContext |                         | Please refer to: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/                                                                                                                                                                                                      | false    |
 
 ### KubeDriver Config Options
 
-| config option |  default value    | description |
-| ---- | ---- | ---- |
-|k8s.build_image_bash_path||The path of command used to build image.|
-|k8s.enable_internal_algorithm|true|Whether enable internal algorithm.|
-|k8s.framework_image_url|hugegraph/hugegraph-computer:latest|The image url of computer framework.|
-|k8s.image_repository_password||The password for login image repository.|
-|k8s.image_repository_registry||The address for login image repository.|
-|k8s.image_repository_url|hugegraph/hugegraph-computer|The url of image repository.|
-|k8s.image_repository_username||The username for login image repository.|
-|k8s.internal_algorithm|[pageRank]|The name list of all internal algorithm.|
-|k8s.internal_algorithm_image_url|hugegraph/hugegraph-computer:latest|The image url of internal algorithm.|
-|k8s.jar_file_dir|/cache/jars/|The directory where the algorithm jar to upload location.|
-|k8s.kube_config|~/.kube/config|The path of k8s config file.|
-|k8s.log4j_xml_path||The log4j.xml path for computer job.|
-|k8s.namespace|hugegraph-computer-system|The namespace of hugegraph-computer system.|
-|k8s.pull_secret_names|[]|The names of pull-secret for pulling image.|
+| config option                    | default value                            | description                                               |
+|----------------------------------|------------------------------------------|-----------------------------------------------------------|
+| k8s.build_image_bash_path        || The path of command used to build image. |
+| k8s.enable_internal_algorithm    | true                                     | Whether enable internal algorithm.                        |
+| k8s.framework_image_url          | hugegraph/hugegraph-computer:latest      | The image url of computer framework.                      |
+| k8s.image_repository_password    || The password for login image repository. |
+| k8s.image_repository_registry    || The address for login image repository.  |
+| k8s.image_repository_url         | hugegraph/hugegraph-computer             | The url of image repository.                              |
+| k8s.image_repository_username    || The username for login image repository. |
+| k8s.internal_algorithm           | [pageRank]                               | The name list of all internal algorithm.                  |
+| k8s.internal_algorithm_image_url | hugegraph/hugegraph-computer:latest      | The image url of internal algorithm.                      |
+| k8s.jar_file_dir                 | /cache/jars/                             | The directory where the algorithm jar to upload location. |
+| k8s.kube_config                  | ~/.kube/config                           | The path of k8s config file.                              |
+| k8s.log4j_xml_path               || The log4j.xml path for computer job.     |
+| k8s.namespace                    | hugegraph-computer-system                | The namespace of hugegraph-computer system.               |
+| k8s.pull_secret_names            | []                                       | The names of pull-secret for pulling image.               |
diff --git a/content/en/docs/config/config-option.md b/content/en/docs/config/config-option.md
index 5e09df89..ed2253b4 100644
--- a/content/en/docs/config/config-option.md
+++ b/content/en/docs/config/config-option.md
@@ -8,7 +8,7 @@ weight: 2
 
 Corresponding configuration file `gremlin-server.yaml`
 
-| config option           | default value                                                                                                | description                                                                      |
+| config option           | default value                                                                                                | description                                                                     |
 |-------------------------|--------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------|
 | host                    | 127.0.0.1                                                                                                    | The host or ip of Gremlin Server.                                               |
 | port                    | 8182                                                                                                         | The listening port of Gremlin Server.                                           |
@@ -21,7 +21,7 @@ Corresponding configuration file `gremlin-server.yaml`
 
 Corresponding configuration file `rest-server.properties`
 
-| config option                      | default value                                    | description                                                                                                                                                                                                    |
+| config option                      | default value                                    | description                                                                                                                                                                                                   |
 |------------------------------------|--------------------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 | graphs                             | [hugegraph:conf/hugegraph.properties]            | The map of graphs' name and config file.                                                                                                                                                                      |
 | server.id                          | server-1                                         | The id of rest server, used for license verification.                                                                                                                                                         |
@@ -41,7 +41,7 @@ Corresponding configuration file `rest-server.properties`
 | batch.max_vertices_per_batch       | 500                                              | The maximum number of vertices submitted per batch.                                                                                                                                                           |
 | batch.max_write_ratio              | 50                                               | The maximum thread ratio for batch writing, only take effect if the batch.max_write_threads is 0.                                                                                                             |
 | batch.max_write_threads            | 0                                                | The maximum threads for batch writing, if the value is 0, the actual value will be set to batch.max_write_ratio * restserver.max_worker_threads.                                                              |
-| auth.authenticator                 |                                                  | The class path of authenticator implementation. e.g., com.baidu.hugegraph.auth.StandardAuthenticator, or com.baidu.hugegraph.auth.ConfigAuthenticator.                                                          |
+| auth.authenticator                 |                                                  | The class path of authenticator implementation. e.g., com.baidu.hugegraph.auth.StandardAuthenticator, or com.baidu.hugegraph.auth.ConfigAuthenticator.                                                        |
 | auth.admin_token                   | 162f7848-0b6d-4faf-b557-3a0797869c55             | Token for administrator operations, only for com.baidu.hugegraph.auth.ConfigAuthenticator.                                                                                                                    |
 | auth.graph_store                   | hugegraph                                        | The name of graph used to store authentication information, like users, only for com.baidu.hugegraph.auth.StandardAuthenticator.                                                                              |
 | auth.user_tokens                   | [hugegraph:9fd95c9c-711b-415b-b85f-d4df46ba5c31] | The map of user tokens with name and password, only for com.baidu.hugegraph.auth.ConfigAuthenticator.                                                                                                         |
@@ -131,7 +131,7 @@ Basic Config Options and Backend Config Options correspond to configuration file
 
 ### RPC server Config Options
 
-| config option               | default value                           | description                                                                                                                                                                                                    |
+| config option               | default value                           | description                                                                                                                                                                                                   |
 |-----------------------------|-----------------------------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
 | rpc.client_connect_timeout  | 20                                      | The timeout(in seconds) of rpc client connect to rpc server.                                                                                                                                                  |
 | rpc.client_load_balancer    | consistentHash                          | The rpc client uses a load-balancing algorithm to access multiple rpc servers in one cluster. Default value is 'consistentHash', means forwarding by request parameters.                                      |
@@ -149,7 +149,7 @@ Basic Config Options and Backend Config Options correspond to configuration file
 
 ### Cassandra Backend Config Options
 
-| config option                  | default value  | description                                                                                                                                     |
+| config option                  | default value  | description                                                                                                                                    |
 |--------------------------------|----------------|------------------------------------------------------------------------------------------------------------------------------------------------|
 | backend                        |                | Must be set to `cassandra`.                                                                                                                    |
 | serializer                     |                | Must be set to `cassandra`.                                                                                                                    |
@@ -167,7 +167,7 @@ Basic Config Options and Backend Config Options correspond to configuration file
 
 ### ScyllaDB Backend Config Options
 
-| config option | default value | description                 |
+| config option | default value | description                |
 |---------------|---------------|----------------------------|
 | backend       |               | Must be set to `scylladb`. |
 | serializer    |               | Must be set to `scylladb`. |
@@ -176,7 +176,7 @@ Other options are consistent with the Cassandra backend.
 
 ### RocksDB Backend Config Options
 
-| config option                                   | default value                                                                                                                        | description                                                                                                                                                                                                                                                                                                       [...]
+| config option                                   | default value                                                                                                                        | description                                                                                                                                                                                                                                                                                                       [...]
 |-------------------------------------------------|--------------------------------------------------------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ [...]
 | backend                                         |                                                                                                                                      | Must be set to `rocksdb`.                                                                                                                                                                                                                                                                                         [...]
 | serializer                                      |                                                                                                                                      | Must be set to `binary`.                                                                                                                                                                                                                                                                                          [...]
@@ -230,7 +230,7 @@ Other options are consistent with the Cassandra backend.
 
 ### HBase Backend Config Options
 
-| config option             | default value                  | description                                                               |
+| config option             | default value                  | description                                                              |
 |---------------------------|--------------------------------|--------------------------------------------------------------------------|
 | backend                   |                                | Must be set to `hbase`.                                                  |
 | serializer                |                                | Must be set to `hbase`.                                                  |
@@ -251,7 +251,7 @@ Other options are consistent with the Cassandra backend.
 
 ### MySQL & PostgreSQL Backend Config Options
 
-| config option                    | default value               | description                                                                          |
+| config option                    | default value               | description                                                                         |
 |----------------------------------|-----------------------------|-------------------------------------------------------------------------------------|
 | backend                          |                             | Must be set to `mysql`.                                                             |
 | serializer                       |                             | Must be set to `mysql`.                                                             |
@@ -267,7 +267,7 @@ Other options are consistent with the Cassandra backend.
 
 ### PostgreSQL Backend Config Options
 
-| config option | default value | description                   |
+| config option | default value | description                  |
 |---------------|---------------|------------------------------|
 | backend       |               | Must be set to `postgresql`. |
 | serializer    |               | Must be set to `postgresql`. |
diff --git a/content/en/docs/contribution-guidelines/validate-release.md b/content/en/docs/contribution-guidelines/validate-release.md
index 535822ed..d69b9de6 100644
--- a/content/en/docs/contribution-guidelines/validate-release.md
+++ b/content/en/docs/contribution-guidelines/validate-release.md
@@ -82,7 +82,7 @@ First confirm the overall integrity/consistency, and then confirm the specific c
 
 #### 4. Check the archive contents
 
-Here it is divided into two aspects: source code package + binary package, The source code package is more strict, it can be said that the core part (Because it is longer,For a complete list refer to the official [Wiki](https://cwiki.apache.org/confluence/display/INCUBATOR/Incubator+Release+Checklist))
+Here it is divided into two aspects: source code package + binary package, The source code package is stricter, it can be said that the core part (Because it is longer,For a complete list refer to the official [Wiki](https://cwiki.apache.org/confluence/display/INCUBATOR/Incubator+Release+Checklist))
 
 First of all, we need to download the package from the apache official `release-candidate` URL to the local (URL: [click to jump](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/))
 
@@ -100,8 +100,11 @@ After decompressing `xxx-hugegraph-source.tar.gz`, Do the following checks:
 8. Finally, make sure the source code works/compiles correctly (then look at tests and specs)
 
 ```bash
-# At the same time, you can also check whether the code style conforms to the specification, and if it does not conform, you can put down an adjustment
-mvn clean test -Dcheckstyle.skip=false
+# prefer to use/switch to java 11 for the following operations (compiling/running)
+# java --version
+
+# try to test in the Unix env to check if it works well
+mvn clean package -Dmaven.test.skip=true -Dcheckstyle.skip=true
 ```
 
 ##### B. binary package
@@ -114,7 +117,7 @@ After decompressing `xxx-hugegraph.tar.gz`, perform the following checks:
 
 **Note:** If a third-party dependency is introduced in the binary package, you need to update the LICENSE and add the third-party dependent LICENSE; if the third-party dependent LICENSE is Apache 2.0, and the corresponding project contains NOTICE, you also need to update Our NOTICE file
 
-#### 5. Check the official website and github and other pages
+#### 5. Check the official website and GitHub and other pages
 
 1. Make sure that the official website at least meets [apache website check](https://whimsy.apache.org/pods/project/hugegraph), and no circular links etc.
 2. Update **download link** and version update instructions
diff --git a/content/en/docs/download/download.md b/content/en/docs/download/download.md
index e535378c..026df86d 100644
--- a/content/en/docs/download/download.md
+++ b/content/en/docs/download/download.md
@@ -4,23 +4,23 @@ linkTitle: "Download"
 weight: 2
 ---
 
-> TODO: update 1.0.0 download link
+> TODO: update 1.0.0 download link with the CN doc
 
 ### Latest version
 
 The latest HugeGraph: **1.0.0**, released on _2022-12-31_.
 
-| components       | description               | download                                                                                                                     |
-| ---------------- | ------------------------- | ---------------------------------------------------------------------------------------------------------------------------- |
-| HugeGraph-Server | HugeGraph 的主程序        | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)           |
+| components       | description     | download                                                                                                                     |
+|------------------|-----------------|------------------------------------------------------------------------------------------------------------------------------|
+| HugeGraph-Server | HugeGraph 的主程序  | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)           |
 | HugeGraph-Hubble | 基于 Web 的可视化图形界面 | 1.0.0                                                                                                                        |
-| HugeGraph-Loader | 数据导入工具              | 1.0.0                                                                                                                        |
-| HugeGraph-Tools  | 命令行工具集              | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-toolchain-incubating-1.0.0.tar.gz) |
+| HugeGraph-Loader | 数据导入工具          | 1.0.0                                                                                                                        |
+| HugeGraph-Tools  | 命令行工具集          | [1.0.0](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-toolchain-incubating-1.0.0.tar.gz) |
 
 ### Binary Versions mapping
 
 | Version | Release Date | server                                                                                                                                                                                                                                                                                                                                                                     | toolchain                                                                                            [...]
-| ------- | ------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------- [...]
+|---------|--------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------------------- [...]
 | 1.0.0   | 2022-12-31   | [[Binary](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz)] [[Sign](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz.asc)] [[SHA512](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0.tar.gz.sha512)] | [[Binary](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-toolchai [...]
 
 ---
@@ -28,7 +28,7 @@ The latest HugeGraph: **1.0.0**, released on _2022-12-31_.
 ### Source Versions mapping
 
 | Version | Release Date | server                                                                                                                                                                                                                                                                                                                                                                                 | toolchain                                                                                [...]
-| ------- | ------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------- [...]
+|---------|--------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------- [...]
 | 1.0.0   | 2022-12-31   | [[Source](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz)] [[Sign](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz.asc)] [[SHA512](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegraph-incubating-1.0.0-src.tar.gz.sha512)] | [[Source](https://dist.apache.org/repos/dist/dev/incubator/hugegraph/1.0.0/apache-hugegr [...]
 
 ---
@@ -36,7 +36,7 @@ The latest HugeGraph: **1.0.0**, released on _2022-12-31_.
 ### outdated Versions mapping
 
 | server                                                                                             | client                                                                                 | loader                                                                                                           | hubble                                                                                                             | common                                                               [...]
-| -------------------------------------------------------------------------------------------------- | -------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------------------------------- | ------------------------------------------------------------------------------------------------------------------ | -------------------------------------------------------------------- [...]
+|----------------------------------------------------------------------------------------------------|----------------------------------------------------------------------------------------|------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------------------------------------------------------|--------------------------------------------------------------------- [...]
 | [0.12.0](https://github.com/hugegraph/hugegraph/releases/download/v0.12.0/hugegraph-0.12.0.tar.gz) | [2.0.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/2.0.1) | [0.12.0](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.12.0/hugegraph-loader-0.12.0.tar.gz) | [1.6.0](https://github.com/hugegraph/hugegraph-hubble/releases/download/v1.6.0/hugegraph-hubble-1.6.0.tar.gz)      | [2.0.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugeg [...]
 | [0.11.2](https://github.com/hugegraph/hugegraph/releases/download/v0.11.2/hugegraph-0.11.2.tar.gz) | [1.9.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/1.9.1) | [0.11.1](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.11.1/hugegraph-loader-0.11.1.tar.gz) | [1.5.0](https://github.com/hugegraph/hugegraph-hubble/releases/download/v1.5.0/hugegraph-hubble-1.5.0.tar.gz)      | [1.8.1](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugeg [...]
 | [0.10.4](https://github.com/hugegraph/hugegraph/releases/download/v0.10.4/hugegraph-0.10.4.tar.gz) | [1.8.0](https://mvnrepository.com/artifact/com.baidu.hugegraph/hugegraph-client/1.8.0) | [0.10.1](https://github.com/hugegraph/hugegraph-loader/releases/download/v0.10.1/hugegraph-loader-0.10.1.tar.gz) | [0.10.0](https://github.com/hugegraph/hugegraph-studio/releases/download/v0.10.0/hugegraph-studio-0.10.0.tar.gz)   | [1.6.16](https://mvnrepository.com/artifact/com.baidu.hugegraph/huge [...]
diff --git a/content/en/docs/guides/faq.md b/content/en/docs/guides/faq.md
index 9503ea8e..71a5a5d5 100644
--- a/content/en/docs/guides/faq.md
+++ b/content/en/docs/guides/faq.md
@@ -10,7 +10,7 @@ weight: 5
 
 - 启动服务时提示:`xxx (core dumped) xxx`
 
-  请检查JDK版本是否为1.8  
+  请检查 JDK 版本是否为 Java 11, 至少要求是 Java 8  
 
 - 启动服务成功了,但是操作图时有类似于"无法连接到后端或连接未打开"的提示
 
diff --git a/content/en/docs/language/hugegraph-example.md b/content/en/docs/language/hugegraph-example.md
index 050b8cad..7ad5bdab 100644
--- a/content/en/docs/language/hugegraph-example.md
+++ b/content/en/docs/language/hugegraph-example.md
@@ -26,9 +26,9 @@ HugeGraph相对于TitanDB而言,其主要特点如下:
 
 本示例通过Property Graph Model图数据模型来描述希腊神话中各人物角色的关系(也被成为人物关系图谱),具体关系详见下图。
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/graph-of-gods.png" alt="image">
-</center>
+</div>
 
 
 其中,圆形节点代表实体(Vertex),箭头代表关系(Edge),方框的内容为属性。
diff --git a/content/en/docs/language/hugegraph-gremlin.md b/content/en/docs/language/hugegraph-gremlin.md
index 47b9ac68..9eda438f 100644
--- a/content/en/docs/language/hugegraph-gremlin.md
+++ b/content/en/docs/language/hugegraph-gremlin.md
@@ -23,7 +23,7 @@ HugeGraph实现了TinkerPop框架,但是并没有实现TinkerPop所有的特
 | Computer             | Determines if the {@code Graph} implementation supports {@link GraphComputer} based processing                                                                                                                                                                                                                                                                                                                                                                | false   |
 | Transactions         | Determines if the {@code Graph} implementations supports transactions.                                                                                                                                                                                                                                                                                                                                                                                        | true    |
 | Persistence          | Determines if the {@code Graph} implementation supports persisting it's contents natively to disk.This feature does not refer to every graph's ability to write to disk via the Gremlin IO packages(.e.g. GraphML), unless the graph natively persists to disk via those options somehow. For example,TinkerGraph does not support this feature as it is a pure in-sideEffects graph.                                                                         | true    |
-| ThreadedTransactions | Determines if the {@code Graph} implementation supports threaded transactions which allow a transaction be executed across multiple threads via {@link Transaction#createThreadedTx()}.                                                                                                                                                                                                                                                                     | false   |
+| ThreadedTransactions | Determines if the {@code Graph} implementation supports threaded transactions which allow a transaction be executed across multiple threads via {@link Transaction#createThreadedTx()}.                                                                                                                                                                                                                                                                       | false   |
 | ConcurrentAccess     | Determines if the {@code Graph} implementation supports more than one connection to the same instance at the same time. For example, Neo4j embedded does not support this feature because concurrent access to the same database files by multiple instances is not possible. However, Neo4j HA could support this feature as each new {@code Graph} instance coordinates with the Neo4j cluster allowing multiple instances to operate on the same database. | false   |
 
 ### Vertex Features
@@ -61,25 +61,25 @@ HugeGraph实现了TinkerPop框架,但是并没有实现TinkerPop所有的特
 
 ### Data Type Features
 
-| Name               | Description                                                                                                                                                                                                                                                         | Support |
-|--------------------|---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------|
-| BooleanValues      |                                                                                                                                                                                                                                                                     | true    |
-| ByteValues         |                                                                                                                                                                                                                                                                     | true    |
-| DoubleValues       |                                                                                                                                                                                                                                                                     | true    |
-| FloatValues        |                                                                                                                                                                                                                                                                     | true    |
-| IntegerValues      |                                                                                                                                                                                                                                                                     | true    |
-| LongValues         |                                                                                                                                                                                                                                                                     | true    |
+| Name               | Description                                                                                                                                                                                                                                                          | Support |
+|--------------------|----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|---------|
+| BooleanValues      |                                                                                                                                                                                                                                                                      | true    |
+| ByteValues         |                                                                                                                                                                                                                                                                      | true    |
+| DoubleValues       |                                                                                                                                                                                                                                                                      | true    |
+| FloatValues        |                                                                                                                                                                                                                                                                      | true    |
+| IntegerValues      |                                                                                                                                                                                                                                                                      | true    |
+| LongValues         |                                                                                                                                                                                                                                                                      | true    |
 | MapValues          | Supports setting of a {@code Map} value. The assumption is that the {@code Map} can contain arbitrary serializable values that may or may not be defined as a feature itself                                                                                         | false   |
 | MixedListValues    | Supports setting of a {@code List} value. The assumption is that the {@code List} can contain arbitrary serializable values that may or may not be defined as a feature itself. As this{@code List} is "mixed" it does not need to contain objects of the same type. | false   |
-| BooleanArrayValues |                                                                                                                                                                                                                                                                     | false   |
-| ByteArrayValues    |                                                                                                                                                                                                                                                                     | true    |
-| DoubleArrayValues  |                                                                                                                                                                                                                                                                     | false   |
-| FloatArrayValues   |                                                                                                                                                                                                                                                                     | false   |
-| IntegerArrayValues |                                                                                                                                                                                                                                                                     | false   |
-| LongArrayValues    |                                                                                                                                                                                                                                                                     | false   |
-| SerializableValues |                                                                                                                                                                                                                                                                     | false   |
-| StringArrayValues  |                                                                                                                                                                                                                                                                     | false   |
-| StringValues       |                                                                                                                                                                                                                                                                     | true    |
+| BooleanArrayValues |                                                                                                                                                                                                                                                                      | false   |
+| ByteArrayValues    |                                                                                                                                                                                                                                                                      | true    |
+| DoubleArrayValues  |                                                                                                                                                                                                                                                                      | false   |
+| FloatArrayValues   |                                                                                                                                                                                                                                                                      | false   |
+| IntegerArrayValues |                                                                                                                                                                                                                                                                      | false   |
+| LongArrayValues    |                                                                                                                                                                                                                                                                      | false   |
+| SerializableValues |                                                                                                                                                                                                                                                                      | false   |
+| StringArrayValues  |                                                                                                                                                                                                                                                                      | false   |
+| StringValues       |                                                                                                                                                                                                                                                                      | true    |
 | UniformListValues  | Supports setting of a {@code List} value. The assumption is that the {@code List} can contain arbitrary serializable values that may or may not be defined as a feature itself. As this{@code List} is "uniform" it must contain objects of the same type.           | false   |
 
 ### Gremlin的步骤
diff --git a/content/en/docs/quickstart/hugegraph-client.md b/content/en/docs/quickstart/hugegraph-client.md
index 268843f9..05ddb620 100644
--- a/content/en/docs/quickstart/hugegraph-client.md
+++ b/content/en/docs/quickstart/hugegraph-client.md
@@ -10,8 +10,8 @@ HugeGraph-Client sends HTTP request to HugeGraph-Server to obtain and parse the
 
 ### 2 What You Need
 
-- JDK 1.8
-- Maven 3.3.9+
+- Java 11 (also support Java 8)
+- Maven 3.5+
 
 ### 3 How To Use
 
diff --git a/content/en/docs/quickstart/hugegraph-computer.md b/content/en/docs/quickstart/hugegraph-computer.md
index 81f885c1..78399216 100644
--- a/content/en/docs/quickstart/hugegraph-computer.md
+++ b/content/en/docs/quickstart/hugegraph-computer.md
@@ -22,7 +22,7 @@ The [`HugeGraph-Computer`](https://github.com/apache/incubator-hugegraph-compute
 
 ### 2.1 Run PageRank algorithm locally
 
-> To run algorithm with HugeGraph-Computer, you need to install 64-bit JRE/JDK 11 or later versions.
+> To run algorithm with HugeGraph-Computer, you need to install 64-bit Java 11 or later versions.
 >
 > You also need to deploy HugeGraph-Server and [Etcd](https://etcd.io/docs/v3.5/quickstart/).
 
diff --git a/content/en/docs/quickstart/hugegraph-loader.md b/content/en/docs/quickstart/hugegraph-loader.md
index 738ad483..9c8272fd 100644
--- a/content/en/docs/quickstart/hugegraph-loader.md
+++ b/content/en/docs/quickstart/hugegraph-loader.md
@@ -84,10 +84,10 @@ This step is the modeling process. Users need to have a clear idea of ​​thei
 For example, if you want to create a graph with two types of vertices and two types of edges, the vertices are "people" and "software", the edges are "people know people" and "people create software", and these vertices and edges have some attributes, For example, the vertex "person" has: "name", "age" and other attributes,
 "Software" includes: "name", "sale price" and other attributes; side "knowledge" includes: "date" attribute and so on.
 
-<center>
+<div style="text-align: center;">
   <img src="/docs/images/demo-graph-model.png" alt="image">
   <p>graph model example</p>
-</center>
+</div>
 
 
 After designing the graph model, we can use `groovy` to write the definition of `schema` and save it to a file, here named `schema.groovy`.
@@ -532,7 +532,7 @@ Mapping file for version 1.0:
 }
 ```
 
-The 1.0 version of the mapping file is centered on the vertex and edge, and sets the input source; while the 2.0 version is centered on the input source, and sets the vertex and edge mapping. Some input sources (such as a file) can generate both vertices and edges. If you write in the 1.0 format, you need to write an input block in each of the vertex and egde mapping blocks. The two input blocks are exactly the same ; and the 2.0 version only needs to write input once. Therefore, compare [...]
+The 1.0 version of the mapping file is centered on the vertex and edge, and sets the input source; while the 2.0 version is centered on the input source, and sets the vertex and edge mapping. Some input sources (such as a file) can generate both vertices and edges. If you write in the 1.0 format, you need to write an input block in each of the vertex and edge mapping blocks. The two input blocks are exactly the same ; and the 2.0 version only needs to write input once. Therefore, compare [...]
 
 In the bin directory of hugegraph-loader-{version}, there is a script tool `mapping-convert.sh` that can directly convert the mapping file of version 1.0 to version 2.0. The usage is as follows:
 
@@ -593,41 +593,41 @@ As mentioned above, it supports multiple relational databases, but because their
 
 **MYSQL**
 
-| Node | Fixed value or common value |
-| --- | --- |
-| vendor | MYSQL |
-| driver | com.mysql.cj.jdbc.Driver |
-| url | jdbc:mysql://127.0.0.1:3306 |
+| Node   | Fixed value or common value |
+|--------|-----------------------------|
+| vendor | MYSQL                       |
+| driver | com.mysql.cj.jdbc.Driver    |
+| url    | jdbc:mysql://127.0.0.1:3306 |
 
 schema: nullable, if filled in, it must be the same as the value of database
 
 **POSTGRESQL**
 
-| Node | Fixed value or common value |
-| --- | --- |
-| vendor | POSTGRESQL |
-| driver | org.postgresql.Driver |
-| url | jdbc:postgresql://127.0.0.1:5432 |
+| Node   | Fixed value or common value      |
+|--------|----------------------------------|
+| vendor | POSTGRESQL                       |
+| driver | org.postgresql.Driver            |
+| url    | jdbc:postgresql://127.0.0.1:5432 |
 
 schema: nullable, default is "public"
 
 **ORACLE**
 
-| Node | Fixed value or common value |
-| --- | --- |
-| vendor | ORACLE |
-| driver | oracle.jdbc.driver.OracleDriver |
-| url | jdbc:oracle:thin:@127.0.0.1:1521 |
+| Node   | Fixed value or common value      |
+|--------|----------------------------------|
+| vendor | ORACLE                           |
+| driver | oracle.jdbc.driver.OracleDriver  |
+| url    | jdbc:oracle:thin:@127.0.0.1:1521 |
 
 schema: nullable, the default value is the same as the username
 
 **SQLSERVER**
 
-| Node | Fixed value or common value |
-| --- | --- |
-| vendor | SQLSERVER |
+| Node   | Fixed value or common value                  |
+|--------|----------------------------------------------|
+| vendor | SQLSERVER                                    |
 | driver | com.microsoft.sqlserver.jdbc.SQLServerDriver |
-| url | jdbc:sqlserver://127.0.0.1:1433 |
+| url    | jdbc:sqlserver://127.0.0.1:1433              |
 
 schema: required
 
diff --git a/content/en/docs/quickstart/hugegraph-server.md b/content/en/docs/quickstart/hugegraph-server.md
index 6069a4e6..7dee5ee0 100644
--- a/content/en/docs/quickstart/hugegraph-server.md
+++ b/content/en/docs/quickstart/hugegraph-server.md
@@ -14,9 +14,9 @@ The Core Module is an implementation of the Tinkerpop interface; The Backend mod
 
 ### 2 Dependency
 
-#### 2.1 Install JDK-1.8
+#### 2.1 Install Java11 (JDK 11)
 
-HugeGraph-Server developed based on jdk-1.8, project's code uses many classes and methods in jdk-1.8, please install and configure by yourself.
+Consider use Java 11 to run `HugeGraph-Server`(also compatible with Java 8), and configure by yourself.
 
 **Be sure to execute the `java -version` command to check the jdk version before reading**