You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@doris.apache.org by mo...@apache.org on 2022/12/27 14:57:47 UTC

[doris] branch master updated: [typo](docs)fix doris docs 404 link (#15400)

This is an automated email from the ASF dual-hosted git repository.

morningman pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/doris.git


The following commit(s) were added to refs/heads/master by this push:
     new aad53d37c7 [typo](docs)fix doris docs 404 link  (#15400)
aad53d37c7 is described below

commit aad53d37c74886680e9efcd4007aafc839b88890
Author: lsy3993 <11...@users.noreply.github.com>
AuthorDate: Tue Dec 27 22:57:40 2022 +0800

    [typo](docs)fix doris docs 404 link  (#15400)
---
 .../docs/admin-manual/cluster-management/elastic-expansion.md  |  2 +-
 docs/en/docs/admin-manual/config/be-config.md                  |  4 ++--
 .../docs/admin-manual/http-actions/fe/table-schema-action.md   |  2 +-
 docs/en/docs/admin-manual/maint-monitor/disk-capacity.md       |  2 +-
 .../maint-monitor/memory-management/be-oom-analysis.md         |  8 ++++----
 .../memory-management/memory-limit-exceeded-analysis.md        |  2 +-
 docs/en/docs/advanced/alter-table/replace-table.md             |  4 ++--
 docs/en/docs/advanced/alter-table/schema-change.md             |  4 ++--
 .../data-operate/import/import-scenes/external-storage-load.md |  2 +-
 docs/en/docs/data-operate/import/import-scenes/jdbc-load.md    |  2 +-
 docs/en/docs/data-table/basic-usage.md                         |  4 ++--
 docs/en/docs/data-table/hit-the-rollup.md                      |  4 ++--
 docs/en/docs/ecosystem/external-table/hive-bitmap-udf.md       |  2 +-
 docs/en/docs/ecosystem/external-table/multi-catalog.md         |  2 +-
 .../Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md  |  2 +-
 .../Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md     |  2 +-
 .../Data-Manipulation-Statements/Load/BROKER-LOAD.md           |  2 +-
 .../Data-Manipulation-Statements/Load/STREAM-LOAD.md           |  4 ++--
 .../Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md      |  2 +-
 .../sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md    |  2 +-
 .../docs/admin-manual/cluster-management/elastic-expansion.md  |  4 ++--
 docs/zh-CN/docs/admin-manual/config/be-config.md               |  4 ++--
 .../docs/admin-manual/http-actions/fe/table-schema-action.md   |  2 +-
 docs/zh-CN/docs/admin-manual/maint-monitor/disk-capacity.md    |  2 +-
 .../maint-monitor/memory-management/be-oom-analysis.md         | 10 +++++-----
 .../memory-management/memory-limit-exceeded-analysis.md        |  4 ++--
 docs/zh-CN/docs/advanced/alter-table/replace-table.md          |  2 +-
 docs/zh-CN/docs/advanced/alter-table/schema-change.md          |  4 ++--
 .../data-operate/import/import-scenes/external-storage-load.md |  4 ++--
 docs/zh-CN/docs/data-operate/import/import-scenes/jdbc-load.md |  2 +-
 docs/zh-CN/docs/data-table/basic-usage.md                      |  4 ++--
 docs/zh-CN/docs/data-table/hit-the-rollup.md                   |  2 +-
 docs/zh-CN/docs/ecosystem/external-table/hive-bitmap-udf.md    |  2 +-
 docs/zh-CN/docs/ecosystem/external-table/multi-catalog.md      |  2 +-
 .../Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md  |  2 +-
 .../Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md     |  2 +-
 .../Data-Manipulation-Statements/Load/STREAM-LOAD.md           |  4 ++--
 .../Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md      |  2 +-
 .../sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md    |  2 +-
 39 files changed, 59 insertions(+), 59 deletions(-)

diff --git a/docs/en/docs/admin-manual/cluster-management/elastic-expansion.md b/docs/en/docs/admin-manual/cluster-management/elastic-expansion.md
index 12cf8fc27e..b921058c1a 100644
--- a/docs/en/docs/admin-manual/cluster-management/elastic-expansion.md
+++ b/docs/en/docs/admin-manual/cluster-management/elastic-expansion.md
@@ -106,7 +106,7 @@ You can also view the BE node through the front-end page connection: ``http://fe
 
 All of the above methods require Doris's root user rights.
 
-The expansion and scaling process of BE nodes does not affect the current system operation and the tasks being performed, and does not affect the performance of the current system. Data balancing is done automatically. Depending on the amount of data available in the cluster, the cluster will be restored to load balancing in a few hours to a day. For cluster load, see the [Tablet Load Balancing Document](../../maint-monitor/tablet-repair-and-balance).
+The expansion and scaling process of BE nodes does not affect the current system operation and the tasks being performed, and does not affect the performance of the current system. Data balancing is done automatically. Depending on the amount of data available in the cluster, the cluster will be restored to load balancing in a few hours to a day. For cluster load, see the [Tablet Load Balancing Document](../maint-monitor/tablet-repair-and-balance).
 
 ### Add BE nodes
 
diff --git a/docs/en/docs/admin-manual/config/be-config.md b/docs/en/docs/admin-manual/config/be-config.md
index 4d109b851d..a6e236f02a 100644
--- a/docs/en/docs/admin-manual/config/be-config.md
+++ b/docs/en/docs/admin-manual/config/be-config.md
@@ -346,7 +346,7 @@ There are two ways to configure BE configuration items:
 #### `doris_max_scan_key_num`
 
 * Type: int
-* Description: Used to limit the maximum number of scan keys that a scan node can split in a query request. When a conditional query request reaches the scan node, the scan node will try to split the conditions related to the key column in the query condition into multiple scan key ranges. After that, these scan key ranges will be assigned to multiple scanner threads for data scanning. A larger value usually means that more scanner threads can be used to increase the parallelism of the s [...]
+* Description: Used to limit the maximum number of scan keys that a scan node can split in a query request. When a conditional query request reaches the scan node, the scan node will try to split the conditions related to the key column in the query condition into multiple scan key ranges. After that, these scan key ranges will be assigned to multiple scanner threads for data scanning. A larger value usually means that more scanner threads can be used to increase the parallelism of the s [...]
   - When the concurrency cannot be improved in high concurrency scenarios, try to reduce this value and observe the impact.
 * Default value: 48
 
@@ -406,7 +406,7 @@ There are two ways to configure BE configuration items:
 #### `max_pushdown_conditions_per_column`
 
 * Type: int
-* Description: Used to limit the maximum number of conditions that can be pushed down to the storage engine for a single column in a query request. During the execution of the query plan, the filter conditions on some columns can be pushed down to the storage engine, so that the index information in the storage engine can be used for data filtering, reducing the amount of data that needs to be scanned by the query. Such as equivalent conditions, conditions in IN predicates, etc. In most  [...]
+* Description: Used to limit the maximum number of conditions that can be pushed down to the storage engine for a single column in a query request. During the execution of the query plan, the filter conditions on some columns can be pushed down to the storage engine, so that the index information in the storage engine can be used for data filtering, reducing the amount of data that needs to be scanned by the query. Such as equivalent conditions, conditions in IN predicates, etc. In most  [...]
 * Default value: 1024
 
 * Example
diff --git a/docs/en/docs/admin-manual/http-actions/fe/table-schema-action.md b/docs/en/docs/admin-manual/http-actions/fe/table-schema-action.md
index 62b6a457b3..4cdb7528dd 100644
--- a/docs/en/docs/admin-manual/http-actions/fe/table-schema-action.md
+++ b/docs/en/docs/admin-manual/http-actions/fe/table-schema-action.md
@@ -97,7 +97,7 @@ None
 	"count": 0
 }
 ```
-Note: The difference is that the `http` method returns more `aggregation_type` fields than the `http v2` method. The `http v2` is enabled by setting `enable_http_server_v2`. For detailed parameter descriptions, see [fe parameter settings](../../../config/fe-config)
+Note: The difference is that the `http` method returns more `aggregation_type` fields than the `http v2` method. The `http v2` is enabled by setting `enable_http_server_v2`. For detailed parameter descriptions, see [fe parameter settings](../../config/fe-config)
 
 ## Examples
 
diff --git a/docs/en/docs/admin-manual/maint-monitor/disk-capacity.md b/docs/en/docs/admin-manual/maint-monitor/disk-capacity.md
index 73da70e49d..047b7f6eb5 100644
--- a/docs/en/docs/admin-manual/maint-monitor/disk-capacity.md
+++ b/docs/en/docs/admin-manual/maint-monitor/disk-capacity.md
@@ -162,6 +162,6 @@ When the disk capacity is higher than High Watermark or even Flood Stage, many o
 
         ```rm -rf data/0/12345/```
 
-    * Delete tablet metadata (refer to [Tablet metadata management tool](./tablet-meta-tool.md))
+    * Delete tablet metadata (refer to [Tablet metadata management tool](./tablet-meta-tool))
 
         ```./lib/meta_tool --operation=delete_header --root_path=/path/to/root_path --tablet_id=12345 --schema_hash= 352781111```
diff --git a/docs/en/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md b/docs/en/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
index aa3061a727..be8662dce6 100644
--- a/docs/en/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
+++ b/docs/en/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
@@ -28,10 +28,10 @@ under the License.
 
 <version since="1.2.0">
 
-Ideally, in [Memory Limit Exceeded Analysis](../admin-manual/memory-management/memory-limit-exceeded-analysis.md), we regularly detect the remaining available memory of the operating system and respond in time when the memory is insufficient , such as triggering the memory GC to release the cache or cancel the memory overrun query, but because refreshing process memory statistics and memory GC both have a certain lag, and it is difficult for us to completely catch all large memory applic [...]
+Ideally, in [Memory Limit Exceeded Analysis](./memory-limit-exceeded-analysis), we regularly detect the remaining available memory of the operating system and respond in time when the memory is insufficient , such as triggering the memory GC to release the cache or cancel the memory overrun query, but because refreshing process memory statistics and memory GC both have a certain lag, and it is difficult for us to completely catch all large memory applications, there are still OOM risk.
 
 ## Solution
-Refer to [BE Configuration Items](../admin-manual/config/be-config.md) to reduce `mem_limit` and increase `max_sys_mem_available_low_water_mark_bytes` in `be.conf`.
+Refer to [BE Configuration Items](../../../admin-manual/config/be-config) to reduce `mem_limit` and increase `max_sys_mem_available_low_water_mark_bytes` in `be.conf`.
 
 ## Memory analysis
 If you want to further understand the memory usage location of the BE process before OOM and reduce the memory usage of the process, you can refer to the following steps to analyze.
@@ -75,8 +75,8 @@ Memory Tracker Summary:
 
 6. `type=load` imports a lot of memory.
 
-7. When the `type=global` memory is used for a long time, continue to check the `type=global` detailed statistics in the second half of the `Memory Tracker Summary` log. When DataPageCache, IndexPageCache, SegmentCache, ChunkAllocator, LastestSuccessChannelCache, etc. use a lot of memory, refer to [BE Configuration Item](../admin-manual/config/be-config.md) to consider modifying the size of the cache; when Orphan memory usage is too large, Continue the analysis as follows.
-  - If the sum of the tracker statistics of `Parent Label=Orphan` only accounts for a small part of the Orphan memory, it means that there is currently a large amount of memory that has no accurate statistics, such as the memory of the brpc process. At this time, you can consider using the heap profile [Memory Tracker]( ../community/developer-guide/debug-tool.md) to further analyze memory locations.
+7. When the `type=global` memory is used for a long time, continue to check the `type=global` detailed statistics in the second half of the `Memory Tracker Summary` log. When DataPageCache, IndexPageCache, SegmentCache, ChunkAllocator, LastestSuccessChannelCache, etc. use a lot of memory, refer to [BE Configuration Item](../../../admin-manual/config/be-config) to consider modifying the size of the cache; when Orphan memory usage is too large, Continue the analysis as follows.
+  - If the sum of the tracker statistics of `Parent Label=Orphan` only accounts for a small part of the Orphan memory, it means that there is currently a large amount of memory that has no accurate statistics, such as the memory of the brpc process. At this time, you can consider using the heap profile [Memory Tracker]( ../../../../community/developer-guide/debug-tool) to further analyze memory locations.
   - If the tracker statistics of `Parent Label=Orphan` account for most of Orphan’s memory, when `Label=TabletManager` uses a lot of memory, further check the number of tablets in the cluster. If there are too many tablets, delete them and they will not be used table or data; when `Label=StorageEngine` uses too much memory, further check the number of segment files in the cluster, and consider manually triggering compaction if the number of segment files is too large;
 
 8. If `be/log/be.INFO` does not print the `Memory Tracker Summary` log before OOM, it means that BE did not detect the memory limit in time, observe Grafana memory monitoring to confirm the memory growth trend of BE before OOM, if OOM is reproducible, consider adding `memory_debug=true` in `be.conf`, after restarting the cluster, the cluster memory statistics will be printed every second, observe the last `Memory Tracker Summary` log before OOM, and continue to step 3 for analysis;
diff --git a/docs/en/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md b/docs/en/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
index fad526c37e..5407303655 100644
--- a/docs/en/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
+++ b/docs/en/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
@@ -49,7 +49,7 @@ The error message is divided into three parts:
 At the same time, you can find the following log in log/be.INFO to confirm whether the memory usage of the current process meets expectations. The log is also divided into three parts:
 1. `Process Memory Summary`: process memory statistics.
 2. `Alloc Stacktrace`: The stack that triggers the memory overrun detection, which is not necessarily the location of the large memory application.
-3. `Memory Tracker Summary`: Process memory tracker statistics, refer to [Memory Tracker](../admin-manual/memory-management/memory-tracker.md) to analyze the location of memory usage.
+3. `Memory Tracker Summary`: Process memory tracker statistics, refer to [Memory Tracker](./memory-tracker) to analyze the location of memory usage.
 Notice:
 1. The printing interval of the process memory overrun log is 1s. After the process memory exceeds the limit, the memory applications in most locations of BE will sense it, and try to make a predetermined callback method, and print the process memory overrun log, so if the log is If the value of Try Alloc is small, you don’t need to pay attention to `Alloc Stacktrace`, just analyze `Memory Tracker Summary` directly.
 2. When the process memory exceeds the limit, BE will trigger memory GC.
diff --git a/docs/en/docs/advanced/alter-table/replace-table.md b/docs/en/docs/advanced/alter-table/replace-table.md
index 21fd3b95d0..204b395a44 100644
--- a/docs/en/docs/advanced/alter-table/replace-table.md
+++ b/docs/en/docs/advanced/alter-table/replace-table.md
@@ -29,7 +29,7 @@ under the License.
 In version 0.14, Doris supports atomic replacement of two tables.
 This operation only applies to OLAP tables.
 
-For partition level replacement operations, please refer to [Temporary Partition Document](../../partition/table-temp-partition)
+For partition level replacement operations, please refer to [Temporary Partition Document](../partition/table-temp-partition)
 
 ## Syntax
 
@@ -69,4 +69,4 @@ If `swap` is `false`, the operation is as follows:
 
 1. Atomic Overwrite Operation
 
-    In some cases, the user wants to be able to rewrite the data of a certain table, but if it is dropped and then imported, there will be a period of time in which the data cannot be viewed. At this time, the user can first use the `CREATE TABLE LIKE` statement to create a new table with the same structure, import the new data into the new table, and replace the old table atomically through the replacement operation to achieve the goal. For partition level atomic overwrite operation, pl [...]
+    In some cases, the user wants to be able to rewrite the data of a certain table, but if it is dropped and then imported, there will be a period of time in which the data cannot be viewed. At this time, the user can first use the `CREATE TABLE LIKE` statement to create a new table with the same structure, import the new data into the new table, and replace the old table atomically through the replacement operation to achieve the goal. For partition level atomic overwrite operation, pl [...]
diff --git a/docs/en/docs/advanced/alter-table/schema-change.md b/docs/en/docs/advanced/alter-table/schema-change.md
index 0503dac489..7e8ca9e3f9 100644
--- a/docs/en/docs/advanced/alter-table/schema-change.md
+++ b/docs/en/docs/advanced/alter-table/schema-change.md
@@ -68,7 +68,7 @@ The basic process of executing a Schema Change is to generate a copy of the inde
 Before starting the conversion of historical data, Doris will obtain a latest transaction ID. And wait for all import transactions before this Transaction ID to complete. This Transaction ID becomes a watershed. This means that Doris guarantees that all import tasks after the watershed will generate data for both the original Index and the new Index. In this way, when the historical data conversion is completed, the data in the new Index can be guaranteed to be complete.
 ## Create Job
 
-The specific syntax for creating a Schema Change can be found in the help [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN.md) for the description of the Schema Change section .
+The specific syntax for creating a Schema Change can be found in the help [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN) for the description of the Schema Change section .
 
 The creation of Schema Change is an asynchronous process. After the job is submitted successfully, the user needs to view the job progress through the `SHOW ALTER TABLE COLUMN` command.
 ## View Job
@@ -283,5 +283,5 @@ SHOW ALTER TABLE COLUMN\G;
 
 ## More Help
 
-For more detailed syntax and best practices used by Schema Change, see [ALTER TABLE COLUMN](../../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN) command manual, you can also enter `HELP ALTER TABLE COLUMN` in the MySql client command line for more help information.
+For more detailed syntax and best practices used by Schema Change, see [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN) command manual, you can also enter `HELP ALTER TABLE COLUMN` in the MySql client command line for more help information.
 
diff --git a/docs/en/docs/data-operate/import/import-scenes/external-storage-load.md b/docs/en/docs/data-operate/import/import-scenes/external-storage-load.md
index ab8bfa7677..eaff843294 100644
--- a/docs/en/docs/data-operate/import/import-scenes/external-storage-load.md
+++ b/docs/en/docs/data-operate/import/import-scenes/external-storage-load.md
@@ -82,7 +82,7 @@ Hdfs load creates an import statement. The import method is basically the same a
 
 3. Check import status
 
-   Broker load is an asynchronous import method. The specific import results can be accessed through [SHOW LOAD](../../../../sql-manual/sql-reference/Show-Statements/SHOW-LOAD) command to view
+   Broker load is an asynchronous import method. The specific import results can be accessed through [SHOW LOAD](../../../sql-manual/sql-reference/Show-Statements/SHOW-LOAD) command to view
    
    ```
    mysql> show load order by createtime desc limit 1\G;
diff --git a/docs/en/docs/data-operate/import/import-scenes/jdbc-load.md b/docs/en/docs/data-operate/import/import-scenes/jdbc-load.md
index 4dd61e18a6..45c743dc6d 100644
--- a/docs/en/docs/data-operate/import/import-scenes/jdbc-load.md
+++ b/docs/en/docs/data-operate/import/import-scenes/jdbc-load.md
@@ -160,4 +160,4 @@ Please note the following:
 
    As mentioned earlier, we recommend that when using INSERT to import data, use the "batch" method to import, rather than a single insert.
 
-   At the same time, we can set a Label for each INSERT operation. Through the [Label mechanism](../load-atomicity), the idempotency and atomicity of operations can be guaranteed, and the data will not be lost or heavy in the end. For the specific usage of Label in INSERT, you can refer to the [INSERT](../../../sql-manual/sql-reference/Data-Manipulation-Statements/Manipulation/INSERT) document.
+   At the same time, we can set a Label for each INSERT operation. Through the [Label mechanism](./load-atomicity), the idempotency and atomicity of operations can be guaranteed, and the data will not be lost or heavy in the end. For the specific usage of Label in INSERT, you can refer to the [INSERT](../../../sql-manual/sql-reference/Data-Manipulation-Statements/Manipulation/INSERT) document.
diff --git a/docs/en/docs/data-table/basic-usage.md b/docs/en/docs/data-table/basic-usage.md
index 3f05aa5e1f..8da31f9652 100644
--- a/docs/en/docs/data-table/basic-usage.md
+++ b/docs/en/docs/data-table/basic-usage.md
@@ -142,7 +142,7 @@ mysql> USE example_db;
 Database changed
 ```
 
-Doris supports [composite partition and single partition](../data-partition)  two table building methods. The following takes the aggregation model as an example to demonstrate how to create two partitioned data tables.
+Doris supports [composite partition and single partition](./data-partition)  two table building methods. The following takes the aggregation model as an example to demonstrate how to create two partitioned data tables.
 
 #### Single partition
 
@@ -674,7 +674,7 @@ mysql> select sum(table1.pv) from table1 join [shuffle] table2 where table1.site
 
 When deploying multiple FE nodes, you can deploy a load balancing layer on top of multiple FEs to achieve high availability of Doris.
 
-Please refer to [Load Balancing](../../admin-manual/cluster-management/load-balancing) for details on installation, deployment, and usage.
+Please refer to [Load Balancing](../admin-manual/cluster-management/load-balancing) for details on installation, deployment, and usage.
 
 ## Data update and deletion
 
diff --git a/docs/en/docs/data-table/hit-the-rollup.md b/docs/en/docs/data-table/hit-the-rollup.md
index 990518c39c..9461696538 100644
--- a/docs/en/docs/data-table/hit-the-rollup.md
+++ b/docs/en/docs/data-table/hit-the-rollup.md
@@ -44,7 +44,7 @@ Because Uniq is only a special case of the Aggregate model, we do not distinguis
 
 Example 1: Get the total consumption per user
 
-Following [Data Model Aggregate Model](./data-model.md) in the **Aggregate Model** section, the Base table structure is as follows:
+Following [Data Model Aggregate Model](./data-model) in the **Aggregate Model** section, the Base table structure is as follows:
 
 | ColumnName        | Type         | AggregationType | Comment                                |
 |-------------------| ------------ | --------------- | -------------------------------------- |
@@ -128,7 +128,7 @@ Doris automatically hits the ROLLUP table.
 
 #### ROLLUP in Duplicate Model
 
-Because the Duplicate model has no aggregate semantics. So the ROLLLUP in this model has lost the meaning of "scroll up". It's just to adjust the column order to hit the prefix index. In the next section, we will introduce prefix index in [data model prefix index](./data-model.md), and how to use ROLLUP to change prefix index in order to achieve better query efficiency.
+Because the Duplicate model has no aggregate semantics. So the ROLLLUP in this model has lost the meaning of "scroll up". It's just to adjust the column order to hit the prefix index. In the next section, we will introduce prefix index in [data model prefix index](./data-model), and how to use ROLLUP to change prefix index in order to achieve better query efficiency.
 
 ## ROLLUP adjusts prefix index
 
diff --git a/docs/en/docs/ecosystem/external-table/hive-bitmap-udf.md b/docs/en/docs/ecosystem/external-table/hive-bitmap-udf.md
index c56fa7a3ae..bf4fdbf950 100644
--- a/docs/en/docs/ecosystem/external-table/hive-bitmap-udf.md
+++ b/docs/en/docs/ecosystem/external-table/hive-bitmap-udf.md
@@ -106,4 +106,4 @@ select k1,bitmap_union(uuid) from hive_bitmap_table group by k1
 
 ## Hive Bitmap import into Doris
 
- see details: [Spark Load](../../../data-operate/import/import-way/spark-load-manual) -> Basic operation -> Create load(Example 3: when the upstream data source is hive binary type table)
+ see details: [Spark Load](../../data-operate/import/import-way/spark-load-manual) -> Basic operation -> Create load(Example 3: when the upstream data source is hive binary type table)
diff --git a/docs/en/docs/ecosystem/external-table/multi-catalog.md b/docs/en/docs/ecosystem/external-table/multi-catalog.md
index 08228e3290..8c52432dcf 100644
--- a/docs/en/docs/ecosystem/external-table/multi-catalog.md
+++ b/docs/en/docs/ecosystem/external-table/multi-catalog.md
@@ -67,7 +67,7 @@ This function will be used as a supplement and enhancement to the previous exter
     
 4. Drop Catalog
 
-    Both Database and Table in External Catalog are read-only. However, the catalog can be deleted (Internal Catalog cannot be deleted). An External Catalog can be dropped via the [DROP CATALOG](../../../sql-manual/sql-reference/Data-Definition-Statements/Drop/DROP-CATALOG) command.
+    Both Database and Table in External Catalog are read-only. However, the catalog can be deleted (Internal Catalog cannot be deleted). An External Catalog can be dropped via the [DROP CATALOG](../../sql-manual/sql-reference/Data-Definition-Statements/Drop/DROP-CATALOG) command.
 
     This operation will only delete the mapping information of the catalog in Doris, and will not modify or change the contents of any external data source.
 
diff --git a/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md b/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
index d8beb732c4..69b81b5c18 100644
--- a/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
+++ b/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
@@ -62,7 +62,7 @@ Notice:
 - The partition is left closed and right open. If the user only specifies the right boundary, the system will automatically determine the left boundary
 - If the bucketing method is not specified, the bucketing method and bucket number used for creating the table would be automatically used
 - If the bucketing method is specified, only the number of buckets can be modified, not the bucketing method or the bucketing column. If the bucketing method is specified but the number of buckets not be specified, the default value `10` will be used for bucket number instead of the number specified when the table is created. If the number of buckets modified, the bucketing method needs to be specified simultaneously.
-- The ["key"="value"] section can set some attributes of the partition, see [CREATE TABLE](../../Create/CREATE-TABLE)
+- The ["key"="value"] section can set some attributes of the partition, see [CREATE TABLE](../Create/CREATE-TABLE)
 - If the user does not explicitly create a partition when creating a table, adding a partition by ALTER is not supported
 
 2. Delete the partition
diff --git a/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md b/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
index c19578901c..596e7cfcc5 100644
--- a/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
+++ b/docs/en/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
@@ -68,7 +68,7 @@ Notice:
 
 - If from_index_name is not specified, it will be created from base index by default
 - Columns in rollup table must be columns already in from_index
-- In properties, the storage format can be specified. For details, see [CREATE TABLE](../../Create/CREATE-TABLE)
+- In properties, the storage format can be specified. For details, see [CREATE TABLE](../Create/CREATE-TABLE)
 
 3. Delete rollup index
 
diff --git a/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md b/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
index d2506675d9..4c7726d666 100644
--- a/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
+++ b/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/BROKER-LOAD.md
@@ -489,7 +489,7 @@ WITH BROKER broker_name
 
 2. Cancel the import task
 
-   Import tasks that have been submitted but not yet completed can be canceled by the [CANCEL LOAD](./CANCEL-LOAD.md) command. After cancellation, the written data will also be rolled back and will not take effect.
+   Import tasks that have been submitted but not yet completed can be canceled by the [CANCEL LOAD](./CANCEL-LOAD) command. After cancellation, the written data will also be rolled back and will not take effect.
 
 3. Label, import transaction, multi-table atomicity
 
diff --git a/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md b/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
index fcb2b92b31..000d0784b9 100644
--- a/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
+++ b/docs/en/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
@@ -431,11 +431,11 @@ curl --location-trusted -u root -H "columns: k1,k2,source_sequence,v1,v2" -H "fu
 
 4. Label, import transaction, multi-table atomicity
 
-   All import tasks in Doris are atomic. And the import of multiple tables in the same import task can also guarantee atomicity. At the same time, Doris can also use the Label mechanism to ensure that the data imported is not lost or heavy. For details, see the [Import Transactions and Atomicity](../../../../../data-operate/import/import-scenes/load-atomicity) documentation.
+   All import tasks in Doris are atomic. And the import of multiple tables in the same import task can also guarantee atomicity. At the same time, Doris can also use the Label mechanism to ensure that the data imported is not lost or heavy. For details, see the [Import Transactions and Atomicity](../../../../data-operate/import/import-scenes/load-atomicity) documentation.
 
 5. Column mapping, derived columns and filtering
 
-   Doris can support very rich column transformation and filtering operations in import statements. Most built-in functions and UDFs are supported. For how to use this function correctly, please refer to the [Column Mapping, Conversion and Filtering](../../../../../data-operate/import/import-scenes/load-data-convert) document.
+   Doris can support very rich column transformation and filtering operations in import statements. Most built-in functions and UDFs are supported. For how to use this function correctly, please refer to the [Column Mapping, Conversion and Filtering](../../../../data-operate/import/import-scenes/load-data-convert) document.
 
 6. Error data filtering
 
diff --git a/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md b/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
index 020935ae99..d14842f947 100644
--- a/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
+++ b/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
@@ -32,7 +32,7 @@ SHOW ALTER TABLE MATERIALIZED VIEW
 
 ### Description
 
-This command is used to view the execution of the Create Materialized View job submitted through the [CREATE-MATERIALIZED-VIEW](../../../sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) statement.
+This command is used to view the execution of the Create Materialized View job submitted through the [CREATE-MATERIALIZED-VIEW](../../sql-reference/Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) statement.
 
 > This statement is equivalent to `SHOW ALTER TABLE ROLLUP`;
 
diff --git a/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md b/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
index 26707d10aa..a531011b61 100644
--- a/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
+++ b/docs/en/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
@@ -32,7 +32,7 @@ SHOW STATUS
 
 ### Description
 
-This command is used to view the execution of the Create Materialized View job submitted through the [CREATE-MATERIALIZED-VIEW](../../Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) statement.
+This command is used to view the execution of the Create Materialized View job submitted through the [CREATE-MATERIALIZED-VIEW](../Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) statement.
 
 > This statement is equivalent to `SHOW ALTER TABLE ROLLUP`;
 
diff --git a/docs/zh-CN/docs/admin-manual/cluster-management/elastic-expansion.md b/docs/zh-CN/docs/admin-manual/cluster-management/elastic-expansion.md
index 20e14bd105..a966461b34 100644
--- a/docs/zh-CN/docs/admin-manual/cluster-management/elastic-expansion.md
+++ b/docs/zh-CN/docs/admin-manual/cluster-management/elastic-expansion.md
@@ -102,7 +102,7 @@ FE 分为 Leader,Follower 和 Observer 三种角色。 默认一个集群,
 
 以上方式,都需要 Doris 的 root 用户权限。
 
-BE 节点的扩容和缩容过程,不影响当前系统运行以及正在执行的任务,并且不会影响当前系统的性能。数据均衡会自动进行。根据集群现有数据量的大小,集群会在几个小时到1天不等的时间内,恢复到负载均衡的状态。集群负载情况,可以参见 [Tablet 负载均衡文档](../../maint-monitor/tablet-repair-and-balance)。
+BE 节点的扩容和缩容过程,不影响当前系统运行以及正在执行的任务,并且不会影响当前系统的性能。数据均衡会自动进行。根据集群现有数据量的大小,集群会在几个小时到1天不等的时间内,恢复到负载均衡的状态。集群负载情况,可以参见 [Tablet 负载均衡文档](../maint-monitor/tablet-repair-and-balance)。
 
 ### 增加 BE 节点
 
@@ -134,7 +134,7 @@ DECOMMISSION 语句如下:
      > 		```CANCEL DECOMMISSION BACKEND "be_host:be_heartbeat_service_port";```  
      > 	命令取消。取消后,该 BE 上的数据将维持当前剩余的数据量。后续 Doris 重新进行负载均衡
 
-**对于多租户部署环境下,BE 节点的扩容和缩容,请参阅 [多租户设计文档](../multi-tenant.md)。**
+**对于多租户部署环境下,BE 节点的扩容和缩容,请参阅 [多租户设计文档](../multi-tenant)。**
 
 ## Broker 扩容缩容
 
diff --git a/docs/zh-CN/docs/admin-manual/config/be-config.md b/docs/zh-CN/docs/admin-manual/config/be-config.md
index bed8d275bb..008274586c 100644
--- a/docs/zh-CN/docs/admin-manual/config/be-config.md
+++ b/docs/zh-CN/docs/admin-manual/config/be-config.md
@@ -357,7 +357,7 @@ BE 重启后该配置将失效。如果想持久化修改结果,使用如下
 #### `doris_max_scan_key_num`
 
 * 类型:int
-* 描述:用于限制一个查询请求中,scan node 节点能拆分的最大 scan key 的个数。当一个带有条件的查询请求到达 scan node 节点时,scan node 会尝试将查询条件中 key 列相关的条件拆分成多个 scan key range。之后这些 scan key range 会被分配给多个 scanner 线程进行数据扫描。较大的数值通常意味着可以使用更多的 scanner 线程来提升扫描操作的并行度。但在高并发场景下,过多的线程可能会带来更大的调度开销和系统负载,反而会降低查询响应速度。一个经验数值为 50。该配置可以单独进行会话级别的配置,具体可参阅 [变量](../../advanced/variables.md) 中 `max_scan_key_num` 的说明。
+* 描述:用于限制一个查询请求中,scan node 节点能拆分的最大 scan key 的个数。当一个带有条件的查询请求到达 scan node 节点时,scan node 会尝试将查询条件中 key 列相关的条件拆分成多个 scan key range。之后这些 scan key range 会被分配给多个 scanner 线程进行数据扫描。较大的数值通常意味着可以使用更多的 scanner 线程来提升扫描操作的并行度。但在高并发场景下,过多的线程可能会带来更大的调度开销和系统负载,反而会降低查询响应速度。一个经验数值为 50。该配置可以单独进行会话级别的配置,具体可参阅 [变量](../../advanced/variables) 中 `max_scan_key_num` 的说明。
   - 当在高并发场景下发下并发度无法提升时,可以尝试降低该数值并观察影响。
 * 默认值:48
 
@@ -417,7 +417,7 @@ BE 重启后该配置将失效。如果想持久化修改结果,使用如下
 #### `max_pushdown_conditions_per_column`
 
 * 类型:int
-* 描述:用于限制一个查询请求中,针对单个列,能够下推到存储引擎的最大条件数量。在查询计划执行的过程中,一些列上的过滤条件可以下推到存储引擎,这样可以利用存储引擎中的索引信息进行数据过滤,减少查询需要扫描的数据量。比如等值条件、IN 谓词中的条件等。这个参数在绝大多数情况下仅影响包含 IN 谓词的查询。如 `WHERE colA IN (1,2,3,4,...)`。较大的数值意味值 IN 谓词中更多的条件可以推送给存储引擎,但过多的条件可能会导致随机读的增加,某些情况下可能会降低查询效率。该配置可以单独进行会话级别的配置,具体可参阅 [变量](../../advanced/variables.md) 中 `max_pushdown_conditions_per_column ` 的说明。
+* 描述:用于限制一个查询请求中,针对单个列,能够下推到存储引擎的最大条件数量。在查询计划执行的过程中,一些列上的过滤条件可以下推到存储引擎,这样可以利用存储引擎中的索引信息进行数据过滤,减少查询需要扫描的数据量。比如等值条件、IN 谓词中的条件等。这个参数在绝大多数情况下仅影响包含 IN 谓词的查询。如 `WHERE colA IN (1,2,3,4,...)`。较大的数值意味值 IN 谓词中更多的条件可以推送给存储引擎,但过多的条件可能会导致随机读的增加,某些情况下可能会降低查询效率。该配置可以单独进行会话级别的配置,具体可参阅 [变量](../../advanced/variables) 中 `max_pushdown_conditions_per_column ` 的说明。
 * 默认值:1024
 
 * 示例
diff --git a/docs/zh-CN/docs/admin-manual/http-actions/fe/table-schema-action.md b/docs/zh-CN/docs/admin-manual/http-actions/fe/table-schema-action.md
index 8f5a456971..6ec41bbd48 100644
--- a/docs/zh-CN/docs/admin-manual/http-actions/fe/table-schema-action.md
+++ b/docs/zh-CN/docs/admin-manual/http-actions/fe/table-schema-action.md
@@ -97,7 +97,7 @@ under the License.
 	"count": 0
 }
 ```
-注意:区别为`http`方式比`http v2`方式多返回`aggregation_type`字段,`http v2`开启是通过`enable_http_server_v2`进行设置,具体参数说明详见[fe参数设置](../../../config/fe-config.md)
+注意:区别为`http`方式比`http v2`方式多返回`aggregation_type`字段,`http v2`开启是通过`enable_http_server_v2`进行设置,具体参数说明详见[fe参数设置](../../config/fe-config)
 
 ## Examples
 
diff --git a/docs/zh-CN/docs/admin-manual/maint-monitor/disk-capacity.md b/docs/zh-CN/docs/admin-manual/maint-monitor/disk-capacity.md
index e83ec4592d..ca3c9172e0 100644
--- a/docs/zh-CN/docs/admin-manual/maint-monitor/disk-capacity.md
+++ b/docs/zh-CN/docs/admin-manual/maint-monitor/disk-capacity.md
@@ -156,6 +156,6 @@ capacity_min_left_bytes_flood_stage 默认 1GB。
 
     `rm -rf data/0/12345/`
 
-  - 删除 Tablet 元数据(具体参考 [Tablet 元数据管理工具](./tablet-meta-tool.md))
+  - 删除 Tablet 元数据(具体参考 [Tablet 元数据管理工具](./tablet-meta-tool))
 
     `./lib/meta_tool --operation=delete_header --root_path=/path/to/root_path --tablet_id=12345 --schema_hash= 352781111`
diff --git a/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md b/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
index 257e93d95c..41db09f3b3 100644
--- a/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
+++ b/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/be-oom-analysis.md
@@ -28,10 +28,10 @@ under the License.
 
 <version since="1.2.0">
 
-理想情况下,在 [Memory Limit Exceeded Analysis](../admin-manual/memory-management/memory-limit-exceeded-analysis.md) 中我们定时检测操作系统剩余可用内存,并在内存不足时及时响应,如触发内存GC释放缓存或cancel内存超限的查询,但因为刷新进程内存统计和内存GC都具有一定的滞后性,同时我们很难完全catch所有大内存申请,在集群压力过大时仍有OOM风险。
+理想情况下,在 [Memory Limit Exceeded Analysis](./memory-limit-exceeded-analysis) 中我们定时检测操作系统剩余可用内存,并在内存不足时及时响应,如触发内存GC释放缓存或cancel内存超限的查询,但因为刷新进程内存统计和内存GC都具有一定的滞后性,同时我们很难完全catch所有大内存申请,在集群压力过大时仍有OOM风险。
 
 ## 解决方法
-参考 [BE 配置项](../admin-manual/config/be-config.md) 在`be.conf`中调小`mem_limit`,调大`max_sys_mem_available_low_water_mark_bytes`。
+参考 [BE 配置项](../../../admin-manual/config/be-config) 在`be.conf`中调小`mem_limit`,调大`max_sys_mem_available_low_water_mark_bytes`。
 
 ## 内存分析
 若希望进一步了解 OOM 前BE进程的内存使用位置,减少进程内存使用,可参考如下步骤分析。
@@ -67,7 +67,7 @@ Memory Tracker Summary:
     MemTrackerLimiter Label=DeleteBitmap AggCache, Type=global, Limit=-1.00 B(-1 B), Used=0(0 B), Peak=0(0 B)
 ```
 
-3. 当 OOM 前 be/log/be.INFO 的最后包含系统内存超限的日志时,参考 [Memory Limit Exceeded Analysis](../admin-manual/memory-management/memory-limit-exceeded-analysis.md) 中的日志分析方法,查看进程每个类别的内存使用情况。若当前是`type=query`内存使用较多,若已知 OOM 前的查询继续步骤4,否则继续步骤5;若当前是`type=load`内存使用多继续步骤6,若当前是`type=global`内存使用多继续步骤7。
+3. 当 OOM 前 be/log/be.INFO 的最后包含系统内存超限的日志时,参考 [Memory Limit Exceeded Analysis](./memory-limit-exceeded-analysis) 中的日志分析方法,查看进程每个类别的内存使用情况。若当前是`type=query`内存使用较多,若已知 OOM 前的查询继续步骤4,否则继续步骤5;若当前是`type=load`内存使用多继续步骤6,若当前是`type=global`内存使用多继续步骤7。
 
 4. `type=query`查询内存使用多,且已知 OOM 前的查询时,比如测试集群或定时任务,重启BE节点,参考 [Memory Tracker](../admin-manual/memory-management/memory-tracker.md) 查看实时 memory tracker 统计,`set global enable_profile=true`后重试查询,观察具体算子的内存使用位置,确认查询内存使用是否合理,进一步考虑优化SQL内存使用,比如调整join顺序。
 
@@ -75,8 +75,8 @@ Memory Tracker Summary:
 
 6. `type=load`导入内存使用多时。
 
-7. `type=global`内存使用多时,继续查看`Memory Tracker Summary`日志后半部分已经打出得`type=global`详细统计。当 DataPageCache、IndexPageCache、SegmentCache、ChunkAllocator、LastestSuccessChannelCache 等内存使用多时,参考 [BE 配置项](../admin-manual/config/be-config.md) 考虑修改cache的大小;当 Orphan 内存使用过多时,如下继续分析。
-  - 若`Parent Label=Orphan`的tracker统计值相加只占 Orphan 内存的小部分,则说明当前有大量内存没有准确统计,比如 brpc 过程的内存,此时可以考虑借助 heap profile [Memory Tracker](../community/developer-guide/debug-tool.md) 中的方法进一步分析内存位置。
+7. `type=global`内存使用多时,继续查看`Memory Tracker Summary`日志后半部分已经打出得`type=global`详细统计。当 DataPageCache、IndexPageCache、SegmentCache、ChunkAllocator、LastestSuccessChannelCache 等内存使用多时,参考 [BE 配置项](../../../admin-manual/config/be-config) 考虑修改cache的大小;当 Orphan 内存使用过多时,如下继续分析。
+  - 若`Parent Label=Orphan`的tracker统计值相加只占 Orphan 内存的小部分,则说明当前有大量内存没有准确统计,比如 brpc 过程的内存,此时可以考虑借助 heap profile [Memory Tracker](../../../../community/developer-guide/debug-tool) 中的方法进一步分析内存位置。
   - 若`Parent Label=Orphan`的tracker统计值相加占 Orphan 内存的大部分,当`Label=TabletManager`内存使用多时,进一步查看集群 Tablet 数量,若 Tablet 数量过多则考虑删除过时不会被使用的表或数据;当`Label=StorageEngine`内存使用过多时,进一步查看集群 Segment 文件个数,若 Segment 文件个数过多则考虑手动触发compaction;
 
 8. 若`be/log/be.INFO`没有在 OOM 前打印出`Memory Tracker Summary`日志,说明 BE 没有及时检测出内存超限,观察 Grafana 内存监控确认BE在 OOM 前的内存增长趋势,若 OOM 可复现,考虑在`be.conf`中增加`memory_debug=true`,重启集群后会每秒打印集群内存统计,观察 OOM 前的最后一次`Memory Tracker Summary`日志,继续步骤3分析;
diff --git a/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md b/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
index 432c3232b6..e464210162 100644
--- a/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
+++ b/docs/zh-CN/docs/admin-manual/maint-monitor/memory-management/memory-limit-exceeded-analysis.md
@@ -49,7 +49,7 @@ ERROR 1105 (HY000): errCode = 2, detailMessage = Memory limit exceeded:<consumin
 同时可以在 log/be.INFO 中找到如下日志,确认当前进程内存使用是否符合预期,日志同样分为三部分:
 1、`Process Memory Summary`:进程内存统计。
 2、`Alloc Stacktrace`:触发内存超限检测的栈,这不一定是大内存申请的位置。
-3、`Memory Tracker Summary`:进程 memory tracker 统计,参考 [Memory Tracker](../admin-manual/memory-management/memory-tracker.md) 分析使用内存的位置。
+3、`Memory Tracker Summary`:进程 memory tracker 统计,参考 [Memory Tracker](./memory-tracker) 分析使用内存的位置。
 注意:
 1、进程内存超限日志的打印间隔是1s,进程内存超限后,BE大多数位置的内存申请都会感知,并尝试做出预定的回调方法,并打印进程内存超限日志,所以如果日志中 Try Alloc 的值很小,则无须关注`Alloc Stacktrace`,直接分析`Memory Tracker Summary`即可。
 2、当进程内存超限后,BE会触发内存GC。
@@ -130,7 +130,7 @@ ERROR 1105 (HY000): errCode = 2, detailMessage = Memory limit exceeded:<consumin
 同时可以在 log/be.INFO 中找到如下日志,确认当前查询内存使用是否符合预期,日志同样分为三部分:
 1、`Process Memory Summary`:进程内存统计。
 2、`Alloc Stacktrace`:触发内存超限检测的栈,这不一定是大内存申请的位置。
-3、`Memory Tracker Summary`:当前查询的 memory tracker 统计,可以看到查询每个算子当前使用的内存和峰值,具体可参考 [Memory Tracker](../admin-manual/memory-management/memory-tracker.md)。
+3、`Memory Tracker Summary`:当前查询的 memory tracker 统计,可以看到查询每个算子当前使用的内存和峰值,具体可参考 [Memory Tracker](./memory-tracker)。
 注意:一个查询在内存超限后只会打印一次日志,此时查询的多个线程都会感知,并尝试等待内存释放,或者cancel当前查询,如果日志中 Try Alloc 的值很小,则无须关注`Alloc Stacktrace`,直接分析`Memory Tracker Summary`即可。
 
 ```
diff --git a/docs/zh-CN/docs/advanced/alter-table/replace-table.md b/docs/zh-CN/docs/advanced/alter-table/replace-table.md
index 043ab9e6fc..fa28de9189 100644
--- a/docs/zh-CN/docs/advanced/alter-table/replace-table.md
+++ b/docs/zh-CN/docs/advanced/alter-table/replace-table.md
@@ -28,7 +28,7 @@ under the License.
 
 在 0.14 版本中,Doris 支持对两个表进行原子的替换操作。 该操作仅适用于 OLAP 表。
 
-分区级别的替换操作,请参阅 [临时分区文档](../../partition/table-temp-partition)
+分区级别的替换操作,请参阅 [临时分区文档](../partition/table-temp-partition)
 
 ## 语法说明
 
diff --git a/docs/zh-CN/docs/advanced/alter-table/schema-change.md b/docs/zh-CN/docs/advanced/alter-table/schema-change.md
index e3bf908008..af820835d6 100644
--- a/docs/zh-CN/docs/advanced/alter-table/schema-change.md
+++ b/docs/zh-CN/docs/advanced/alter-table/schema-change.md
@@ -71,7 +71,7 @@ under the License.
 
 ## 创建作业
 
-创建 Schema Change 的具体语法可以查看帮助 [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN.md) 中 Schema Change 部分的说明。
+创建 Schema Change 的具体语法可以查看帮助 [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN) 中 Schema Change 部分的说明。
 
 Schema Change 的创建是一个异步过程,作业提交成功后,用户需要通过 `SHOW ALTER TABLE COLUMN` 命令来查看作业进度。
 
@@ -291,4 +291,4 @@ SHOW ALTER TABLE COLUMN\G;
 
 ## 更多帮助
 
-关于Schema Change使用的更多详细语法及最佳实践,请参阅 [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN.md) 命令手册,你也可以在 MySql 客户端命令行下输入 `HELP ALTER TABLE COLUMN`  获取更多帮助信息。
+关于Schema Change使用的更多详细语法及最佳实践,请参阅 [ALTER TABLE COLUMN](../../sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-COLUMN) 命令手册,你也可以在 MySql 客户端命令行下输入 `HELP ALTER TABLE COLUMN`  获取更多帮助信息。
diff --git a/docs/zh-CN/docs/data-operate/import/import-scenes/external-storage-load.md b/docs/zh-CN/docs/data-operate/import/import-scenes/external-storage-load.md
index 990c841a77..ac29d0e93c 100644
--- a/docs/zh-CN/docs/data-operate/import/import-scenes/external-storage-load.md
+++ b/docs/zh-CN/docs/data-operate/import/import-scenes/external-storage-load.md
@@ -49,7 +49,7 @@ Hdfs load 创建导入语句,导入方式和[Broker Load](../../../data-operat
 
 1. 创建一张表
 
-   通过 `CREATE TABLE` 命令在`demo`创建一张表用于存储待导入的数据。具体的导入方式请查阅 [CREATE TABLE](../../../sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-TABLE.md) 命令手册。示例如下:
+   通过 `CREATE TABLE` 命令在`demo`创建一张表用于存储待导入的数据。具体的导入方式请查阅 [CREATE TABLE](../../../sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-TABLE) 命令手册。示例如下:
 
    ```sql
    CREATE TABLE IF NOT EXISTS load_hdfs_file_test
@@ -86,7 +86,7 @@ Hdfs load 创建导入语句,导入方式和[Broker Load](../../../data-operat
   
 3. 查看导入状态
    
-   Broker load 是一个异步的导入方式,具体导入结果可以通过[SHOW LOAD](../../../../sql-manual/sql-reference/Show-Statements/SHOW-LOAD)命令查看
+   Broker load 是一个异步的导入方式,具体导入结果可以通过[SHOW LOAD](../../../sql-manual/sql-reference/Show-Statements/SHOW-LOAD)命令查看
    
    ```
    mysql> show load order by createtime desc limit 1\G;
diff --git a/docs/zh-CN/docs/data-operate/import/import-scenes/jdbc-load.md b/docs/zh-CN/docs/data-operate/import/import-scenes/jdbc-load.md
index ada0d26ddc..36725343f0 100644
--- a/docs/zh-CN/docs/data-operate/import/import-scenes/jdbc-load.md
+++ b/docs/zh-CN/docs/data-operate/import/import-scenes/jdbc-load.md
@@ -160,4 +160,4 @@ public class DorisJDBCDemo {
 
    前面提到,我们建议在使用 INSERT 导入数据时,采用 ”批“ 的方式进行导入,而不是单条插入。
 
-   同时,我们可以为每次 INSERT 操作设置一个 Label。通过 [Label 机制](../load-atomicity) 可以保证操作的幂等性和原子性,最终做到数据的不丢不重。关于 INSERT 中 Label 的具体用法,可以参阅 [INSERT](../../../sql-manual/sql-reference/Data-Manipulation-Statements/Manipulation/INSERT) 文档。
+   同时,我们可以为每次 INSERT 操作设置一个 Label。通过 [Label 机制](./load-atomicity) 可以保证操作的幂等性和原子性,最终做到数据的不丢不重。关于 INSERT 中 Label 的具体用法,可以参阅 [INSERT](../../../sql-manual/sql-reference/Data-Manipulation-Statements/Manipulation/INSERT) 文档。
diff --git a/docs/zh-CN/docs/data-table/basic-usage.md b/docs/zh-CN/docs/data-table/basic-usage.md
index e46c72e688..2098d4960d 100644
--- a/docs/zh-CN/docs/data-table/basic-usage.md
+++ b/docs/zh-CN/docs/data-table/basic-usage.md
@@ -165,7 +165,7 @@ mysql> USE example_db;
 Database changed
 ```
 
-Doris支持[复合分区和单分区](../data-partition)两种建表方式。下面以聚合模型为例,分别演示如何创建两种分区的数据表。
+Doris支持[复合分区和单分区](./data-partition)两种建表方式。下面以聚合模型为例,分别演示如何创建两种分区的数据表。
 
 #### 单分区
 
@@ -701,7 +701,7 @@ mysql> select sum(table1.pv) from table1 join [shuffle] table2 where table1.site
 
 当部署多个 FE 节点时,用户可以在多个 FE 之上部署负载均衡层来实现 Doris 的高可用。
 
-具体安装部署及使用方式请参照 [负载均衡](../../admin-manual/cluster-management/load-balancing)
+具体安装部署及使用方式请参照 [负载均衡](../admin-manual/cluster-management/load-balancing)
 
 ## 数据更新和删除
 
diff --git a/docs/zh-CN/docs/data-table/hit-the-rollup.md b/docs/zh-CN/docs/data-table/hit-the-rollup.md
index 42e21bc7dd..6baf8ac2c7 100644
--- a/docs/zh-CN/docs/data-table/hit-the-rollup.md
+++ b/docs/zh-CN/docs/data-table/hit-the-rollup.md
@@ -44,7 +44,7 @@ ROLLUP 表的基本作用,在于在 Base 表的基础上,获得更粗粒度
 
 1. 示例1:获得每个用户的总消费
 
-接 **[数据模型Aggregate 模型](../data-model)**小节的**示例2**,Base 表结构如下:
+接 **[数据模型Aggregate 模型](./data-model)**小节的**示例2**,Base 表结构如下:
 
 | ColumnName      | Type        | AggregationType | Comment                |
 | --------------- | ----------- | --------------- | ---------------------- |
diff --git a/docs/zh-CN/docs/ecosystem/external-table/hive-bitmap-udf.md b/docs/zh-CN/docs/ecosystem/external-table/hive-bitmap-udf.md
index 06cfd5a890..d2b7aae243 100644
--- a/docs/zh-CN/docs/ecosystem/external-table/hive-bitmap-udf.md
+++ b/docs/zh-CN/docs/ecosystem/external-table/hive-bitmap-udf.md
@@ -115,4 +115,4 @@ select k1,bitmap_union(uuid) from hive_bitmap_table group by k1
 
 ## Hive bitmap 导入 doris
 
- 详见: [Spark Load](../../../data-operate/import/import-way/spark-load-manual) -> 基本操作  -> 创建导入 (示例3:上游数据源是hive binary类型情况)
+ 详见: [Spark Load](../../data-operate/import/import-way/spark-load-manual) -> 基本操作  -> 创建导入 (示例3:上游数据源是hive binary类型情况)
diff --git a/docs/zh-CN/docs/ecosystem/external-table/multi-catalog.md b/docs/zh-CN/docs/ecosystem/external-table/multi-catalog.md
index b66159b978..cc670eefe1 100644
--- a/docs/zh-CN/docs/ecosystem/external-table/multi-catalog.md
+++ b/docs/zh-CN/docs/ecosystem/external-table/multi-catalog.md
@@ -67,7 +67,7 @@ under the License.
     
 4. 删除 Catalog
 
-    External Catalog 中的 Database 和 Table 都是只读的。但是可以删除 Catalog(Internal Catalog无法删除)。可以通过 [DROP CATALOG](../../../sql-manual/sql-reference/Data-Definition-Statements/Drop/DROP-CATALOG) 命令删除一个 External Catalog。
+    External Catalog 中的 Database 和 Table 都是只读的。但是可以删除 Catalog(Internal Catalog无法删除)。可以通过 [DROP CATALOG](../../sql-manual/sql-reference/Data-Definition-Statements/Drop/DROP-CATALOG) 命令删除一个 External Catalog。
     
     该操作仅会删除 Doris 中该 Catalog 的映射信息,并不会修改或变更任何外部数据目录的内容。
 
diff --git a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
index 1f1c4085e2..67ce90fe86 100644
--- a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
+++ b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-PARTITION.md
@@ -62,7 +62,7 @@ partition_desc ["key"="value"]
 - 分区为左闭右开区间,如果用户仅指定右边界,系统会自动确定左边界
 - 如果没有指定分桶方式,则自动使用建表使用的分桶方式和分桶数。
 - 如指定分桶方式,只能修改分桶数,不可修改分桶方式或分桶列。如果指定了分桶方式,但是没有指定分桶数,则分桶数会使用默认值10,不会使用建表时指定的分桶数。如果要指定分桶数,则必须指定分桶方式。
-- ["key"="value"] 部分可以设置分区的一些属性,具体说明见 [CREATE TABLE](../../Create/CREATE-TABLE)
+- ["key"="value"] 部分可以设置分区的一些属性,具体说明见 [CREATE TABLE](../Create/CREATE-TABLE)
 - 如果建表时用户未显式创建Partition,则不支持通过ALTER的方式增加分区
 
 2. 删除分区
diff --git a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
index b755c39bc8..cd72797b57 100644
--- a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
+++ b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Definition-Statements/Alter/ALTER-TABLE-ROLLUP.md
@@ -68,7 +68,7 @@ ADD ROLLUP [rollup_name (column_name1, column_name2, ...)
 
 - 如果没有指定 from_index_name,则默认从 base index 创建
 - rollup 表中的列必须是 from_index 中已有的列
-- 在 properties 中,可以指定存储格式。具体请参阅 [CREATE TABLE](../../Create/CREATE-TABLE)
+- 在 properties 中,可以指定存储格式。具体请参阅 [CREATE TABLE](../Create/CREATE-TABLE)
 
 3. 删除 rollup index
 
diff --git a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
index 5dadc7f0e7..f0974a161e 100644
--- a/docs/zh-CN/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
+++ b/docs/zh-CN/docs/sql-manual/sql-reference/Data-Manipulation-Statements/Load/STREAM-LOAD.md
@@ -405,11 +405,11 @@ curl --location-trusted -u root -H "columns: k1,k2,source_sequence,v1,v2" -H "fu
 
 4. Label、导入事务、多表原子性
 
-   Doris 中所有导入任务都是原子生效的。并且在同一个导入任务中对多张表的导入也能够保证原子性。同时,Doris 还可以通过 Label 的机制来保证数据导入的不丢不重。具体说明可以参阅 [导入事务和原子性](../../../../data-operate/import/import-scenes/load-atomicity.md) 文档。
+   Doris 中所有导入任务都是原子生效的。并且在同一个导入任务中对多张表的导入也能够保证原子性。同时,Doris 还可以通过 Label 的机制来保证数据导入的不丢不重。具体说明可以参阅 [导入事务和原子性](../../../../data-operate/import/import-scenes/load-atomicity) 文档。
 
 5. 列映射、衍生列和过滤
 
-   Doris 可以在导入语句中支持非常丰富的列转换和过滤操作。支持绝大多数内置函数和 UDF。关于如何正确的使用这个功能,可参阅 [列的映射,转换与过滤](../../../../data-operate/import/import-scenes/load-data-convert.md) 文档。
+   Doris 可以在导入语句中支持非常丰富的列转换和过滤操作。支持绝大多数内置函数和 UDF。关于如何正确的使用这个功能,可参阅 [列的映射,转换与过滤](../../../../data-operate/import/import-scenes/load-data-convert) 文档。
 
 6. 错误数据过滤
 
diff --git a/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md b/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
index fb4171d69f..ba1f659052 100644
--- a/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
+++ b/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-ALTER-TABLE-MATERIALIZED-VIEW.md
@@ -32,7 +32,7 @@ SHOW ALTER TABLE MATERIALIZED VIEW
 
 ### Description
 
-该命令用于查看通过 [CREATE-MATERIALIZED-VIEW](../../../sql-manual/sql-reference/Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) 语句提交的创建物化视图作业的执行情况。
+该命令用于查看通过 [CREATE-MATERIALIZED-VIEW](../../sql-reference/Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW) 语句提交的创建物化视图作业的执行情况。
 
 > 该语句等同于 `SHOW ALTER TABLE ROLLUP`;
 
diff --git a/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md b/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
index 5ab40e24c1..a34345f068 100644
--- a/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
+++ b/docs/zh-CN/docs/sql-manual/sql-reference/Show-Statements/SHOW-STATUS.md
@@ -32,7 +32,7 @@ SHOW STATUS
 
 ### Description
 
-该命令用于查看通过[创建物化视图](../../Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW)语句提交的创建物化视图作业的执行情况。
+该命令用于查看通过[创建物化视图](../Data-Definition-Statements/Create/CREATE-MATERIALIZED-VIEW)语句提交的创建物化视图作业的执行情况。
 
 > 该语句相当于`SHOW ALTER TABLE ROLLUP`;
 


---------------------------------------------------------------------
To unsubscribe, e-mail: commits-unsubscribe@doris.apache.org
For additional commands, e-mail: commits-help@doris.apache.org