You are viewing a plain text version of this content. The canonical link for it is here.
Posted to reviews@yunikorn.apache.org by yu...@apache.org on 2022/03/25 07:48:21 UTC

[incubator-yunikorn-site] branch master updated: [YUNIKORN-1151] Fix the yaml comment syntax in Performance/Benchmarking Tutorial (#142)

This is an automated email from the ASF dual-hosted git repository.

yuteng pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/incubator-yunikorn-site.git


The following commit(s) were added to refs/heads/master by this push:
     new 1fd5ef8  [YUNIKORN-1151] Fix the yaml comment syntax in Performance/Benchmarking Tutorial (#142)
1fd5ef8 is described below

commit 1fd5ef823a98a73aabd4ed098b2fc2db479586c5
Author: 9501sam <44...@users.noreply.github.com>
AuthorDate: Fri Mar 25 15:48:17 2022 +0800

    [YUNIKORN-1151] Fix the yaml comment syntax in Performance/Benchmarking Tutorial (#142)
    
    let comments in yaml file begin with ```#```
    
    Co-authored-by: lab <la...@debian-BULLSEYE-live-builder-AMD64>
---
 docs/performance/performance_tutorial.md               | 16 ++++++++--------
 .../current/performance/performance_tutorial.md        | 18 +++++++++---------
 .../version-0.12.1/performance/performance_tutorial.md | 18 +++++++++---------
 .../version-0.12.2/performance/performance_tutorial.md | 18 +++++++++---------
 .../version-0.12.1/performance/performance_tutorial.md | 18 +++++++++---------
 .../version-0.12.2/performance/performance_tutorial.md | 18 +++++++++---------
 6 files changed, 53 insertions(+), 53 deletions(-)

diff --git a/docs/performance/performance_tutorial.md b/docs/performance/performance_tutorial.md
index 4c06c0a..e2aa9f2 100644
--- a/docs/performance/performance_tutorial.md
+++ b/docs/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
diff --git a/i18n/zh-cn/docusaurus-plugin-content-docs/current/performance/performance_tutorial.md b/i18n/zh-cn/docusaurus-plugin-content-docs/current/performance/performance_tutorial.md
index 2d34025..ac305e7 100644
--- a/i18n/zh-cn/docusaurus-plugin-content-docs/current/performance/performance_tutorial.md
+++ b/i18n/zh-cn/docusaurus-plugin-content-docs/current/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
@@ -449,4 +449,4 @@ systemctl restart kubelet
 
 ## Summary
 
-With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
\ No newline at end of file
+With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
diff --git a/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.1/performance/performance_tutorial.md b/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.1/performance/performance_tutorial.md
index 8dfa57c..7d10def 100644
--- a/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.1/performance/performance_tutorial.md
+++ b/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.1/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
@@ -449,4 +449,4 @@ systemctl restart kubelet
 
 ## Summary
 
-With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
\ No newline at end of file
+With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
diff --git a/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.2/performance/performance_tutorial.md b/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.2/performance/performance_tutorial.md
index 8dfa57c..7d10def 100644
--- a/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.2/performance/performance_tutorial.md
+++ b/i18n/zh-cn/docusaurus-plugin-content-docs/version-0.12.2/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
@@ -449,4 +449,4 @@ systemctl restart kubelet
 
 ## Summary
 
-With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
\ No newline at end of file
+With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
diff --git a/versioned_docs/version-0.12.1/performance/performance_tutorial.md b/versioned_docs/version-0.12.1/performance/performance_tutorial.md
index 2d34025..ac305e7 100644
--- a/versioned_docs/version-0.12.1/performance/performance_tutorial.md
+++ b/versioned_docs/version-0.12.1/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
@@ -449,4 +449,4 @@ systemctl restart kubelet
 
 ## Summary
 
-With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
\ No newline at end of file
+With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
diff --git a/versioned_docs/version-0.12.2/performance/performance_tutorial.md b/versioned_docs/version-0.12.2/performance/performance_tutorial.md
index 2d34025..ac305e7 100644
--- a/versioned_docs/version-0.12.2/performance/performance_tutorial.md
+++ b/versioned_docs/version-0.12.2/performance/performance_tutorial.md
@@ -157,7 +157,7 @@ metadata:
   name: hollow-node
   namespace: kubemark
 spec:
-  replicas: 2000  // the node number you want to simulate
+  replicas: 2000  # the node number you want to simulate
   selector:
       name: hollow-node
   template:
@@ -165,13 +165,13 @@ spec:
       labels:
         name: hollow-node
     spec:
-      nodeSelector:  // leverage label to allocate to native node
+      nodeSelector:  # leverage label to allocate to native node
         tag: tagName  
       initContainers:
       - name: init-inotify-limit
         image: docker.io/busybox:latest
         imagePullPolicy: IfNotPresent
-        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] // set as same as max_user_instance in actual node 
+        command: ['sysctl', '-w', 'fs.inotify.max_user_instances=200'] # set as same as max_user_instance in actual node 
         securityContext:
           privileged: true
       volumes:
@@ -183,7 +183,7 @@ spec:
           path: /var/log
       containers:
       - name: hollow-kubelet
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         ports:
         - containerPort: 4194
@@ -209,13 +209,13 @@ spec:
         - name: logs-volume
           mountPath: /var/log
         resources:
-          requests:    // the resource of hollow pod, can modify it.
+          requests:    # the resource of hollow pod, can modify it.
             cpu: 20m
             memory: 50M
         securityContext:
           privileged: true
       - name: hollow-proxy
-        image: 0yukali0/kubemark:1.20.10 // the kubemark image you build 
+        image: 0yukali0/kubemark:1.20.10 # the kubemark image you build 
         imagePullPolicy: IfNotPresent
         env:
         - name: NODE_NAME
@@ -237,7 +237,7 @@ spec:
           readOnly: true
         - name: logs-volume
           mountPath: /var/log
-        resources:  // the resource of hollow pod, can modify it.
+        resources:  # the resource of hollow pod, can modify it.
           requests:
             cpu: 20m
             memory: 50M
@@ -342,7 +342,7 @@ scrape_configs:
     metrics_path: '/ws/v1/metrics'
     static_configs:
     - targets: ['docker.for.mac.host.internal:9080'] 
-    // 9080 is internal port, need port forward or modify 9080 to service's port
+    # 9080 is internal port, need port forward or modify 9080 to service's port
 ```
 
 ### 3. Launch Prometheus
@@ -449,4 +449,4 @@ systemctl restart kubelet
 
 ## Summary
 
-With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.
\ No newline at end of file
+With Kubemark and Prometheus, we can easily run benchmark testing, collect YuniKorn metrics and analyze the performance. This helps us to identify the performance bottleneck in the scheduler and further eliminate them. The YuniKorn community will continue to improve these tools in the future, and continue to gain more performance improvements.