You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@yunikorn.apache.org by "Yu-Lin Chen (Jira)" <ji...@apache.org> on 2024/01/06 17:34:00 UTC

[jira] [Created] (YUNIKORN-2313) Flaky E2E Test: "Verify_basic_preemption" try to request more resource than the node’s available resource

Yu-Lin Chen created YUNIKORN-2313:
-------------------------------------

             Summary: Flaky E2E Test:  "Verify_basic_preemption" try to request more resource than the node’s available resource
                 Key: YUNIKORN-2313
                 URL: https://issues.apache.org/jira/browse/YUNIKORN-2313
             Project: Apache YuniKorn
          Issue Type: Sub-task
          Components: test - e2e
            Reporter: Yu-Lin Chen
            Assignee: Yu-Lin Chen
         Attachments: 2_e2e-tests (v1.28.0) (Verify_basic_preemption).txt

The failed e2e test "Verify_basic_preemption" attempted to request more resources than the available resource on the K8S node. The root cause is that some pods running on the node are invisible to YuniKorn scheduler.

The e2e test mistakenly believed that it could acquire all the available resources record in YuniKorn.

[https://github.com/apache/yunikorn-k8shim/actions/runs/7420644533/job/20212796775?pr=759#step:6:1483]

 

*Logs:* (Verify_basic_preemption)

{*}K8S describe node result{*}: (node: yk8s-worker)
allocable resource: 14610305024 B
non-terminated pods's memory usage:
 * sleepjob1 - 4870M (33%)
 * sleepjob2 - 4870M (33%)
 * kindnet-tr9jc - {color:#de350b}50Mi{color}(0%)
 * kube-proxy-k58rn - 0 B
 * yunikorn-admission-controller-56c8c8b766-55kz8 - {color:#de350b}500Mi{color} (3%)

→ Total 70%

{*}YuniKorn Node REST API response{*}: (node: yk8s-worker)
 * memory capacity: 14610305024 B
 * memory allocated: 9740000000 B (4870M + 4870M) (66%)
 * memory available: 4870305024 B

Node only have 30% resource remaining, so the sleepjob3(4870M, 33%) never entered Running state. 

YuniKorn is not aware of 'kindnet-tr9jc' and 'yunikorn-admission-controller-56c8c8b766-55kz8' running on the node. The “YuniKorn node available resource” is not equal to “K8S node available resource”.

Remove dynamic limit for the sleepJob and set a fixed pod/queue limit could solve the issue on the e2e test. But still need to investigate why those running pod are invisible to YuniKorn.
 



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscribe@yunikorn.apache.org
For additional commands, e-mail: issues-help@yunikorn.apache.org