You are viewing a plain text version of this content. The canonical link for it is here.
Posted to notifications@linkis.apache.org by "liygit (via GitHub)" <gi...@apache.org> on 2023/02/08 10:13:19 UTC

[GitHub] [linkis] liygit opened a new issue, #4161: [Question] linkis submit python task to yarn failed

liygit opened a new issue, #4161:
URL: https://github.com/apache/linkis/issues/4161

   ### Before asking
   
   - [X] I have read the [How to ask for help](https://linkis.apache.org/community/how-to-ask-for-help) (我已阅读[如何寻求帮助](https://linkis.apache.org/zh-CN/community/how-to-ask-for-help)指引文档).
   
   
   ### Your environment
   
   <!-- Please describe the linkis version you are using and basic environment information -->
   <!-- 请描述您使用的linkis版本和基本环境信息 -->
   - Linkis version used: 1.3.1
   - Environment name and version:
       - hive-2.3.3
       - spark-2.4.3
       - scala-2.11.12
       - jdk 1.8.0_65
       - python(anaconda3)-3.7.4
   
   
   ### Describe your questions
   
   <!-- Please describe your questions -->
   <!-- 请描述您的问题 -->
   Q1.  the python task can not be submitted to yarn,there is no error messages
   
   
   ### Eureka service list
   
   <!-- You can directly copy and paste the screenshot of the list of registered service instances displayed on the Eureka page here -->
   <!--  可以将Eureka页面显示的注册服务实例列表的的截图 直接复制并粘贴到这里。 粘贴后,会自动生成如下格式的url-->
   
   ![image](https://user-images.githubusercontent.com/35162447/217499355-7de9a423-f362-440b-b30e-9d3dc03985dc.png)
   
   
   ### Some logs info or acctch file
   
   <!-- Please add the log in the following format, the key error log must be pasted in the code style, and the text as much as possible, the detailed log or other files can be dragged and uploaded as an attachment -->
   <!-- 请按如下格式 添加日志,关键错误日志请贴到[````log][````]之间,尽可能贴文字,详细日志或则其它文件,可以拖拽上传作为附件 -->
   
   linkis-cg-entrance.log:
   ````log
   2023-02-03 11:17:34.286 [INFO ] [qtp1843167615-26                        ] o.a.l.e.r.EntranceRestfulApi (98) [execute] - Begin to get an execID
   2023-02-03 11:17:34.351 [INFO ] [qtp1843167615-26                        ] o.a.l.e.s.DefaultEntranceServer (79) [execute] - received a request,convert JobRequest{id=749, reqId='null', submitUser='hadoop', executeUser='hadoop', labels=[[key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2], [key: codeType, value: {"codeType":"python"}, str: python], [key: userCreator, value: {"creator":"IDE","user":"hadoop"}, str: hadoop-IDE]], params={variable={}, configuration={}}, status=Inited}
   2023-02-03 11:17:34.582 [INFO ] [qtp1843167615-26                        ] o.a.l.s.u.FileSystemUtils$ (80) [createNewFileWithFileSystem] - doesn't need to call setOwner
   2023-02-03 11:17:34.583 [INFO ] [qtp1843167615-26                        ] o.a.l.e.l.CacheLogManager (99) [createLogWriter] - job 749 create cacheLogWriter
   2023-02-03 11:17:34.585 [INFO ] [IDE_hadoop_pythonConsumerThread         ] o.a.l.e.j.EntranceExecutionJob (97) [transition] - jobGroupId:749 execID:IDE_hadoop_python_1 change status Inited => Scheduled.
   2023-02-03 11:17:34.585 [INFO ] [qtp1843167615-26                        ] o.a.l.e.s.DefaultEntranceServer (156) [apply] - 2023-02-03 11:17:34.017 INFO Job with jobId : 749 and execID : IDE_hadoop_python_1 submitted 
   2023-02-03 11:17:34.586 [INFO ] [qtp1843167615-26                        ] o.a.l.s.u.ModuleUserUtils (63) [getProxyUserEntity] - user hadoop proxy to null operation execute task,id: 749
   2023-02-03 11:17:34.589 [INFO ] [qtp1843167615-26                        ] o.a.l.e.r.EntranceRestfulApi (141) [execute] - End to get an an execID: exec_id018015linkis-cg-entrancedss-master:9104IDE_hadoop_python_1, taskID: 749
   2023-02-03 11:17:34.729 [INFO ] [EntranceJobScheduler-ThreadPool-3       ] o.a.l.e.j.EntranceExecutionJob (97) [transition] - jobGroupId:749 execID:IDE_hadoop_python_1 change status Scheduled => Running.
   2023-02-03 11:17:34.764 [INFO ] [EntranceJobScheduler-ThreadPool-3       ] o.a.l.o.c.i.SessionStateImpl$$anon$4 (51) [convert] - Finished to convert JobReq(749) to AstJob(TaskID_749_otJobId_astJob_1).
   2023-02-03 11:17:34.766 [INFO ] [EntranceJobScheduler-ThreadPool-3       ] o.a.l.o.s.a.AsyncTaskManager (88) [putExecTask] - submit execTask TaskID_749_otJobId_astJob_1_job_2 to taskManager get executionTask execution_1
   2023-02-03 11:17:34.766 [INFO ] [EntranceJobScheduler-ThreadPool-3       ] o.a.l.e.e.DefaultEntranceExecutor (271) [apply] - JobRequest (749) was submitted to Orchestrator.
   2023-02-03 11:17:34.767 [INFO ] [CodeReheaterNotifyTaskConsumer          ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
   2023-02-03 11:17:34.768 [INFO ] [EntranceJobScheduler-ThreadPool-3       ] o.a.l.e.e.DefaultEntranceExecutor (300) [apply] - For job 749 and orchestrator task id 749 to create EngineExecuteAsyncReturn
   2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_job_3) to running
   2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (77) [run] - Succeed to execute ExecTask(TaskID_749_otJobId_astJob_1_job_3)
   2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_job_3 change status Inited => Succeed.
   2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5              ] o.a.l.o.s.a.AsyncTaskManager (249) [addCompletedTask] - TaskID_749_otJobId_astJob_1_job_3 task completed, now remove from taskManager
   2023-02-03 11:17:34.770 [INFO ] [BaseTaskScheduler-Thread-5              ] o.a.l.o.s.a.AsyncTaskManager (283) [apply] - executionTask(execution_1) no task running, user key hadoop-IDE,python-python2, minusNumber: 0
   2023-02-03 11:17:34.772 [INFO ] [CodeReheaterNotifyTaskConsumer          ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
   2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_stage_3) to running
   2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (77) [run] - Succeed to execute ExecTask(TaskID_749_otJobId_astJob_1_stage_3)
   2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_stage_3 change status Inited => Succeed.
   2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6              ] o.a.l.o.s.a.AsyncTaskManager (249) [addCompletedTask] - TaskID_749_otJobId_astJob_1_stage_3 task completed, now remove from taskManager
   2023-02-03 11:17:34.774 [INFO ] [BaseTaskScheduler-Thread-6              ] o.a.l.o.s.a.AsyncTaskManager (283) [apply] - executionTask(execution_1) no task running, user key hadoop-IDE,python-python2, minusNumber: 0
   2023-02-03 11:17:34.775 [INFO ] [CodeReheaterNotifyTaskConsumer          ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
   2023-02-03 11:17:34.776 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_codeExec_1) to running
   2023-02-03 11:17:34.778 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.c.p.CodeLogicalUnitExecTask (84) [execute] - Start to execute CodeLogicalUnitExecTask(TaskID_749_otJobId_astJob_1_codeExec_1).
   2023-02-03 11:17:34.778 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.e.ComputationEngineConnManager (194) [addMark] - add mark mark_1
   2023-02-03 11:17:34.779 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (121) [createExecutor] - create Executor for execId TaskID_749_otJobId_astJob_1_codeExec_1 mark id is mark_1, user hadoop
   2023-02-03 11:17:34.779 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.e.ComputationEngineConnManager (125) [getAvailableEngineConnExecutor] - mark mark_1 start to  getAvailableEngineConnExecutor
   2023-02-03 11:17:34.916 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (138) [createExecutor] - Finished to create Executor for execId TaskID_749_otJobId_astJob_1_codeExec_1 mark id is mark_1, user hadoop
   2023-02-03 11:17:34.917 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (97) [askExecutor] - Finished to askExecutor for execId TaskID_749_otJobId_astJob_1_codeExec_1
   2023-02-03 11:17:34.940 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.e.s.i.ComputationEngineConnExecutor (83) [execute] - Succeed to submit task TaskID_749_otJobId_astJob_1_codeExec_1 to engineConn(ServiceInstance(linkis-cg-engineconn, dss-master:41636)), Get asyncResponse execID is SubmitResponse(2)
   2023-02-03 11:17:34.940 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (257) [addEngineConnTaskID] - To add codeExecTaskExecutor  engineConn ServiceInstance(linkis-cg-engineconn, dss-master:41636) execTask TaskID_749_otJobId_astJob_1_codeExec_1 mark mark_1 engineConnTaskId 2 to instanceToExecutors
   2023-02-03 11:17:34.943 [INFO ] [BaseTaskScheduler-Thread-7              ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_codeExec_1 change status Inited => Running.
   2023-02-03 11:17:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-4       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 4
   2023-02-03 11:18:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-12      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:19:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-8       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:20:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-12      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:21:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-2       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:22:17.764 [INFO ] [AsyncResolver-bootstrap-executor-0      ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
   2023-02-03 11:22:39.834 [INFO ] [Orchestrator-Listener-Asyn-Thread-Thread-0] o.a.l.e.l.HDFSCacheLogWriter (115) [writeToFile] - hdfs:///wedata/linkis/log/2023-02-03/IDE/hadoop/749.log write first one line log
   2023-02-03 11:22:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-6       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:23:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-15      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:24:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-16      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:25:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-8       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:26:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-18      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:27:17.765 [INFO ] [AsyncResolver-bootstrap-executor-0      ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
   2023-02-03 11:27:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-4       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:28:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-15      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:29:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-10      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:30:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-5       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:31:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-3       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:32:17.765 [INFO ] [AsyncResolver-bootstrap-executor-0      ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
   2023-02-03 11:32:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-6       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:33:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-7       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:34:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-2       ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:35:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-19      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:36:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-17      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:37:17.766 [INFO ] [AsyncResolver-bootstrap-executor-0      ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
   2023-02-03 11:37:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-19      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   2023-02-03 11:38:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-17      ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from  taskFutureCache size 0
   
   ````
   linkis-cg-linkismanager.log:
   ````log
   2023-02-03 17:59:51.520 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (100) [apply] - Task: 836 start to async(dss-master:9101_12) createEngine, dssmark_id: mark_88
   2023-02-03 17:59:51.520 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEngineCreateService (114) [createEngine] - Task: 836 start to create Engine for request: EngineCreateRequest{labels={userCreator=hadoop-LINKISCLI, codeType=python, engineType=python-python2, executeOnce=}, timeOut=660000, user='hadoop', createService='dssmark_id: mark_88', description='null'}.
   2023-02-03 17:59:51.547 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEngineCreateService (156) [createEngine] - Suitable ems size is 2
   2023-02-03 17:59:51.603 [INFO ] [AskEngineService-Thread-25              ] o.a.l.e.s.s.DefaultEngineConnResourceFactoryService (49) [createEngineResource] - To invoke createEngineResource TimeoutEngineResourceRequest(660000,hadoop,[[key: userCreator, value: {"creator":"LINKISCLI","user":"hadoop"}, str: hadoop-LINKISCLI], [key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2]],{jobId=836, wds.linkis.rm.yarnqueue.memory.max=300G, wds.linkis.rm.client.core.max=10, wds.linkis.rm.instance=10, wds.linkis.engineconn.java.driver.memory=1g, wds.linkis.rm.yarnqueue.instance.max=30, wds.linkis.rm.yarnqueue.cores.max=150, wds.linkis.rm.client.memory.max=20G, wds.linkis.rm.yarnqueue=default, python.version=python2, wds.linkis.engineconn.max.free.time=1h})
   2023-02-03 17:59:51.614 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.ResourceLockService (52) [apply$mcZ$sp] - successfully locked labelhadoop-LINKISCLI,python-python2
   2023-02-03 17:59:51.616 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.i.DefaultReqResourceService (62) [canRequest] - ResourceInit: hadoop-LINKISCLI,python-python2 
   2023-02-03 17:59:51.616 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.u.UserConfiguration$ (102) [apply] - hadoop on creator LINKISCLI available engine python resource:Number of instances(实例数):10,(RAM)内存:20.0 GB,cpu:10
   2023-02-03 17:59:51.643 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.ResourceLockService (52) [apply$mcZ$sp] - successfully locked labellinkis-cg-engineconnmanager-zdww4:9102
   2023-02-03 17:59:51.652 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.i.DefaultResourceManager (306) [apply] - ResourceChanged:linkis-cg-engineconnmanager-zdww4:9102 --> CommonNodeResource{resourceType=LoadInstance, maxResource=Number of instances(实例数):50,(RAM)内存:100.0 GB,cpu:100, minResource=Number of instances(实例数):2,(RAM)内存:4.0 GB,cpu:2, usedResource=Number of instances(实例数):2,(RAM)内存:2.0 GB,cpu:2, lockedResource=Number of instances(实例数):1,(RAM)内存:1024.0 MB,cpu:1, expectedResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, leftResource=Number of instances(实例数):47,(RAM)内存:97.0 GB,cpu:97}
   2023-02-03 17:59:51.660 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.i.DefaultResourceManager (306) [apply] - ResourceChanged:hadoop-LINKISCLI,python-python2 --> CommonNodeResource{resourceType=LoadInstance, maxResource=Number of instances(实例数):10,(RAM)内存:20.0 GB,cpu:10, minResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, usedResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, lockedResource=Number of instances(实例数):1,(RAM)内存:1024.0 MB,cpu:1, expectedResource=null, leftResource=Number of instances(实例数):9,(RAM)内存:19.0 GB,cpu:9}
   2023-02-03 17:59:51.664 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.ResourceLockService (64) [apply$mcV$sp] - unlocked hadoop-LINKISCLI,python-python2
   2023-02-03 17:59:51.669 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.r.s.ResourceLockService (64) [apply$mcV$sp] - unlocked linkis-cg-engineconnmanager-zdww4:9102
   2023-02-03 17:59:51.669 [WARN ] [AskEngineService-Thread-25              ] o.a.l.m.r.d.RMLabelContainer (156) [getEngineInstanceLabel] - EngineInstanceLabel not found
   2023-02-03 17:59:51.730 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEMEngineService (70) [createEngine] - EM ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102) start to create Engine EngineConnBuildRequestImpl(e84fc301-889c-4231-9c33-b46abebaafc5,[[key: userCreator, value: {"creator":"LINKISCLI","user":"hadoop"}, str: hadoop-LINKISCLI], [key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2]],org.apache.linkis.manager.engineplugin.common.resource.UserNodeResource@4c4ee70e,EngineConnCreationDescImpl(dssmark_id: mark_88,null,{jobId=836, wds.linkis.rm.yarnqueue.memory.max=300G, wds.linkis.rm.client.core.max=10, wds.linkis.rm.instance=10, wds.linkis.engineconn.java.driver.memory=1g, wds.linkis.rm.yarnqueue.instance.max=30, wds.linkis.rm.yarnqueue.cores.max=150, wds.linkis.rm.client.memory.max=20G, wds.linkis.rm.yarnqueue=default, python.version=python2, wds.linkis.engineconn.max.free.time=1h}))
   2023-02-03 17:59:51.730 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.p.DefaultEMNodPointer (40) [createEngine] - Start to createEngine ask em ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102)
   2023-02-03 17:59:51.751 [INFO ] [qtp2126981566-581                       ] o.a.l.m.e.p.l.PythonProcessEngineConnLaunchBuilder (101) [getEnvironment] - Setting up the launch environment for engineconn.
   2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.p.DefaultEMNodPointer (43) [createEngine] - Succeed to createEngine ask em ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102), engineNode AMEngineNode{nodeStatus=null, lock='null', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'} 
   2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEMEngineService (72) [createEngine] - EM ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102) Finished to create Engine e84fc301-889c-4231-9c33-b46abebaafc5
   2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEngineCreateService (204) [createEngine] - Task: 836 finished to create  engineConn AMEngineNode{nodeStatus=null,
    lock='null', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'}. ticketId is e84fc301-889c-4231-9c33-b46abebaafc5
    
   2023-02-03 17:59:51.828 [INFO ] [AskEngineService-Thread-25              ] o.a.l.m.a.s.e.DefaultEngineCreateService (365) [apply$mcV$sp] - Start to wait engineConn(AMEngineNode{nodeStatus=null, lock='null', se
   rviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'}) to be available, but only 11.0 m left.
   
   2023-02-03 17:59:55.743 [INFO ] [qtp2126981566-568                       ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (71) [askEngine] - received task: 837, engineAskRequest EngineAskRequest{labels={userCreat
   or=hadoop-IDE, codeType=sql, engineType=spark-2.4.3}, timeOut=660000, user='hadoop', createService='dssmark_id: mark_89', description='null'}
   2023-02-03 17:59:55.744 [INFO ] [qtp2126981566-568                       ] o.a.l.m.a.s.e.DefaultEngineReuseService (72) [reuseEngine] - Start to reuse Engine for request: EngineReuseRequest{timeOut=660000, reu
   seCount=0, user='hadoop'}
   2023-02-03 17:59:55.769 [INFO ] [qtp2126981566-568                       ] o.a.l.m.a.s.e.DefaultEngineReuseService (152) [org$apache$linkis$manager$am$service$engine$DefaultEngineReuseService$$selectEngineToRe
   use$1] - prepare to reuse engineNode: ServiceInstance(linkis-cg-engineconn, zdww4:42094)
   2023-02-03 17:59:55.814 [INFO ] [qtp2126981566-568                       ] o.a.l.m.a.s.e.DefaultEngineReuseService (185) [reuseEngine] - Finished to reuse Engine for request: EngineReuseRequest{timeOut=660000,
    reuseCount=0, user='hadoop'} get EngineNode AMEngineNode{nodeStatus=Unlock, lock='java.util.concurrent.Semaphore@30e82d09[Permits = 0]', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:42094), own
   er='null'}, time taken 44
   2023-02-03 17:59:55.815 [INFO ] [qtp2126981566-568                       ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (91) [askEngine] - Finished to ask engine for task: 837 user hadoop by reuse node AMEngine
   Node{nodeStatus=Unlock, lock='java.util.concurrent.Semaphore@30e82d09[Permits = 0]', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:42094), owner='null'}
   2023-02-03 17:59:55.819 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Idle, serviceInstance=Se
   rviceInstance(linkis-cg-engineconn, zdww4:42094)}
   2023-02-03 17:59:55.844 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Busy, serviceInstance=Se
   rviceInstance(linkis-cg-engineconn, zdww4:42094)}
   2023-02-03 18:00:00.837 [INFO ] [Linkis-Default-Scheduler-Thread-1       ] o.a.l.m.a.s.m.NodeHeartbeatMonitor (99) [apply$mcV$sp] - Start to check the health of the node
   2023-02-03 18:00:00.841 [INFO ] [Linkis-Default-Scheduler-Thread-1       ] o.a.l.m.a.s.m.NodeHeartbeatMonitor (130) [apply$mcV$sp] - Finished to check the health of the node
   2023-02-03 18:00:01.487 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.l.DefaultEngineNodeLocker (50) [releaseLock] - clientServiceInstance(linkis-cg-entrance, dss-master:9104) Start to unlock en
   gine ServiceInstance(linkis-cg-engineconn, zdww4:42094)
   2023-02-03 18:00:01.519 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-1       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Idle, serviceInstance=Se
   rviceInstance(linkis-cg-engineconn, zdww4:42094)}
   2023-02-03 18:00:01.531 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.l.DefaultEngineNodeLocker (56) [releaseLock] - clientServiceInstance(linkis-cg-entrance, dss-master:9104) Finished to unlock
    engine ServiceInstance(linkis-cg-engineconn, zdww4:42094)
   2023-02-03 18:00:01.537 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Unlock, serviceInstance=
   ServiceInstance(linkis-cg-engineconn, zdww4:42094)}
   2023-02-03 18:00:01.614 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
   =ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102)}
   2023-02-03 18:00:03.385 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
   =ServiceInstance(linkis-cg-engineconnmanager, zdww3:9102)}
   2023-02-03 18:00:03.386 [WARN ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.p.i.DefaultNodeMetricManagerPersistence (89) [addOrupdateNodeMetrics] - The request of update node metrics was ignored, becaus
   e the node zdww3:9102 is not exist.
   2023-02-03 18:00:09.559 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0       ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
   =ServiceInstance(linkis-cg-engineconnmanager, dss-master:9102)}
   ````


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org


[GitHub] [linkis] casionone commented on issue #4161: [Question] linkis submit python task to yarn failed

Posted by "casionone (via GitHub)" <gi...@apache.org>.
casionone commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1445282198

   Without further feedback, I closed this issue. You can reopen if necessary


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org


[GitHub] [linkis] casionone closed issue #4161: [Question] linkis submit python task to yarn failed

Posted by "casionone (via GitHub)" <gi...@apache.org>.
casionone closed issue #4161: [Question] linkis submit python task to yarn failed
URL: https://github.com/apache/linkis/issues/4161


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org


[GitHub] [linkis] github-actions[bot] commented on issue #4161: [Question] linkis submit python task to yarn failed

Posted by "github-actions[bot] (via GitHub)" <gi...@apache.org>.
github-actions[bot] commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1422352439

   ## :blush:  Welcome to the Apache Linkis community!!
   We are glad that you are contributing by opening this issue.
   
   Please make sure to include all the relevant context.
   We will be here shortly.
   
   If you are interested in contributing to our website project, please let us know!
   You can check out our contributing guide on
    :point_right:  [How to Participate in Project Contribution](https://linkis.apache.org/community/how-to-contribute).
   
   
   ### Community
   
   |WeChat Assistant|WeChat Public Account|
   |-|-|
   |<img src="https://linkis.apache.org/Images/wedatasphere_contact_01.png" width="128"/>|<img src="https://linkis.apache.org/Images/gzh_01.png" width="128"/>|
   
   
   ### Mailing Lists
   |Name|Description|Subscribe|Unsubscribe|Archive|
   |:-----|:--------|:------|:-------|:-----|
   | [dev@linkis.apache.org](mailto:dev@linkis.apache.org) | community activity information | [subscribe](mailto:dev-subscribe@linkis.apache.org) | [unsubscribe](mailto:dev-unsubscribe@linkis.apache.org) | [archive](http://mail-archives.apache.org/mod_mbox/linkis-dev) |


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org


[GitHub] [linkis] QuantumXiecao commented on issue #4161: [Question] linkis submit python task to yarn failed

Posted by "QuantumXiecao (via GitHub)" <gi...@apache.org>.
QuantumXiecao commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1422595620

   @liygit  Could you please supply your submit code to help locating the problem?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org