You are viewing a plain text version of this content. The canonical link for it is here.
Posted to notifications@linkis.apache.org by "liygit (via GitHub)" <gi...@apache.org> on 2023/02/08 10:13:19 UTC
[GitHub] [linkis] liygit opened a new issue, #4161: [Question] linkis submit python task to yarn failed
liygit opened a new issue, #4161:
URL: https://github.com/apache/linkis/issues/4161
### Before asking
- [X] I have read the [How to ask for help](https://linkis.apache.org/community/how-to-ask-for-help) (我已阅读[如何寻求帮助](https://linkis.apache.org/zh-CN/community/how-to-ask-for-help)指引文档).
### Your environment
<!-- Please describe the linkis version you are using and basic environment information -->
<!-- 请描述您使用的linkis版本和基本环境信息 -->
- Linkis version used: 1.3.1
- Environment name and version:
- hive-2.3.3
- spark-2.4.3
- scala-2.11.12
- jdk 1.8.0_65
- python(anaconda3)-3.7.4
### Describe your questions
<!-- Please describe your questions -->
<!-- 请描述您的问题 -->
Q1. the python task can not be submitted to yarn,there is no error messages
### Eureka service list
<!-- You can directly copy and paste the screenshot of the list of registered service instances displayed on the Eureka page here -->
<!-- 可以将Eureka页面显示的注册服务实例列表的的截图 直接复制并粘贴到这里。 粘贴后,会自动生成如下格式的url-->
![image](https://user-images.githubusercontent.com/35162447/217499355-7de9a423-f362-440b-b30e-9d3dc03985dc.png)
### Some logs info or acctch file
<!-- Please add the log in the following format, the key error log must be pasted in the code style, and the text as much as possible, the detailed log or other files can be dragged and uploaded as an attachment -->
<!-- 请按如下格式 添加日志,关键错误日志请贴到[````log][````]之间,尽可能贴文字,详细日志或则其它文件,可以拖拽上传作为附件 -->
linkis-cg-entrance.log:
````log
2023-02-03 11:17:34.286 [INFO ] [qtp1843167615-26 ] o.a.l.e.r.EntranceRestfulApi (98) [execute] - Begin to get an execID
2023-02-03 11:17:34.351 [INFO ] [qtp1843167615-26 ] o.a.l.e.s.DefaultEntranceServer (79) [execute] - received a request,convert JobRequest{id=749, reqId='null', submitUser='hadoop', executeUser='hadoop', labels=[[key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2], [key: codeType, value: {"codeType":"python"}, str: python], [key: userCreator, value: {"creator":"IDE","user":"hadoop"}, str: hadoop-IDE]], params={variable={}, configuration={}}, status=Inited}
2023-02-03 11:17:34.582 [INFO ] [qtp1843167615-26 ] o.a.l.s.u.FileSystemUtils$ (80) [createNewFileWithFileSystem] - doesn't need to call setOwner
2023-02-03 11:17:34.583 [INFO ] [qtp1843167615-26 ] o.a.l.e.l.CacheLogManager (99) [createLogWriter] - job 749 create cacheLogWriter
2023-02-03 11:17:34.585 [INFO ] [IDE_hadoop_pythonConsumerThread ] o.a.l.e.j.EntranceExecutionJob (97) [transition] - jobGroupId:749 execID:IDE_hadoop_python_1 change status Inited => Scheduled.
2023-02-03 11:17:34.585 [INFO ] [qtp1843167615-26 ] o.a.l.e.s.DefaultEntranceServer (156) [apply] - 2023-02-03 11:17:34.017 INFO Job with jobId : 749 and execID : IDE_hadoop_python_1 submitted
2023-02-03 11:17:34.586 [INFO ] [qtp1843167615-26 ] o.a.l.s.u.ModuleUserUtils (63) [getProxyUserEntity] - user hadoop proxy to null operation execute task,id: 749
2023-02-03 11:17:34.589 [INFO ] [qtp1843167615-26 ] o.a.l.e.r.EntranceRestfulApi (141) [execute] - End to get an an execID: exec_id018015linkis-cg-entrancedss-master:9104IDE_hadoop_python_1, taskID: 749
2023-02-03 11:17:34.729 [INFO ] [EntranceJobScheduler-ThreadPool-3 ] o.a.l.e.j.EntranceExecutionJob (97) [transition] - jobGroupId:749 execID:IDE_hadoop_python_1 change status Scheduled => Running.
2023-02-03 11:17:34.764 [INFO ] [EntranceJobScheduler-ThreadPool-3 ] o.a.l.o.c.i.SessionStateImpl$$anon$4 (51) [convert] - Finished to convert JobReq(749) to AstJob(TaskID_749_otJobId_astJob_1).
2023-02-03 11:17:34.766 [INFO ] [EntranceJobScheduler-ThreadPool-3 ] o.a.l.o.s.a.AsyncTaskManager (88) [putExecTask] - submit execTask TaskID_749_otJobId_astJob_1_job_2 to taskManager get executionTask execution_1
2023-02-03 11:17:34.766 [INFO ] [EntranceJobScheduler-ThreadPool-3 ] o.a.l.e.e.DefaultEntranceExecutor (271) [apply] - JobRequest (749) was submitted to Orchestrator.
2023-02-03 11:17:34.767 [INFO ] [CodeReheaterNotifyTaskConsumer ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
2023-02-03 11:17:34.768 [INFO ] [EntranceJobScheduler-ThreadPool-3 ] o.a.l.e.e.DefaultEntranceExecutor (300) [apply] - For job 749 and orchestrator task id 749 to create EngineExecuteAsyncReturn
2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_job_3) to running
2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (77) [run] - Succeed to execute ExecTask(TaskID_749_otJobId_astJob_1_job_3)
2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_job_3 change status Inited => Succeed.
2023-02-03 11:17:34.769 [INFO ] [BaseTaskScheduler-Thread-5 ] o.a.l.o.s.a.AsyncTaskManager (249) [addCompletedTask] - TaskID_749_otJobId_astJob_1_job_3 task completed, now remove from taskManager
2023-02-03 11:17:34.770 [INFO ] [BaseTaskScheduler-Thread-5 ] o.a.l.o.s.a.AsyncTaskManager (283) [apply] - executionTask(execution_1) no task running, user key hadoop-IDE,python-python2, minusNumber: 0
2023-02-03 11:17:34.772 [INFO ] [CodeReheaterNotifyTaskConsumer ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_stage_3) to running
2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (77) [run] - Succeed to execute ExecTask(TaskID_749_otJobId_astJob_1_stage_3)
2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_stage_3 change status Inited => Succeed.
2023-02-03 11:17:34.773 [INFO ] [BaseTaskScheduler-Thread-6 ] o.a.l.o.s.a.AsyncTaskManager (249) [addCompletedTask] - TaskID_749_otJobId_astJob_1_stage_3 task completed, now remove from taskManager
2023-02-03 11:17:34.774 [INFO ] [BaseTaskScheduler-Thread-6 ] o.a.l.o.s.a.AsyncTaskManager (283) [apply] - executionTask(execution_1) no task running, user key hadoop-IDE,python-python2, minusNumber: 0
2023-02-03 11:17:34.775 [INFO ] [CodeReheaterNotifyTaskConsumer ] o.a.l.o.s.a.AsyncTaskManager (195) [apply] - user key hadoop-IDE,python-python2, executionTaskId execution_1 to addNumber: 1
2023-02-03 11:17:34.776 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (70) [run] - ExecTaskRunner Submit execTask(TaskID_749_otJobId_astJob_1_codeExec_1) to running
2023-02-03 11:17:34.778 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.c.p.CodeLogicalUnitExecTask (84) [execute] - Start to execute CodeLogicalUnitExecTask(TaskID_749_otJobId_astJob_1_codeExec_1).
2023-02-03 11:17:34.778 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.e.ComputationEngineConnManager (194) [addMark] - add mark mark_1
2023-02-03 11:17:34.779 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (121) [createExecutor] - create Executor for execId TaskID_749_otJobId_astJob_1_codeExec_1 mark id is mark_1, user hadoop
2023-02-03 11:17:34.779 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.e.ComputationEngineConnManager (125) [getAvailableEngineConnExecutor] - mark mark_1 start to getAvailableEngineConnExecutor
2023-02-03 11:17:34.916 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (138) [createExecutor] - Finished to create Executor for execId TaskID_749_otJobId_astJob_1_codeExec_1 mark id is mark_1, user hadoop
2023-02-03 11:17:34.917 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (97) [askExecutor] - Finished to askExecutor for execId TaskID_749_otJobId_astJob_1_codeExec_1
2023-02-03 11:17:34.940 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.e.s.i.ComputationEngineConnExecutor (83) [execute] - Succeed to submit task TaskID_749_otJobId_astJob_1_codeExec_1 to engineConn(ServiceInstance(linkis-cg-engineconn, dss-master:41636)), Get asyncResponse execID is SubmitResponse(2)
2023-02-03 11:17:34.940 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.c.e.DefaultCodeExecTaskExecutorManager (257) [addEngineConnTaskID] - To add codeExecTaskExecutor engineConn ServiceInstance(linkis-cg-engineconn, dss-master:41636) execTask TaskID_749_otJobId_astJob_1_codeExec_1 mark mark_1 engineConnTaskId 2 to instanceToExecutors
2023-02-03 11:17:34.943 [INFO ] [BaseTaskScheduler-Thread-7 ] o.a.l.o.s.a.AsyncExecTaskRunnerImpl (105) [transientStatus] - TaskID_749_otJobId_astJob_1_codeExec_1 change status Inited => Running.
2023-02-03 11:17:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-4 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 4
2023-02-03 11:18:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-12 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:19:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-8 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:20:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-12 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:21:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-2 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:22:17.764 [INFO ] [AsyncResolver-bootstrap-executor-0 ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
2023-02-03 11:22:39.834 [INFO ] [Orchestrator-Listener-Asyn-Thread-Thread-0] o.a.l.e.l.HDFSCacheLogWriter (115) [writeToFile] - hdfs:///wedata/linkis/log/2023-02-03/IDE/hadoop/749.log write first one line log
2023-02-03 11:22:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-6 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:23:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-15 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:24:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-16 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:25:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-8 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:26:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-18 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:27:17.765 [INFO ] [AsyncResolver-bootstrap-executor-0 ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
2023-02-03 11:27:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-4 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:28:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-15 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:29:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-10 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:30:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-5 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:31:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-3 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:32:17.765 [INFO ] [AsyncResolver-bootstrap-executor-0 ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
2023-02-03 11:32:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-6 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:33:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-7 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:34:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-2 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:35:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-19 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:36:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-17 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:37:17.766 [INFO ] [AsyncResolver-bootstrap-executor-0 ] c.n.d.s.r.a.ConfigClusterResolver (43) [getClusterEndpoints] - Resolving eureka endpoints via configuration
2023-02-03 11:37:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-19 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
2023-02-03 11:38:40.701 [INFO ] [Linkis-Default-Scheduler-Thread-17 ] o.a.l.o.e.i.BaseTaskScheduler (50) [apply$mcV$sp] - Clear finished task from taskFutureCache size 0
````
linkis-cg-linkismanager.log:
````log
2023-02-03 17:59:51.520 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (100) [apply] - Task: 836 start to async(dss-master:9101_12) createEngine, dssmark_id: mark_88
2023-02-03 17:59:51.520 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEngineCreateService (114) [createEngine] - Task: 836 start to create Engine for request: EngineCreateRequest{labels={userCreator=hadoop-LINKISCLI, codeType=python, engineType=python-python2, executeOnce=}, timeOut=660000, user='hadoop', createService='dssmark_id: mark_88', description='null'}.
2023-02-03 17:59:51.547 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEngineCreateService (156) [createEngine] - Suitable ems size is 2
2023-02-03 17:59:51.603 [INFO ] [AskEngineService-Thread-25 ] o.a.l.e.s.s.DefaultEngineConnResourceFactoryService (49) [createEngineResource] - To invoke createEngineResource TimeoutEngineResourceRequest(660000,hadoop,[[key: userCreator, value: {"creator":"LINKISCLI","user":"hadoop"}, str: hadoop-LINKISCLI], [key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2]],{jobId=836, wds.linkis.rm.yarnqueue.memory.max=300G, wds.linkis.rm.client.core.max=10, wds.linkis.rm.instance=10, wds.linkis.engineconn.java.driver.memory=1g, wds.linkis.rm.yarnqueue.instance.max=30, wds.linkis.rm.yarnqueue.cores.max=150, wds.linkis.rm.client.memory.max=20G, wds.linkis.rm.yarnqueue=default, python.version=python2, wds.linkis.engineconn.max.free.time=1h})
2023-02-03 17:59:51.614 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.ResourceLockService (52) [apply$mcZ$sp] - successfully locked labelhadoop-LINKISCLI,python-python2
2023-02-03 17:59:51.616 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.i.DefaultReqResourceService (62) [canRequest] - ResourceInit: hadoop-LINKISCLI,python-python2
2023-02-03 17:59:51.616 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.u.UserConfiguration$ (102) [apply] - hadoop on creator LINKISCLI available engine python resource:Number of instances(实例数):10,(RAM)内存:20.0 GB,cpu:10
2023-02-03 17:59:51.643 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.ResourceLockService (52) [apply$mcZ$sp] - successfully locked labellinkis-cg-engineconnmanager-zdww4:9102
2023-02-03 17:59:51.652 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.i.DefaultResourceManager (306) [apply] - ResourceChanged:linkis-cg-engineconnmanager-zdww4:9102 --> CommonNodeResource{resourceType=LoadInstance, maxResource=Number of instances(实例数):50,(RAM)内存:100.0 GB,cpu:100, minResource=Number of instances(实例数):2,(RAM)内存:4.0 GB,cpu:2, usedResource=Number of instances(实例数):2,(RAM)内存:2.0 GB,cpu:2, lockedResource=Number of instances(实例数):1,(RAM)内存:1024.0 MB,cpu:1, expectedResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, leftResource=Number of instances(实例数):47,(RAM)内存:97.0 GB,cpu:97}
2023-02-03 17:59:51.660 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.i.DefaultResourceManager (306) [apply] - ResourceChanged:hadoop-LINKISCLI,python-python2 --> CommonNodeResource{resourceType=LoadInstance, maxResource=Number of instances(实例数):10,(RAM)内存:20.0 GB,cpu:10, minResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, usedResource=Number of instances(实例数):0,(RAM)内存:0.0 B,cpu:0, lockedResource=Number of instances(实例数):1,(RAM)内存:1024.0 MB,cpu:1, expectedResource=null, leftResource=Number of instances(实例数):9,(RAM)内存:19.0 GB,cpu:9}
2023-02-03 17:59:51.664 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.ResourceLockService (64) [apply$mcV$sp] - unlocked hadoop-LINKISCLI,python-python2
2023-02-03 17:59:51.669 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.r.s.ResourceLockService (64) [apply$mcV$sp] - unlocked linkis-cg-engineconnmanager-zdww4:9102
2023-02-03 17:59:51.669 [WARN ] [AskEngineService-Thread-25 ] o.a.l.m.r.d.RMLabelContainer (156) [getEngineInstanceLabel] - EngineInstanceLabel not found
2023-02-03 17:59:51.730 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEMEngineService (70) [createEngine] - EM ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102) start to create Engine EngineConnBuildRequestImpl(e84fc301-889c-4231-9c33-b46abebaafc5,[[key: userCreator, value: {"creator":"LINKISCLI","user":"hadoop"}, str: hadoop-LINKISCLI], [key: engineType, value: {"engineType":"python","version":"python2"}, str: python-python2]],org.apache.linkis.manager.engineplugin.common.resource.UserNodeResource@4c4ee70e,EngineConnCreationDescImpl(dssmark_id: mark_88,null,{jobId=836, wds.linkis.rm.yarnqueue.memory.max=300G, wds.linkis.rm.client.core.max=10, wds.linkis.rm.instance=10, wds.linkis.engineconn.java.driver.memory=1g, wds.linkis.rm.yarnqueue.instance.max=30, wds.linkis.rm.yarnqueue.cores.max=150, wds.linkis.rm.client.memory.max=20G, wds.linkis.rm.yarnqueue=default, python.version=python2, wds.linkis.engineconn.max.free.time=1h}))
2023-02-03 17:59:51.730 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.p.DefaultEMNodPointer (40) [createEngine] - Start to createEngine ask em ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102)
2023-02-03 17:59:51.751 [INFO ] [qtp2126981566-581 ] o.a.l.m.e.p.l.PythonProcessEngineConnLaunchBuilder (101) [getEnvironment] - Setting up the launch environment for engineconn.
2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.p.DefaultEMNodPointer (43) [createEngine] - Succeed to createEngine ask em ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102), engineNode AMEngineNode{nodeStatus=null, lock='null', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'}
2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEMEngineService (72) [createEngine] - EM ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102) Finished to create Engine e84fc301-889c-4231-9c33-b46abebaafc5
2023-02-03 17:59:51.773 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEngineCreateService (204) [createEngine] - Task: 836 finished to create engineConn AMEngineNode{nodeStatus=null,
lock='null', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'}. ticketId is e84fc301-889c-4231-9c33-b46abebaafc5
2023-02-03 17:59:51.828 [INFO ] [AskEngineService-Thread-25 ] o.a.l.m.a.s.e.DefaultEngineCreateService (365) [apply$mcV$sp] - Start to wait engineConn(AMEngineNode{nodeStatus=null, lock='null', se
rviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:39310), owner='hadoop'}) to be available, but only 11.0 m left.
2023-02-03 17:59:55.743 [INFO ] [qtp2126981566-568 ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (71) [askEngine] - received task: 837, engineAskRequest EngineAskRequest{labels={userCreat
or=hadoop-IDE, codeType=sql, engineType=spark-2.4.3}, timeOut=660000, user='hadoop', createService='dssmark_id: mark_89', description='null'}
2023-02-03 17:59:55.744 [INFO ] [qtp2126981566-568 ] o.a.l.m.a.s.e.DefaultEngineReuseService (72) [reuseEngine] - Start to reuse Engine for request: EngineReuseRequest{timeOut=660000, reu
seCount=0, user='hadoop'}
2023-02-03 17:59:55.769 [INFO ] [qtp2126981566-568 ] o.a.l.m.a.s.e.DefaultEngineReuseService (152) [org$apache$linkis$manager$am$service$engine$DefaultEngineReuseService$$selectEngineToRe
use$1] - prepare to reuse engineNode: ServiceInstance(linkis-cg-engineconn, zdww4:42094)
2023-02-03 17:59:55.814 [INFO ] [qtp2126981566-568 ] o.a.l.m.a.s.e.DefaultEngineReuseService (185) [reuseEngine] - Finished to reuse Engine for request: EngineReuseRequest{timeOut=660000,
reuseCount=0, user='hadoop'} get EngineNode AMEngineNode{nodeStatus=Unlock, lock='java.util.concurrent.Semaphore@30e82d09[Permits = 0]', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:42094), own
er='null'}, time taken 44
2023-02-03 17:59:55.815 [INFO ] [qtp2126981566-568 ] o.a.l.m.a.s.e.DefaultEngineAskEngineService (91) [askEngine] - Finished to ask engine for task: 837 user hadoop by reuse node AMEngine
Node{nodeStatus=Unlock, lock='java.util.concurrent.Semaphore@30e82d09[Permits = 0]', serviceInstance=ServiceInstance(linkis-cg-engineconn, zdww4:42094), owner='null'}
2023-02-03 17:59:55.819 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Idle, serviceInstance=Se
rviceInstance(linkis-cg-engineconn, zdww4:42094)}
2023-02-03 17:59:55.844 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Busy, serviceInstance=Se
rviceInstance(linkis-cg-engineconn, zdww4:42094)}
2023-02-03 18:00:00.837 [INFO ] [Linkis-Default-Scheduler-Thread-1 ] o.a.l.m.a.s.m.NodeHeartbeatMonitor (99) [apply$mcV$sp] - Start to check the health of the node
2023-02-03 18:00:00.841 [INFO ] [Linkis-Default-Scheduler-Thread-1 ] o.a.l.m.a.s.m.NodeHeartbeatMonitor (130) [apply$mcV$sp] - Finished to check the health of the node
2023-02-03 18:00:01.487 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.l.DefaultEngineNodeLocker (50) [releaseLock] - clientServiceInstance(linkis-cg-entrance, dss-master:9104) Start to unlock en
gine ServiceInstance(linkis-cg-engineconn, zdww4:42094)
2023-02-03 18:00:01.519 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-1 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Idle, serviceInstance=Se
rviceInstance(linkis-cg-engineconn, zdww4:42094)}
2023-02-03 18:00:01.531 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.l.DefaultEngineNodeLocker (56) [releaseLock] - clientServiceInstance(linkis-cg-entrance, dss-master:9104) Finished to unlock
engine ServiceInstance(linkis-cg-engineconn, zdww4:42094)
2023-02-03 18:00:01.537 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Unlock, serviceInstance=
ServiceInstance(linkis-cg-engineconn, zdww4:42094)}
2023-02-03 18:00:01.614 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
=ServiceInstance(linkis-cg-engineconnmanager, zdww4:9102)}
2023-02-03 18:00:03.385 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
=ServiceInstance(linkis-cg-engineconnmanager, zdww3:9102)}
2023-02-03 18:00:03.386 [WARN ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.p.i.DefaultNodeMetricManagerPersistence (89) [addOrupdateNodeMetrics] - The request of update node metrics was ignored, becaus
e the node zdww3:9102 is not exist.
2023-02-03 18:00:09.559 [INFO ] [RPC-Receiver-Asyn-Thread-Thread-0 ] o.a.l.m.a.s.h.AMHeartbeatService (70) [heartbeatEventDeal] - Am deal nodeHeartbeatMsg NodeHeartbeatMsg{status=Running, serviceInstance
=ServiceInstance(linkis-cg-engineconnmanager, dss-master:9102)}
````
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org
[GitHub] [linkis] casionone commented on issue #4161: [Question] linkis submit python task to yarn failed
Posted by "casionone (via GitHub)" <gi...@apache.org>.
casionone commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1445282198
Without further feedback, I closed this issue. You can reopen if necessary
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org
[GitHub] [linkis] casionone closed issue #4161: [Question] linkis submit python task to yarn failed
Posted by "casionone (via GitHub)" <gi...@apache.org>.
casionone closed issue #4161: [Question] linkis submit python task to yarn failed
URL: https://github.com/apache/linkis/issues/4161
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org
[GitHub] [linkis] github-actions[bot] commented on issue #4161: [Question] linkis submit python task to yarn failed
Posted by "github-actions[bot] (via GitHub)" <gi...@apache.org>.
github-actions[bot] commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1422352439
## :blush: Welcome to the Apache Linkis community!!
We are glad that you are contributing by opening this issue.
Please make sure to include all the relevant context.
We will be here shortly.
If you are interested in contributing to our website project, please let us know!
You can check out our contributing guide on
:point_right: [How to Participate in Project Contribution](https://linkis.apache.org/community/how-to-contribute).
### Community
|WeChat Assistant|WeChat Public Account|
|-|-|
|<img src="https://linkis.apache.org/Images/wedatasphere_contact_01.png" width="128"/>|<img src="https://linkis.apache.org/Images/gzh_01.png" width="128"/>|
### Mailing Lists
|Name|Description|Subscribe|Unsubscribe|Archive|
|:-----|:--------|:------|:-------|:-----|
| [dev@linkis.apache.org](mailto:dev@linkis.apache.org) | community activity information | [subscribe](mailto:dev-subscribe@linkis.apache.org) | [unsubscribe](mailto:dev-unsubscribe@linkis.apache.org) | [archive](http://mail-archives.apache.org/mod_mbox/linkis-dev) |
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org
[GitHub] [linkis] QuantumXiecao commented on issue #4161: [Question] linkis submit python task to yarn failed
Posted by "QuantumXiecao (via GitHub)" <gi...@apache.org>.
QuantumXiecao commented on issue #4161:
URL: https://github.com/apache/linkis/issues/4161#issuecomment-1422595620
@liygit Could you please supply your submit code to help locating the problem?
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
---------------------------------------------------------------------
To unsubscribe, e-mail: notifications-unsubscribe@linkis.apache.org
For additional commands, e-mail: notifications-help@linkis.apache.org