You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@flink.apache.org by "zhisheng (Jira)" <ji...@apache.org> on 2022/05/11 14:40:00 UTC

[jira] [Updated] (FLINK-27576) Flink will request new pod when jm pod is delete, but will remove when TaskExecutor exceeded the idle timeout

     [ https://issues.apache.org/jira/browse/FLINK-27576?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

zhisheng updated FLINK-27576:
-----------------------------
    Description: 
flink 1.12.0 enable the ha(zk) and checkpoint, when i use kubectl delete the jm pod, the job will  request new jm pod failover from the last checkpoint , it is ok.  But it will request new tm pod again, but not use actually, the new tm pod will closed when TaskExecutor exceeded the idle timeout . actually it will use the old tm, why need to request for new tm pod? whether the job will fail if the cluster has no resource for the new tm?Can we optimize and reuse the old tm directly?

 

[^jobmanager_log.txt]

^!image-2022-05-11-20-06-58-955.png!^

^!image-2022-05-11-20-08-01-739.png|width=857,height=324!^

  was:
flink 1.12.0 enable the ha(zk) and checkpoint, when i use kubectl delete the jm pod, the job will  request new jm pod failover from the last checkpoint , it is ok.  But it will request new tm pod again, but not use actually, the new tm pod will closed when TaskExecutor exceeded the idle timeout . actually it will use the old tm, why need to request for new tm pod? whether the job will fail if the cluster has no resource for the new tm?Can we optimize and reuse the old tm directly?

 

[^jobmanager_log.txt]

^!image-2022-05-11-20-06-58-955.png!^

^!image-2022-05-11-20-08-01-739.png!^


> Flink will request new pod when jm pod is delete, but will remove when TaskExecutor exceeded the idle timeout 
> --------------------------------------------------------------------------------------------------------------
>
>                 Key: FLINK-27576
>                 URL: https://issues.apache.org/jira/browse/FLINK-27576
>             Project: Flink
>          Issue Type: Bug
>          Components: Deployment / Kubernetes
>    Affects Versions: 1.12.0
>            Reporter: zhisheng
>            Priority: Major
>         Attachments: image-2022-05-11-20-06-58-955.png, image-2022-05-11-20-08-01-739.png, jobmanager_log.txt
>
>
> flink 1.12.0 enable the ha(zk) and checkpoint, when i use kubectl delete the jm pod, the job will  request new jm pod failover from the last checkpoint , it is ok.  But it will request new tm pod again, but not use actually, the new tm pod will closed when TaskExecutor exceeded the idle timeout . actually it will use the old tm, why need to request for new tm pod? whether the job will fail if the cluster has no resource for the new tm?Can we optimize and reuse the old tm directly?
>  
> [^jobmanager_log.txt]
> ^!image-2022-05-11-20-06-58-955.png!^
> ^!image-2022-05-11-20-08-01-739.png|width=857,height=324!^



--
This message was sent by Atlassian Jira
(v8.20.7#820007)