You are viewing a plain text version of this content. The canonical link for it is here.
Posted to dev@flink.apache.org by "Gyula Fora (Jira)" <ji...@apache.org> on 2022/05/26 12:42:00 UTC

[jira] [Created] (FLINK-27802) Job submission errors are swallowed for Flink 1.15 + HA

Gyula Fora created FLINK-27802:
----------------------------------

             Summary: Job submission errors are swallowed for Flink 1.15 + HA
                 Key: FLINK-27802
                 URL: https://issues.apache.org/jira/browse/FLINK-27802
             Project: Flink
          Issue Type: Improvement
            Reporter: Gyula Fora


We are currently setting both a result store and the "execution.submit-failed-job-on-application-error" config for HA jobs.

This leads to swallowed job submission errors that only show up in the result store, but the flink job is not actually displayed in the failed state:


2022-05-26 12:34:43,497 WARN org.apache.flink.runtime.dispatcher.StandaloneDispatcher [] - Ignoring JobGraph submission 'State machine job' (00000000000000000000000000000000) because the job already reached a globally-terminal state (i.e. FAILED, CANCELED, FINISHED) in a previous execution.
2022-05-26 12:34:43,552 INFO org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap [] - Application completed SUCCESSFULLY


The easiest way to reproduce this is to create a new deployment and set initialSavepointPath to a random missing path.

I consider this a bug in Flink but we should simply disable the execution.submit-failed-job-on-application-error config.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)