You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@streampark.apache.org by be...@apache.org on 2023/09/24 08:05:26 UTC

[incubator-streampark] branch dev updated: [Improve] Unified Naming Rules for Enumeration Classes (#3133) (#3175)

This is an automated email from the ASF dual-hosted git repository.

benjobs pushed a commit to branch dev
in repository https://gitbox.apache.org/repos/asf/incubator-streampark.git


The following commit(s) were added to refs/heads/dev by this push:
     new 2274d5502 [Improve] Unified Naming Rules for Enumeration Classes (#3133) (#3175)
2274d5502 is described below

commit 2274d5502d529fc7f0ab7e12841b1841082ae335
Author: xsymin <49...@users.noreply.github.com>
AuthorDate: Sun Sep 24 16:05:20 2023 +0800

    [Improve] Unified Naming Rules for Enumeration Classes (#3133) (#3175)
    
    * [Improve] Unified Naming Rules for Enumeration Classes (#3133)
---
 ...plicationType.java => ApplicationTypeEnum.java} |   8 +-
 .../{ClusterState.java => ClusterStateEnum.java}   |  26 +--
 ...velopmentMode.java => DevelopmentModeEnum.java} |  12 +-
 .../{ExecutionMode.java => ExecutionModeEnum.java} |  32 ++--
 ...dType.java => FlinkK8sRestExposedTypeEnum.java} |   8 +-
 ....java => FlinkSqlValidationFailedTypeEnum.java} |   8 +-
 .../{ResolveOrder.java => ResolveOrderEnum.java}   |   8 +-
 .../{RestoreMode.java => RestoreModeEnum.java}     |  12 +-
 .../enums/{Semantic.java => SemanticEnum.java}     |  10 +-
 .../{StorageType.java => StorageTypeEnum.java}     |   8 +-
 .../apache/streampark/common/conf/Workspace.scala  |  14 +-
 .../enums/{ApiType.scala => ApiTypeEnum.scala}     |   4 +-
 .../{PlannerType.scala => PlannerTypeEnum.scala}   |   4 +-
 .../enums/{TableMode.scala => TableModeEnum.scala} |   4 +-
 .../apache/streampark/common/fs/FsOperator.scala   |  12 +-
 .../interceptor/UploadFileTypeInterceptor.java     |   2 +-
 .../streampark/console/base/util/GitUtils.java     |   6 +-
 .../console/core/annotation/PermissionAction.java  |   4 +-
 .../console/core/aspect/StreamParkAspect.java      |  12 +-
 .../console/core/bean/AlertProbeMsg.java           |   4 +-
 .../console/core/bean/AlertTemplate.java           |  25 +--
 .../ApplicationBuildPipelineController.java        |   8 +-
 .../core/controller/ApplicationController.java     |  28 ++--
 .../controller/ApplicationHistoryController.java   |   4 +-
 .../core/controller/FlinkClusterController.java    |   6 +-
 .../console/core/controller/MessageController.java |   6 +-
 .../console/core/controller/ProjectController.java |   4 +-
 .../console/core/entity/AppBuildPipeline.java      |  30 ++--
 .../console/core/entity/Application.java           |  86 +++++-----
 .../console/core/entity/ApplicationConfig.java     |   4 +-
 .../streampark/console/core/entity/Effective.java  |   6 +-
 .../console/core/entity/FlinkCluster.java          |  28 ++--
 .../streampark/console/core/entity/Message.java    |   7 +-
 .../streampark/console/core/entity/Project.java    |  12 +-
 .../streampark/console/core/entity/Resource.java   |   8 +-
 ...ssTokenState.java => AccessTokenStateEnum.java} |   4 +-
 .../enums/{AlertType.java => AlertTypeEnum.java}   |  26 +--
 ...AppExistsState.java => AppExistsStateEnum.java} |   4 +-
 .../enums/{BuildState.java => BuildStateEnum.java} |   6 +-
 .../{CandidateType.java => CandidateTypeEnum.java} |   6 +-
 ...kPointStatus.java => CheckPointStatusEnum.java} |   6 +-
 ...CheckPointType.java => CheckPointTypeEnum.java} |   6 +-
 ...ConfigFileType.java => ConfigFileTypeEnum.java} |   6 +-
 .../{EffectiveType.java => EffectiveTypeEnum.java} |   4 +-
 .../enums/{EngineType.java => EngineTypeEnum.java} |   6 +-
 ...overStrategy.java => FailoverStrategyEnum.java} |   6 +-
 .../{FlinkAppState.java => FlinkAppStateEnum.java} |  46 +++---
 ...rizedError.java => GitAuthorizedErrorEnum.java} |   8 +-
 .../{GitCredential.java => GitCredentialEnum.java} |   8 +-
 .../enums/{LoginType.java => LoginTypeEnum.java}   |   8 +-
 .../enums/{NoticeType.java => NoticeTypeEnum.java} |   6 +-
 .../enums/{Operation.java => OperationEnum.java}   |   6 +-
 .../{OptionState.java => OptionStateEnum.java}     |   6 +-
 ...PermissionType.java => PermissionTypeEnum.java} |   6 +-
 ...aceholderType.java => PlaceholderTypeEnum.java} |   4 +-
 .../{ReleaseState.java => ReleaseStateEnum.java}   |   6 +-
 .../{ResourceFrom.java => ResourceFromEnum.java}   |   6 +-
 .../{ResourceType.java => ResourceTypeEnum.java}   |   6 +-
 .../enums/{StopFrom.java => StopFromEnum.java}     |   2 +-
 .../enums/{UserType.java => UserTypeEnum.java}     |   6 +-
 .../console/core/metrics/flink/CheckPoints.java    |  16 +-
 .../console/core/runner/EnvInitializer.java        |  24 +--
 .../console/core/service/AppBuildPipeService.java  |   4 +-
 .../console/core/service/EffectiveService.java     |   8 +-
 .../console/core/service/FlinkClusterService.java  |   8 +-
 .../console/core/service/FlinkSqlService.java      |   6 +-
 .../console/core/service/MessageService.java       |   4 +-
 .../console/core/service/YarnQueueService.java     |   4 +-
 .../core/service/alert/impl/AlertServiceImpl.java  |  12 +-
 .../application/ApplicationInfoService.java        |   4 +-
 .../application/ApplicationManageService.java      |   6 +-
 .../impl/ApplicationActionServiceImpl.java         | 113 ++++++-------
 .../impl/ApplicationInfoServiceImpl.java           |  50 +++---
 .../impl/ApplicationManageServiceImpl.java         |  85 +++++-----
 .../core/service/impl/AppBuildPipeServiceImpl.java |  56 +++----
 .../service/impl/ApplicationBackUpServiceImpl.java |  15 +-
 .../service/impl/ApplicationConfigServiceImpl.java |  12 +-
 .../core/service/impl/EffectiveServiceImpl.java    |  12 +-
 .../core/service/impl/ExternalLinkServiceImpl.java |   8 +-
 .../core/service/impl/FlinkClusterServiceImpl.java |  56 +++----
 .../core/service/impl/FlinkSqlServiceImpl.java     |  26 +--
 .../core/service/impl/MessageServiceImpl.java      |   6 +-
 .../core/service/impl/ProjectServiceImpl.java      |  26 +--
 .../core/service/impl/ResourceServiceImpl.java     |  10 +-
 .../core/service/impl/SavePointServiceImpl.java    |  29 ++--
 .../core/service/impl/SqlWorkBenchServiceImpl.java |  20 +--
 .../core/service/impl/VariableServiceImpl.java     |   4 +-
 .../core/service/impl/YarnQueueServiceImpl.java    |  13 +-
 .../console/core/task/FlinkAppHttpWatcher.java     | 175 +++++++++++----------
 .../console/core/task/FlinkAppLostWatcher.java     |  10 +-
 .../core/task/FlinkCheckpointProcessor.java        |  22 +--
 .../console/core/task/FlinkClusterWatcher.java     |  58 +++----
 .../core/task/FlinkK8sChangeEventListener.java     |  32 ++--
 .../console/core/task/FlinkK8sWatcherWrapper.java  |  24 +--
 .../console/core/task/ProjectBuildTask.java        |  14 +-
 .../system/controller/AccessTokenController.java   |  10 +-
 .../system/controller/MemberController.java        |   8 +-
 .../console/system/controller/SsoController.java   |   4 +-
 .../console/system/controller/UserController.java  |   8 +-
 .../streampark/console/system/entity/User.java     |   8 +-
 .../system/security/impl/AuthenticatorImpl.java    |  26 +--
 .../system/service/impl/MenuServiceImpl.java       |   6 +-
 .../system/service/impl/TeamServiceImpl.java       |   4 +-
 .../system/service/impl/UserServiceImpl.java       |   6 +-
 .../resources/mapper/core/AlertConfigMapper.xml    |   2 +-
 .../resources/mapper/core/ApplicationMapper.xml    |  16 +-
 .../resources/mapper/core/FlinkClusterMapper.xml   |   6 +-
 .../main/resources/mapper/core/ProjectMapper.xml   |   8 +-
 .../main/resources/mapper/system/UserMapper.xml    |   2 +-
 .../console/core/task/FlinkK8sObserverBroker.scala |  32 ++--
 .../core/utils/FlinkK8sDataTypeConverter.scala     |  32 ++--
 .../streampark/console/SpringUnitTestBase.java     |   8 +-
 .../console/core/entity/ProjectTest.java           |   8 +-
 .../{AlertTypeTest.java => AlertTypeEnumTest.java} |   8 +-
 .../service/ApplicationManageServiceITest.java     |  14 +-
 .../core/service/ApplicationManageServiceTest.java |  10 +-
 .../core/service/FlinkClusterServiceTest.java      |   4 +-
 .../console/core/service/SavePointServiceTest.java |  30 ++--
 .../console/core/service/UserServiceTest.java      |  20 +--
 .../console/core/service/YarnQueueServiceTest.java |  12 +-
 .../core/service/alert/AlertServiceTest.java       |   4 +-
 .../src/views/base/login/LoginForm.vue             |  12 +-
 .../src/views/flink/app/Add.vue                    |   8 +-
 .../src/views/flink/app/Detail.vue                 |   2 +-
 .../src/views/flink/app/EditFlink.vue              |  12 +-
 .../src/views/flink/app/EditStreamPark.vue         |  16 +-
 .../src/views/flink/app/View.vue                   |   6 +-
 .../components/AppView/StartApplicationModal.vue   |   6 +-
 .../src/views/flink/project/Add.vue                |   2 +-
 .../src/views/flink/project/Edit.vue               |  10 +-
 .../src/views/flink/project/View.vue               |  12 +-
 .../views/flink/project/components/ListItem.vue    |  14 +-
 .../flink/project/components/RepositoryGroup.tsx   |  10 +-
 .../views/setting/Alarm/components/AlertModal.vue  |  24 +--
 .../setting/Alarm/components/AlertTypeInfo.vue     |  18 +--
 .../src/views/setting/Alarm/index.data.ts          |   2 +-
 .../src/views/setting/Alarm/index.vue              |  14 +-
 .../src/views/setting/FlinkCluster/EditCluster.vue |   4 +-
 .../src/views/setting/FlinkCluster/index.vue       |  18 +--
 .../src/views/system/user/User.vue                 |   2 +-
 .../flink/client/bean/CancelRequest.scala          |   4 +-
 .../flink/client/bean/DeployRequest.scala          |   7 +-
 .../flink/client/bean/KubernetesSubmitParam.scala  |   6 +-
 .../flink/client/bean/SavepointRequestTrait.scala  |   4 +-
 .../flink/client/bean/ShutDownRequest.scala        |   4 +-
 .../flink/client/bean/SubmitRequest.scala          |  16 +-
 .../client/bean/TriggerSavepointRequest.scala      |   4 +-
 .../flink/client/FlinkClientEndpoint.scala         |   6 +-
 .../impl/KubernetesNativeApplicationClient.scala   |   4 +-
 .../impl/KubernetesNativeSessionClient.scala       |  10 +-
 .../flink/client/impl/YarnApplicationClient.scala  |   6 +-
 .../flink/client/trait/FlinkClientTrait.scala      |  20 +--
 .../client/trait/KubernetesNativeClientTrait.scala |  13 +-
 .../flink/kubernetes/DefaultFlinkK8sWatcher.scala  |   8 +-
 .../flink/kubernetes/KubernetesRetriever.scala     |   4 +-
 ...FlinkJobState.scala => FlinkJobStateEnum.scala} |   6 +-
 ...uteMode.scala => FlinkK8sExecuteModeEnum.scala} |  20 +--
 .../kubernetes/event/FlinkJobStateEvent.scala      |   4 +-
 .../flink/kubernetes/model/ClusterKey.scala        |   4 +-
 .../flink/kubernetes/model/JobStatusCV.scala       |   4 +-
 .../flink/kubernetes/model/TrackId.scala           |  12 +-
 .../kubernetes/watcher/FlinkJobStatusWatcher.scala |  52 +++---
 ...PipelineStatus.java => PipelineStatusEnum.java} |   8 +-
 ...StepStatus.java => PipelineStepStatusEnum.java} |   8 +-
 .../{PipelineType.java => PipelineTypeEnum.java}   |  12 +-
 .../flink/packer/pipeline/BuildPipeline.scala      |  32 ++--
 .../flink/packer/pipeline/BuildRequest.scala       |  22 +--
 .../flink/packer/pipeline/PipeSnapshot.scala       |  10 +-
 .../impl/FlinkK8sApplicationBuildPipeline.scala    |   8 +-
 .../impl/FlinkK8sApplicationBuildPipelineV2.scala  |   8 +-
 .../impl/FlinkK8sSessionBuildPipeline.scala        |   2 +-
 .../pipeline/impl/FlinkRemoteBuildPipeline.scala   |  10 +-
 .../impl/FlinkYarnApplicationBuildPipeline.scala   |   8 +-
 .../streampark/flink/core/FlinkSqlValidator.scala  |   8 +-
 .../flink/core/FlinkStreamingInitializer.scala     |  12 +-
 .../flink/core/FlinkTableInitializer.scala         |  30 ++--
 .../streampark/flink/core/SqlCommandParser.scala   |  10 +-
 ...erationStatus.java => OperationStatusEnum.java} |   8 +-
 ...hOrientation.java => FetchOrientationEnum.java} |   2 +-
 .../streampark/gateway/results/OperationInfo.java  |   8 +-
 .../{ResultKind.java => ResultKindEnum.java}       |   2 +-
 .../gateway/results/ResultQueryCondition.java      |   8 +-
 .../streampark/gateway/results/ResultSet.java      |  16 +-
 .../streampark/gateway/results/TableInfo.java      |  17 +-
 .../results/{TableKind.java => TableKindEnum.java} |   2 +-
 .../gateway/service/SqlGatewayService.java         |   4 +-
 .../gateway/flink/FlinkSqlGatewayImpl.java         |  14 +-
 187 files changed, 1313 insertions(+), 1298 deletions(-)

diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationType.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationTypeEnum.java
similarity index 87%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationType.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationTypeEnum.java
index 70c992d82..557dd4cf8 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationType.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/ApplicationTypeEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.common.enums;
 
-public enum ApplicationType {
+public enum ApplicationTypeEnum {
 
   /** StreamPark Flink */
   STREAMPARK_FLINK(1, "StreamPark Flink"),
@@ -34,7 +34,7 @@ public enum ApplicationType {
   private final int type;
   private final String name;
 
-  ApplicationType(int type, String name) {
+  ApplicationTypeEnum(int type, String name) {
     this.type = type;
     this.name = name;
   }
@@ -47,8 +47,8 @@ public enum ApplicationType {
     return name;
   }
 
-  public static ApplicationType of(int type) {
-    for (ApplicationType appType : ApplicationType.values()) {
+  public static ApplicationTypeEnum of(int type) {
+    for (ApplicationTypeEnum appType : ApplicationTypeEnum.values()) {
       if (appType.getType() == type) {
         return appType;
       }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterState.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterStateEnum.java
similarity index 71%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterState.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterStateEnum.java
index b2599b74c..3cd992f87 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterState.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/ClusterStateEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.common.enums;
 
 /** @since 1.2.3 */
-public enum ClusterState {
+public enum ClusterStateEnum {
 
   /** The cluster was just created but not started */
   CREATED(0),
@@ -49,33 +49,33 @@ public enum ClusterState {
 
   private final Integer state;
 
-  ClusterState(Integer state) {
+  ClusterStateEnum(Integer state) {
     this.state = state;
   }
 
-  public static ClusterState of(Integer value) {
-    for (ClusterState clusterState : values()) {
-      if (clusterState.state.equals(value)) {
-        return clusterState;
+  public static ClusterStateEnum of(Integer value) {
+    for (ClusterStateEnum clusterStateEnum : values()) {
+      if (clusterStateEnum.state.equals(value)) {
+        return clusterStateEnum;
       }
     }
-    return ClusterState.UNKNOWN;
+    return ClusterStateEnum.UNKNOWN;
   }
 
-  public static ClusterState of(String name) {
-    for (ClusterState clusterState : values()) {
-      if (clusterState.name().equals(name)) {
-        return clusterState;
+  public static ClusterStateEnum of(String name) {
+    for (ClusterStateEnum clusterStateEnum : values()) {
+      if (clusterStateEnum.name().equals(name)) {
+        return clusterStateEnum;
       }
     }
-    return ClusterState.UNKNOWN;
+    return ClusterStateEnum.UNKNOWN;
   }
 
   public Integer getState() {
     return state;
   }
 
-  public static boolean isRunning(ClusterState state) {
+  public static boolean isRunning(ClusterStateEnum state) {
     return RUNNING.equals(state);
   }
 }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentMode.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentModeEnum.java
similarity index 80%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentMode.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentModeEnum.java
index 00580a2cd..31377c1f7 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentMode.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/DevelopmentModeEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.common.enums;
 
-public enum DevelopmentMode {
+public enum DevelopmentModeEnum {
 
   /** custom code */
   CUSTOM_CODE("Custom Code", 1),
@@ -32,15 +32,15 @@ public enum DevelopmentMode {
 
   private final Integer mode;
 
-  DevelopmentMode(String name, Integer mode) {
+  DevelopmentModeEnum(String name, Integer mode) {
     this.name = name;
     this.mode = mode;
   }
 
-  public static DevelopmentMode of(Integer value) {
-    for (DevelopmentMode developmentMode : values()) {
-      if (developmentMode.mode.equals(value)) {
-        return developmentMode;
+  public static DevelopmentModeEnum of(Integer value) {
+    for (DevelopmentModeEnum developmentModeEnum : values()) {
+      if (developmentModeEnum.mode.equals(value)) {
+        return developmentModeEnum;
       }
     }
     return null;
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionMode.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionModeEnum.java
similarity index 76%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionMode.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionModeEnum.java
index c903a5bcd..06bedf34b 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionMode.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/ExecutionModeEnum.java
@@ -21,7 +21,7 @@ import com.google.common.collect.Lists;
 
 import java.util.List;
 
-public enum ExecutionMode {
+public enum ExecutionModeEnum {
 
   /** Local mode */
   LOCAL(0, "local"),
@@ -48,24 +48,24 @@ public enum ExecutionMode {
 
   private final String name;
 
-  ExecutionMode(Integer mode, String name) {
+  ExecutionModeEnum(Integer mode, String name) {
     this.mode = mode;
     this.name = name;
   }
 
-  public static ExecutionMode of(Integer value) {
-    for (ExecutionMode executionMode : values()) {
-      if (executionMode.mode.equals(value)) {
-        return executionMode;
+  public static ExecutionModeEnum of(Integer value) {
+    for (ExecutionModeEnum executionModeEnum : values()) {
+      if (executionModeEnum.mode.equals(value)) {
+        return executionModeEnum;
       }
     }
     return null;
   }
 
-  public static ExecutionMode of(String name) {
-    for (ExecutionMode executionMode : values()) {
-      if (executionMode.name.equals(name)) {
-        return executionMode;
+  public static ExecutionModeEnum of(String name) {
+    for (ExecutionModeEnum executionModeEnum : values()) {
+      if (executionModeEnum.name.equals(name)) {
+        return executionModeEnum;
       }
     }
     return null;
@@ -79,17 +79,17 @@ public enum ExecutionMode {
     return name;
   }
 
-  public static boolean isYarnMode(ExecutionMode mode) {
+  public static boolean isYarnMode(ExecutionModeEnum mode) {
     return YARN_PER_JOB == mode || YARN_APPLICATION == mode || YARN_SESSION == mode;
   }
 
   // TODO: We'll inline this method back to the corresponding caller lines
   //  after dropping the yarn perjob mode.
-  public static boolean isYarnPerJobOrAppMode(ExecutionMode mode) {
+  public static boolean isYarnPerJobOrAppMode(ExecutionModeEnum mode) {
     return YARN_PER_JOB == mode || YARN_APPLICATION == mode;
   }
 
-  public static boolean isYarnSessionMode(ExecutionMode mode) {
+  public static boolean isYarnSessionMode(ExecutionModeEnum mode) {
     return YARN_SESSION == mode;
   }
 
@@ -101,7 +101,7 @@ public enum ExecutionMode {
     return KUBERNETES_NATIVE_SESSION == of(value);
   }
 
-  public static boolean isKubernetesMode(ExecutionMode mode) {
+  public static boolean isKubernetesMode(ExecutionModeEnum mode) {
     return KUBERNETES_NATIVE_SESSION == mode || KUBERNETES_NATIVE_APPLICATION == mode;
   }
 
@@ -118,7 +118,7 @@ public enum ExecutionMode {
         KUBERNETES_NATIVE_SESSION.getMode(), KUBERNETES_NATIVE_APPLICATION.getMode());
   }
 
-  public static boolean isSessionMode(ExecutionMode mode) {
+  public static boolean isSessionMode(ExecutionModeEnum mode) {
     return KUBERNETES_NATIVE_SESSION == mode || YARN_SESSION == mode;
   }
 
@@ -126,7 +126,7 @@ public enum ExecutionMode {
     return isRemoteMode(of(value));
   }
 
-  public static boolean isRemoteMode(ExecutionMode mode) {
+  public static boolean isRemoteMode(ExecutionModeEnum mode) {
     return REMOTE == mode;
   }
 }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedType.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedTypeEnum.java
similarity index 85%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedType.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedTypeEnum.java
index c87c5a3a6..df6504453 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedType.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkK8sRestExposedTypeEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.common.enums;
 
 /** kubernetes.rest-service.exposed.type */
-public enum FlinkK8sRestExposedType {
+public enum FlinkK8sRestExposedTypeEnum {
 
   /** LoadBalancer */
   LOAD_BALANCER("LoadBalancer", 0),
@@ -33,13 +33,13 @@ public enum FlinkK8sRestExposedType {
 
   private final Integer type;
 
-  FlinkK8sRestExposedType(String name, Integer type) {
+  FlinkK8sRestExposedTypeEnum(String name, Integer type) {
     this.name = name;
     this.type = type;
   }
 
-  public static FlinkK8sRestExposedType of(Integer value) {
-    for (FlinkK8sRestExposedType order : values()) {
+  public static FlinkK8sRestExposedTypeEnum of(Integer value) {
+    for (FlinkK8sRestExposedTypeEnum order : values()) {
       if (order.type.equals(value)) {
         return order;
       }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedType.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedTypeEnum.java
similarity index 85%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedType.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedTypeEnum.java
index 6f9b8a067..8c32a55b1 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedType.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/FlinkSqlValidationFailedTypeEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.common.enums;
 
-public enum FlinkSqlValidationFailedType {
+public enum FlinkSqlValidationFailedTypeEnum {
 
   /** Basic test failed (such as null, etc.) */
   VERIFY_FAILED(1),
@@ -36,12 +36,12 @@ public enum FlinkSqlValidationFailedType {
 
   private final int failedType;
 
-  FlinkSqlValidationFailedType(int failedType) {
+  FlinkSqlValidationFailedTypeEnum(int failedType) {
     this.failedType = failedType;
   }
 
-  public static FlinkSqlValidationFailedType of(Integer value) {
-    for (FlinkSqlValidationFailedType type : values()) {
+  public static FlinkSqlValidationFailedTypeEnum of(Integer value) {
+    for (FlinkSqlValidationFailedTypeEnum type : values()) {
       if (type.failedType == value) {
         return type;
       }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrder.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrderEnum.java
similarity index 87%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrder.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrderEnum.java
index e69823e27..0ab2a8821 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrder.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/ResolveOrderEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.common.enums;
 
 /** classloader.resolve-order */
-public enum ResolveOrder {
+public enum ResolveOrderEnum {
 
   /** parent-first */
   PARENT_FIRST("parent-first", 0),
@@ -30,13 +30,13 @@ public enum ResolveOrder {
 
   private final Integer order;
 
-  ResolveOrder(String name, Integer order) {
+  ResolveOrderEnum(String name, Integer order) {
     this.name = name;
     this.order = order;
   }
 
-  public static ResolveOrder of(Integer value) {
-    for (ResolveOrder order : values()) {
+  public static ResolveOrderEnum of(Integer value) {
+    for (ResolveOrderEnum order : values()) {
       if (order.order.equals(value)) {
         return order;
       }
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreMode.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreModeEnum.java
similarity index 88%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreMode.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreModeEnum.java
index e4a49e8fc..bb48c1d04 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreMode.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/RestoreModeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.common.enums;
 
 import java.util.Objects;
 
-public enum RestoreMode {
+public enum RestoreModeEnum {
 
   /**
    * In this mode Flink claims ownership of the snapshot and essentially treats it like a
@@ -51,7 +51,7 @@ public enum RestoreMode {
     return this.mode;
   }
 
-  RestoreMode(int mode) {
+  RestoreModeEnum(int mode) {
     this.mode = mode;
   }
 
@@ -59,10 +59,10 @@ public enum RestoreMode {
     return this.toString();
   }
 
-  public static RestoreMode of(Integer value) {
-    for (RestoreMode restoreMode : values()) {
-      if (Objects.equals(restoreMode.mode, value)) {
-        return restoreMode;
+  public static RestoreModeEnum of(Integer value) {
+    for (RestoreModeEnum restoreModeEnum : values()) {
+      if (Objects.equals(restoreModeEnum.mode, value)) {
+        return restoreModeEnum;
       }
     }
     return null;
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/Semantic.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/SemanticEnum.java
similarity index 85%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/Semantic.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/SemanticEnum.java
index f405f2818..9f5ab9f47 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/Semantic.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/SemanticEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.common.enums;
 
 /** Flink consistency semantics */
-public enum Semantic {
+public enum SemanticEnum {
 
   /**
    * Ensure that the counting results obtained after a fault are consistent with the correct values.
@@ -31,10 +31,10 @@ public enum Semantic {
   /** After the fault occurs, the counting results may be lost. */
   NONE;
 
-  public static Semantic of(String name) {
-    for (Semantic semantic : Semantic.values()) {
-      if (name.equals(semantic.name())) {
-        return semantic;
+  public static SemanticEnum of(String name) {
+    for (SemanticEnum semanticEnum : SemanticEnum.values()) {
+      if (name.equals(semanticEnum.name())) {
+        return semanticEnum;
       }
     }
     return null;
diff --git a/streampark-common/src/main/java/org/apache/streampark/common/enums/StorageType.java b/streampark-common/src/main/java/org/apache/streampark/common/enums/StorageTypeEnum.java
similarity index 88%
rename from streampark-common/src/main/java/org/apache/streampark/common/enums/StorageType.java
rename to streampark-common/src/main/java/org/apache/streampark/common/enums/StorageTypeEnum.java
index 11be2cbec..8d49a5f98 100644
--- a/streampark-common/src/main/java/org/apache/streampark/common/enums/StorageType.java
+++ b/streampark-common/src/main/java/org/apache/streampark/common/enums/StorageTypeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.common.enums;
 
 import org.apache.commons.lang3.StringUtils;
 
-public enum StorageType {
+public enum StorageTypeEnum {
 
   /** hdfs */
   HDFS("hdfs"),
@@ -29,7 +29,7 @@ public enum StorageType {
 
   private final String type;
 
-  StorageType(String type) {
+  StorageTypeEnum(String type) {
     this.type = type;
   }
 
@@ -37,11 +37,11 @@ public enum StorageType {
     return type;
   }
 
-  public static StorageType of(String identifier) {
+  public static StorageTypeEnum of(String identifier) {
     if (StringUtils.isBlank(identifier)) {
       return LFS;
     }
-    for (StorageType type : values()) {
+    for (StorageTypeEnum type : values()) {
       if (type.type.equals(identifier)) {
         return type;
       }
diff --git a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
index 8f3e7b10c..d7d7ee13e 100644
--- a/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
+++ b/streampark-common/src/main/scala/org/apache/streampark/common/conf/Workspace.scala
@@ -16,7 +16,7 @@
  */
 package org.apache.streampark.common.conf
 
-import org.apache.streampark.common.enums.StorageType
+import org.apache.streampark.common.enums.StorageTypeEnum
 import org.apache.streampark.common.util.{HdfsUtils, SystemPropertyUtils}
 import org.apache.streampark.common.util.ImplicitsUtils._
 
@@ -24,11 +24,11 @@ import java.net.URI
 
 object Workspace {
 
-  def of(storageType: StorageType): Workspace = Workspace(storageType)
+  def of(storageType: StorageTypeEnum): Workspace = Workspace(storageType)
 
-  lazy val local: Workspace = Workspace.of(StorageType.LFS)
+  lazy val local: Workspace = Workspace.of(StorageTypeEnum.LFS)
 
-  lazy val remote: Workspace = Workspace.of(StorageType.HDFS)
+  lazy val remote: Workspace = Workspace.of(StorageTypeEnum.HDFS)
 
   private[this] lazy val localWorkspace = local.WORKSPACE
 
@@ -52,7 +52,7 @@ object Workspace {
 
 }
 
-case class Workspace(storageType: StorageType) {
+case class Workspace(storageType: StorageTypeEnum) {
 
   private[this] def getConfigValue[T](option: InternalOption): T = {
     val s = SystemPropertyUtils.get(option.key)
@@ -68,11 +68,11 @@ case class Workspace(storageType: StorageType) {
 
   lazy val WORKSPACE: String = {
     storageType match {
-      case StorageType.LFS =>
+      case StorageTypeEnum.LFS =>
         val path: String = getConfigValue[String](CommonConfig.STREAMPARK_WORKSPACE_LOCAL)
         require(path != null, "[StreamPark] streampark.workspace.local must not be null")
         path
-      case StorageType.HDFS =>
+      case StorageTypeEnum.HDFS =>
         val path: String = getConfigValue[String](CommonConfig.STREAMPARK_WORKSPACE_REMOTE)
         path match {
           case p if p.isEmpty =>
diff --git a/streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiType.scala b/streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiTypeEnum.scala
similarity index 92%
rename from streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiType.scala
rename to streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiTypeEnum.scala
index d261dc678..22578a37e 100644
--- a/streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiType.scala
+++ b/streampark-common/src/main/scala/org/apache/streampark/common/enums/ApiTypeEnum.scala
@@ -16,7 +16,7 @@
  */
 package org.apache.streampark.common.enums
 
-object ApiType extends Enumeration {
+object ApiTypeEnum extends Enumeration {
   type ApiType = Value
-  val java, scala = Value
+  val JAVA, SCALA = Value
 }
diff --git a/streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerType.scala b/streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerTypeEnum.scala
similarity index 92%
rename from streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerType.scala
rename to streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerTypeEnum.scala
index 9951f8a65..2f4e0f006 100644
--- a/streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerType.scala
+++ b/streampark-common/src/main/scala/org/apache/streampark/common/enums/PlannerTypeEnum.scala
@@ -16,7 +16,7 @@
  */
 package org.apache.streampark.common.enums
 
-object PlannerType extends Enumeration {
+object PlannerTypeEnum extends Enumeration {
   type PlannerType = Value
-  val blink, old, any = Value
+  val BLINK, OLD, ANY = Value
 }
diff --git a/streampark-common/src/main/scala/org/apache/streampark/common/enums/TableMode.scala b/streampark-common/src/main/scala/org/apache/streampark/common/enums/TableModeEnum.scala
similarity index 92%
rename from streampark-common/src/main/scala/org/apache/streampark/common/enums/TableMode.scala
rename to streampark-common/src/main/scala/org/apache/streampark/common/enums/TableModeEnum.scala
index 50bd9628d..88a78188c 100644
--- a/streampark-common/src/main/scala/org/apache/streampark/common/enums/TableMode.scala
+++ b/streampark-common/src/main/scala/org/apache/streampark/common/enums/TableModeEnum.scala
@@ -16,7 +16,7 @@
  */
 package org.apache.streampark.common.enums
 
-object TableMode extends Enumeration {
+object TableModeEnum extends Enumeration {
   type TableMode = Value
-  val batch, streaming = Value
+  val BATCH, STREAMING = Value
 }
diff --git a/streampark-common/src/main/scala/org/apache/streampark/common/fs/FsOperator.scala b/streampark-common/src/main/scala/org/apache/streampark/common/fs/FsOperator.scala
index b7565371c..d54014af4 100644
--- a/streampark-common/src/main/scala/org/apache/streampark/common/fs/FsOperator.scala
+++ b/streampark-common/src/main/scala/org/apache/streampark/common/fs/FsOperator.scala
@@ -17,18 +17,18 @@
 
 package org.apache.streampark.common.fs
 
-import org.apache.streampark.common.enums.StorageType
+import org.apache.streampark.common.enums.StorageTypeEnum
 
 object FsOperator {
 
-  lazy val lfs: FsOperator = FsOperator.of(StorageType.LFS)
+  lazy val lfs: FsOperator = FsOperator.of(StorageTypeEnum.LFS)
 
-  lazy val hdfs: FsOperator = FsOperator.of(StorageType.HDFS)
+  lazy val hdfs: FsOperator = FsOperator.of(StorageTypeEnum.HDFS)
 
-  def of(storageType: StorageType): FsOperator = {
+  def of(storageType: StorageTypeEnum): FsOperator = {
     storageType match {
-      case StorageType.HDFS => HdfsOperator
-      case StorageType.LFS => LfsOperator
+      case StorageTypeEnum.HDFS => HdfsOperator
+      case StorageTypeEnum.LFS => LfsOperator
       case _ => throw new UnsupportedOperationException(s"Unsupported storageType:$storageType")
     }
   }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/interceptor/UploadFileTypeInterceptor.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/interceptor/UploadFileTypeInterceptor.java
index e575c7824..2dde6014c 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/interceptor/UploadFileTypeInterceptor.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/interceptor/UploadFileTypeInterceptor.java
@@ -75,7 +75,7 @@ public class UploadFileTypeInterceptor implements HandlerInterceptor {
     try {
       Metadata metadata = new Metadata();
       AutoDetectParser parser = new AutoDetectParser();
-      parser.parse(input, new DefaultHandler(), metadata, new ParseContext());
+      parser.parse(stream, new DefaultHandler(), metadata, new ParseContext());
       String mimeType = metadata.get(HttpHeaders.CONTENT_TYPE);
       return contentType.contains("text/x-python")
           && MediaType.TEXT_PLAIN.toString().equals(mimeType);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/util/GitUtils.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/util/GitUtils.java
index 57531baa9..7aab84af6 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/util/GitUtils.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/base/util/GitUtils.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.base.util;
 import org.apache.streampark.common.util.FileUtils;
 import org.apache.streampark.common.util.SystemPropertyUtils;
 import org.apache.streampark.console.core.entity.Project;
-import org.apache.streampark.console.core.enums.GitCredential;
+import org.apache.streampark.console.core.enums.GitCredentialEnum;
 
 import org.apache.commons.lang3.StringUtils;
 
@@ -79,8 +79,8 @@ public class GitUtils {
   }
 
   private static void setCredentials(TransportCommand<?, ?> transportCommand, Project project) {
-    GitCredential gitCredential = GitCredential.of(project.getGitCredential());
-    switch (gitCredential) {
+    GitCredentialEnum gitCredentialEnum = GitCredentialEnum.of(project.getGitCredential());
+    switch (gitCredentialEnum) {
       case HTTPS:
         if (!StringUtils.isAllBlank(project.getUserName(), project.getPassword())) {
           UsernamePasswordCredentialsProvider credentialsProvider =
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/annotation/PermissionAction.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/annotation/PermissionAction.java
index 04d964c61..5c6a4ad7d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/annotation/PermissionAction.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/annotation/PermissionAction.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.annotation;
 
-import org.apache.streampark.console.core.enums.PermissionType;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
 
 import java.lang.annotation.ElementType;
 import java.lang.annotation.Retention;
@@ -29,5 +29,5 @@ import java.lang.annotation.Target;
 public @interface PermissionAction {
   String id();
 
-  PermissionType type();
+  PermissionTypeEnum type();
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/aspect/StreamParkAspect.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/aspect/StreamParkAspect.java
index a15583333..a16765d80 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/aspect/StreamParkAspect.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/aspect/StreamParkAspect.java
@@ -24,8 +24,8 @@ import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.core.annotation.ApiAccess;
 import org.apache.streampark.console.core.annotation.PermissionAction;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.PermissionType;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.console.core.task.FlinkAppHttpWatcher;
@@ -109,13 +109,13 @@ public class StreamParkAspect {
     User currentUser = commonService.getCurrentUser();
     ApiAlertException.throwIfNull(currentUser, "Permission denied, please login first.");
 
-    boolean isAdmin = currentUser.getUserType() == UserType.ADMIN;
+    boolean isAdmin = currentUser.getUserTypeEnum() == UserTypeEnum.ADMIN;
 
     if (!isAdmin) {
-      PermissionType permissionType = permissionAction.type();
+      PermissionTypeEnum permissionTypeEnum = permissionAction.type();
       Long paramId = getParamId(joinPoint, methodSignature, permissionAction.id());
 
-      switch (permissionType) {
+      switch (permissionTypeEnum) {
         case USER:
           ApiAlertException.throwIfTrue(
               !currentUser.getUserId().equals(paramId),
@@ -137,7 +137,7 @@ public class StreamParkAspect {
           break;
         default:
           throw new IllegalArgumentException(
-              String.format("Permission type %s is not supported.", permissionType));
+              String.format("Permission type %s is not supported.", permissionTypeEnum));
       }
     }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertProbeMsg.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertProbeMsg.java
index f9bb31604..af6069993 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertProbeMsg.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertProbeMsg.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.bean;
 
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 
 import lombok.Data;
 import lombok.NoArgsConstructor;
@@ -40,7 +40,7 @@ public class AlertProbeMsg {
 
   private Integer cancelledJobs = 0;
 
-  public void compute(FlinkAppState state) {
+  public void compute(FlinkAppStateEnum state) {
     this.probeJobs++;
     switch (state) {
       case LOST:
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertTemplate.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertTemplate.java
index be407078a..e05195637 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertTemplate.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/bean/AlertTemplate.java
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.console.core.bean;
 
-import org.apache.streampark.common.enums.ClusterState;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ClusterStateEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.DateUtils;
 import org.apache.streampark.common.util.YarnUtils;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkCluster;
-import org.apache.streampark.console.core.enums.CheckPointStatus;
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.CheckPointStatusEnum;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 
 import lombok.Data;
 
@@ -57,7 +57,7 @@ public class AlertTemplate implements Serializable {
   private Integer lostJobs;
   private Integer cancelledJobs;
 
-  public static AlertTemplate of(Application application, FlinkAppState appState) {
+  public static AlertTemplate of(Application application, FlinkAppStateEnum appState) {
     return new AlertTemplateBuilder()
         .setDuration(application.getStartTime(), application.getEndTime())
         .setJobName(application.getJobName())
@@ -74,7 +74,8 @@ public class AlertTemplate implements Serializable {
         .build();
   }
 
-  public static AlertTemplate of(Application application, CheckPointStatus checkPointStatus) {
+  public static AlertTemplate of(
+      Application application, CheckPointStatusEnum checkPointStatusEnum) {
     return new AlertTemplateBuilder()
         .setDuration(application.getStartTime(), application.getEndTime())
         .setJobName(application.getJobName())
@@ -90,7 +91,7 @@ public class AlertTemplate implements Serializable {
         .build();
   }
 
-  public static AlertTemplate of(FlinkCluster cluster, ClusterState clusterState) {
+  public static AlertTemplate of(FlinkCluster cluster, ClusterStateEnum clusterStateEnum) {
     return new AlertTemplateBuilder()
         .setDuration(cluster.getStartTime(), cluster.getEndTime())
         .setJobName(cluster.getClusterName())
@@ -98,10 +99,10 @@ public class AlertTemplate implements Serializable {
         .setStartTime(cluster.getStartTime())
         .setEndTime(cluster.getEndTime())
         .setType(3)
-        .setTitle(String.format("Notify: %s %s", cluster.getClusterName(), clusterState.name()))
+        .setTitle(String.format("Notify: %s %s", cluster.getClusterName(), clusterStateEnum.name()))
         .setSubject(
-            String.format("StreamPark Alert: %s %s", cluster.getClusterName(), clusterState))
-        .setStatus(clusterState.name())
+            String.format("StreamPark Alert: %s %s", cluster.getClusterName(), clusterStateEnum))
+        .setStatus(clusterStateEnum.name())
         .setAllJobs(cluster.getAllJobs())
         .setAffectedJobs(cluster.getAffectedJobs())
         .build();
@@ -186,8 +187,8 @@ public class AlertTemplate implements Serializable {
       return this;
     }
 
-    public AlertTemplateBuilder setLink(ExecutionMode mode, String appId) {
-      if (ExecutionMode.isYarnMode(mode)) {
+    public AlertTemplateBuilder setLink(ExecutionModeEnum mode, String appId) {
+      if (ExecutionModeEnum.isYarnMode(mode)) {
         String format = "%s/proxy/%s/";
         String url = String.format(format, YarnUtils.getRMWebAppURL(false), appId);
         alertTemplate.setLink(url);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationBuildPipelineController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationBuildPipelineController.java
index 14dcf7326..a692e56fa 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationBuildPipelineController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationBuildPipelineController.java
@@ -23,10 +23,10 @@ import org.apache.streampark.console.core.annotation.ApiAccess;
 import org.apache.streampark.console.core.annotation.PermissionAction;
 import org.apache.streampark.console.core.bean.AppBuildDockerResolvedDetail;
 import org.apache.streampark.console.core.entity.AppBuildPipeline;
-import org.apache.streampark.console.core.enums.PermissionType;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
 import org.apache.streampark.console.core.service.AppBuildPipeService;
 import org.apache.streampark.flink.packer.pipeline.DockerResolvedSnapshot;
-import org.apache.streampark.flink.packer.pipeline.PipelineType;
+import org.apache.streampark.flink.packer.pipeline.PipelineTypeEnum;
 
 import org.apache.shiro.authz.annotation.RequiresPermissions;
 
@@ -75,7 +75,7 @@ public class ApplicationBuildPipelineController {
         schema = @Schema(defaultValue = "false", implementation = boolean.class))
   })
   @ApiAccess
-  @PermissionAction(id = "#appId", type = PermissionType.APP)
+  @PermissionAction(id = "#appId", type = PermissionTypeEnum.APP)
   @PostMapping(value = "build")
   @RequiresPermissions("app:create")
   public RestResponse buildApplication(Long appId, boolean forceBuild) {
@@ -103,7 +103,7 @@ public class ApplicationBuildPipelineController {
     details.put("pipeline", pipeline.map(AppBuildPipeline::toView).orElse(null));
 
     if (pipeline.isPresent()
-        && PipelineType.FLINK_NATIVE_K8S_APPLICATION == pipeline.get().getPipeType()) {
+        && PipelineTypeEnum.FLINK_NATIVE_K8S_APPLICATION == pipeline.get().getPipeType()) {
       DockerResolvedSnapshot dockerProgress =
           appBuildPipeService.getDockerProgressDetailSnapshot(appId);
       details.put("docker", AppBuildDockerResolvedDetail.of(dockerProgress));
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationController.java
index 048f95f2c..dfe803bce 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationController.java
@@ -29,8 +29,8 @@ import org.apache.streampark.console.core.annotation.PermissionAction;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ApplicationBackUp;
 import org.apache.streampark.console.core.entity.ApplicationLog;
-import org.apache.streampark.console.core.enums.AppExistsState;
-import org.apache.streampark.console.core.enums.PermissionType;
+import org.apache.streampark.console.core.enums.AppExistsStateEnum;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
 import org.apache.streampark.console.core.service.ApplicationLogService;
 import org.apache.streampark.console.core.service.ResourceService;
@@ -92,7 +92,7 @@ public class ApplicationController {
 
   @Operation(summary = "Create application")
   @ApiAccess
-  @PermissionAction(id = "#app.teamId", type = PermissionType.TEAM)
+  @PermissionAction(id = "#app.teamId", type = PermissionTypeEnum.TEAM)
   @PostMapping("create")
   @RequiresPermissions("app:create")
   public RestResponse create(Application app) throws IOException {
@@ -118,7 +118,7 @@ public class ApplicationController {
     @Parameter(name = "args", description = "new application args", in = ParameterIn.QUERY)
   })
   @ApiAccess
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping(value = "copy")
   @RequiresPermissions("app:copy")
   public RestResponse copy(@Parameter(hidden = true) Application app) throws IOException {
@@ -128,7 +128,7 @@ public class ApplicationController {
 
   @Operation(summary = "Update application")
   @AppUpdated
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping("update")
   @RequiresPermissions("app:update")
   public RestResponse update(Application app) {
@@ -163,7 +163,7 @@ public class ApplicationController {
 
   @Operation(summary = "Revoke application")
   @AppUpdated
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping("revoke")
   @RequiresPermissions("app:release")
   public RestResponse revoke(Application app) {
@@ -202,7 +202,7 @@ public class ApplicationController {
         schema = @Schema(implementation = boolean.class, defaultValue = "false"))
   })
   @ApiAccess
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping(value = "start")
   @RequiresPermissions("app:start")
   public RestResponse start(@Parameter(hidden = true) Application app) {
@@ -253,7 +253,7 @@ public class ApplicationController {
         example = "false",
         schema = @Schema(implementation = boolean.class, defaultValue = "false"))
   })
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping(value = "cancel")
   @RequiresPermissions("app:cancel")
   public RestResponse cancel(@Parameter(hidden = true) Application app) throws Exception {
@@ -264,7 +264,7 @@ public class ApplicationController {
   @Operation(summary = "Clean application")
   @AppUpdated
   @ApiAccess
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping("clean")
   @RequiresPermissions("app:clean")
   public RestResponse clean(Application app) {
@@ -274,7 +274,7 @@ public class ApplicationController {
 
   /** force stop(stop normal start or in progress) */
   @Operation(summary = "Force stop application")
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping("forcedStop")
   @RequiresPermissions("app:cancel")
   public RestResponse forcedStop(Application app) {
@@ -298,7 +298,7 @@ public class ApplicationController {
   @Operation(summary = "Check the application exist status")
   @PostMapping("checkName")
   public RestResponse checkName(Application app) {
-    AppExistsState exists = applicationInfoService.checkExists(app);
+    AppExistsStateEnum exists = applicationInfoService.checkExists(app);
     return RestResponse.success(exists.get());
   }
 
@@ -331,7 +331,7 @@ public class ApplicationController {
   }
 
   @Operation(summary = "Delete application operation log")
-  @PermissionAction(id = "#applicationLog.appId", type = PermissionType.APP)
+  @PermissionAction(id = "#applicationLog.appId", type = PermissionTypeEnum.APP)
   @PostMapping("deleteOperationLog")
   @RequiresPermissions("app:delete")
   public RestResponse deleteOperationLog(ApplicationLog applicationLog) {
@@ -340,7 +340,7 @@ public class ApplicationController {
   }
 
   @Operation(summary = "Delete application")
-  @PermissionAction(id = "#app.id", type = PermissionType.APP)
+  @PermissionAction(id = "#app.id", type = PermissionTypeEnum.APP)
   @PostMapping("delete")
   @RequiresPermissions("app:delete")
   public RestResponse delete(Application app) throws InternalException {
@@ -349,7 +349,7 @@ public class ApplicationController {
   }
 
   @Operation(summary = "Backup application when deleted")
-  @PermissionAction(id = "#backUp.appId", type = PermissionType.APP)
+  @PermissionAction(id = "#backUp.appId", type = PermissionTypeEnum.APP)
   @PostMapping("deletebak")
   public RestResponse deleteBak(ApplicationBackUp backUp) throws InternalException {
     Boolean deleted = backUpService.delete(backUp.getId());
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationHistoryController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationHistoryController.java
index 324ca4066..4a09164da 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationHistoryController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ApplicationHistoryController.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.controller;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.core.service.application.ApplicationInfoService;
 
@@ -65,7 +65,7 @@ public class ApplicationHistoryController {
   @RequiresPermissions("app:create")
   public RestResponse listSessionClusterId(int executionMode) {
     List<String> clusterIds;
-    switch (ExecutionMode.of(executionMode)) {
+    switch (ExecutionModeEnum.of(executionMode)) {
       case KUBERNETES_NATIVE_SESSION:
       case YARN_SESSION:
       case REMOTE:
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/FlinkClusterController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/FlinkClusterController.java
index a2b0048c2..64f207214 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/FlinkClusterController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/FlinkClusterController.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.controller;
 
-import org.apache.streampark.common.enums.ClusterState;
+import org.apache.streampark.common.enums.ClusterStateEnum;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.base.exception.InternalException;
 import org.apache.streampark.console.core.bean.ResponseResult;
@@ -93,7 +93,7 @@ public class FlinkClusterController {
   @Operation(summary = "Start flink cluster")
   @PostMapping("start")
   public RestResponse start(FlinkCluster cluster) {
-    flinkClusterService.updateClusterState(cluster.getId(), ClusterState.STARTING);
+    flinkClusterService.updateClusterState(cluster.getId(), ClusterStateEnum.STARTING);
     flinkClusterService.start(cluster);
     return RestResponse.success();
   }
@@ -102,7 +102,7 @@ public class FlinkClusterController {
   @PostMapping("shutdown")
   public RestResponse shutdown(FlinkCluster cluster) {
     if (flinkClusterService.allowShutdownCluster(cluster)) {
-      flinkClusterService.updateClusterState(cluster.getId(), ClusterState.CANCELLING);
+      flinkClusterService.updateClusterState(cluster.getId(), ClusterStateEnum.CANCELLING);
       flinkClusterService.shutdown(cluster);
     }
     return RestResponse.success();
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/MessageController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/MessageController.java
index 87d763783..7fc544c4d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/MessageController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/MessageController.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.controller;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.core.entity.Message;
-import org.apache.streampark.console.core.enums.NoticeType;
+import org.apache.streampark.console.core.enums.NoticeTypeEnum;
 import org.apache.streampark.console.core.service.MessageService;
 
 import com.baomidou.mybatisplus.core.metadata.IPage;
@@ -45,8 +45,8 @@ public class MessageController {
   @Operation(summary = "List notices")
   @PostMapping("notice")
   public RestResponse notice(Integer type, RestRequest request) {
-    NoticeType noticeType = NoticeType.of(type);
-    IPage<Message> pages = messageService.getUnRead(noticeType, request);
+    NoticeTypeEnum noticeTypeEnum = NoticeTypeEnum.of(type);
+    IPage<Message> pages = messageService.getUnRead(noticeTypeEnum, request);
     return RestResponse.success(pages);
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ProjectController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ProjectController.java
index 859261933..57476dfed 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ProjectController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/controller/ProjectController.java
@@ -22,7 +22,7 @@ import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.core.annotation.AppUpdated;
 import org.apache.streampark.console.core.entity.Project;
-import org.apache.streampark.console.core.enums.GitAuthorizedError;
+import org.apache.streampark.console.core.enums.GitAuthorizedErrorEnum;
 import org.apache.streampark.console.core.service.ProjectService;
 
 import org.apache.shiro.authz.annotation.RequiresPermissions;
@@ -120,7 +120,7 @@ public class ProjectController {
   @Operation(summary = "Authenticate git project")
   @PostMapping("gitcheck")
   public RestResponse gitCheck(Project project) {
-    GitAuthorizedError error = project.gitCheck();
+    GitAuthorizedErrorEnum error = project.gitCheck();
     return RestResponse.success().data(error.getType());
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/AppBuildPipeline.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/AppBuildPipeline.java
index fddcd80a2..75b5e386c 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/AppBuildPipeline.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/AppBuildPipeline.java
@@ -23,9 +23,9 @@ import org.apache.streampark.flink.packer.pipeline.BuildPipeline;
 import org.apache.streampark.flink.packer.pipeline.BuildResult;
 import org.apache.streampark.flink.packer.pipeline.PipeError;
 import org.apache.streampark.flink.packer.pipeline.PipeSnapshot;
-import org.apache.streampark.flink.packer.pipeline.PipelineStatus;
-import org.apache.streampark.flink.packer.pipeline.PipelineStepStatus;
-import org.apache.streampark.flink.packer.pipeline.PipelineType;
+import org.apache.streampark.flink.packer.pipeline.PipelineStatusEnum;
+import org.apache.streampark.flink.packer.pipeline.PipelineStepStatusEnum;
+import org.apache.streampark.flink.packer.pipeline.PipelineTypeEnum;
 
 import org.apache.commons.lang3.StringUtils;
 
@@ -96,37 +96,37 @@ public class AppBuildPipeline {
 
   @Nonnull
   @JsonIgnore
-  public PipelineType getPipeType() {
-    return PipelineType.of(pipeTypeCode);
+  public PipelineTypeEnum getPipeType() {
+    return PipelineTypeEnum.of(pipeTypeCode);
   }
 
   @JsonIgnore
-  public AppBuildPipeline setPipeType(@Nonnull PipelineType pipeType) {
+  public AppBuildPipeline setPipeType(@Nonnull PipelineTypeEnum pipeType) {
     this.pipeTypeCode = pipeType.getCode();
     return this;
   }
 
   @Nonnull
   @JsonIgnore
-  public PipelineStatus getPipelineStatus() {
-    return PipelineStatus.of(pipeStatusCode);
+  public PipelineStatusEnum getPipelineStatus() {
+    return PipelineStatusEnum.of(pipeStatusCode);
   }
 
   @JsonIgnore
-  public AppBuildPipeline setPipeStatus(@Nonnull PipelineStatus pipeStatus) {
+  public AppBuildPipeline setPipeStatus(@Nonnull PipelineStatusEnum pipeStatus) {
     this.pipeStatusCode = pipeStatus.getCode();
     return this;
   }
 
   @Nonnull
   @JsonIgnore
-  public Map<Integer, PipelineStepStatus> getStepStatus() {
+  public Map<Integer, PipelineStepStatusEnum> getStepStatus() {
     if (StringUtils.isBlank(stepStatusJson)) {
       return Collections.emptyMap();
     }
     try {
       return JacksonUtils.read(
-          stepStatusJson, new TypeReference<HashMap<Integer, PipelineStepStatus>>() {});
+          stepStatusJson, new TypeReference<HashMap<Integer, PipelineStepStatusEnum>>() {});
     } catch (JsonProcessingException e) {
       log.error(
           "json parse error on ApplicationBuildPipeline, stepStatusJson={}", stepStatusJson, e);
@@ -135,7 +135,7 @@ public class AppBuildPipeline {
   }
 
   @JsonIgnore
-  public AppBuildPipeline setStepStatus(@Nonnull Map<Integer, PipelineStepStatus> stepStatus) {
+  public AppBuildPipeline setStepStatus(@Nonnull Map<Integer, PipelineStepStatusEnum> stepStatus) {
     try {
       this.stepStatusJson = JacksonUtils.write(stepStatus);
     } catch (JsonProcessingException e) {
@@ -235,7 +235,7 @@ public class AppBuildPipeline {
   @Nullable
   @JsonIgnore
   public <R extends BuildResult> R getBuildResult() {
-    PipelineType pipeType = getPipeType();
+    PipelineTypeEnum pipeType = getPipeType();
     if (pipeType.isUnknown() || buildResultJson == null) {
       return null;
     }
@@ -292,7 +292,7 @@ public class AppBuildPipeline {
     public static View of(@Nonnull AppBuildPipeline pipe) {
       // combine step info
       Map<Integer, String> stepDesc = pipe.getPipeType().getSteps();
-      Map<Integer, PipelineStepStatus> stepStatus = pipe.getStepStatus();
+      Map<Integer, PipelineStepStatusEnum> stepStatus = pipe.getStepStatus();
       Map<Integer, Long> stepTs = pipe.getStepStatusTimestamp();
       List<Step> steps = new ArrayList<>(stepDesc.size());
       for (int i = 1; i <= pipe.getPipeType().getSteps().size(); i++) {
@@ -300,7 +300,7 @@ public class AppBuildPipeline {
             new Step()
                 .setSeq(i)
                 .setDesc(stepDesc.getOrDefault(i, "unknown step"))
-                .setStatus(stepStatus.getOrDefault(i, PipelineStepStatus.unknown).getCode());
+                .setStatus(stepStatus.getOrDefault(i, PipelineStepStatusEnum.unknown).getCode());
         Long st = stepTs.get(i);
         if (st != null) {
           step.setTs(new Date(st));
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Application.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Application.java
index 2e8a06b61..b92c3dd62 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Application.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Application.java
@@ -20,18 +20,18 @@ package org.apache.streampark.console.core.entity;
 import org.apache.streampark.common.conf.ConfigConst;
 import org.apache.streampark.common.conf.K8sFlinkConfig;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.ApplicationType;
-import org.apache.streampark.common.enums.DevelopmentMode;
-import org.apache.streampark.common.enums.ExecutionMode;
-import org.apache.streampark.common.enums.FlinkK8sRestExposedType;
-import org.apache.streampark.common.enums.StorageType;
+import org.apache.streampark.common.enums.ApplicationTypeEnum;
+import org.apache.streampark.common.enums.DevelopmentModeEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
+import org.apache.streampark.common.enums.FlinkK8sRestExposedTypeEnum;
+import org.apache.streampark.common.enums.StorageTypeEnum;
 import org.apache.streampark.common.fs.FsOperator;
 import org.apache.streampark.console.base.util.JacksonUtils;
 import org.apache.streampark.console.core.bean.AppControl;
 import org.apache.streampark.console.core.bean.Dependency;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.ReleaseState;
-import org.apache.streampark.console.core.enums.ResourceFrom;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
+import org.apache.streampark.console.core.enums.ResourceFromEnum;
 import org.apache.streampark.console.core.metrics.flink.JobsOverview;
 import org.apache.streampark.console.core.utils.YarnQueueLabelExpression;
 import org.apache.streampark.flink.kubernetes.model.K8sPodTemplates;
@@ -276,8 +276,8 @@ public class Application implements Serializable {
   }
 
   public void setYarnQueueByHotParams() {
-    if (!(ExecutionMode.YARN_APPLICATION == this.getExecutionModeEnum()
-        || ExecutionMode.YARN_PER_JOB == this.getExecutionModeEnum())) {
+    if (!(ExecutionModeEnum.YARN_APPLICATION == this.getExecutionModeEnum()
+        || ExecutionModeEnum.YARN_PER_JOB == this.getExecutionModeEnum())) {
       return;
     }
 
@@ -336,28 +336,28 @@ public class Application implements Serializable {
   }
 
   @JsonIgnore
-  public ReleaseState getReleaseState() {
-    return ReleaseState.of(release);
+  public ReleaseStateEnum getReleaseState() {
+    return ReleaseStateEnum.of(release);
   }
 
   @JsonIgnore
-  public DevelopmentMode getDevelopmentMode() {
-    return DevelopmentMode.of(jobType);
+  public DevelopmentModeEnum getDevelopmentMode() {
+    return DevelopmentModeEnum.of(jobType);
   }
 
   @JsonIgnore
-  public FlinkAppState getStateEnum() {
-    return FlinkAppState.of(state);
+  public FlinkAppStateEnum getStateEnum() {
+    return FlinkAppStateEnum.of(state);
   }
 
   @JsonIgnore
-  public FlinkK8sRestExposedType getK8sRestExposedTypeEnum() {
-    return FlinkK8sRestExposedType.of(this.k8sRestExposedType);
+  public FlinkK8sRestExposedTypeEnum getK8sRestExposedTypeEnum() {
+    return FlinkK8sRestExposedTypeEnum.of(this.k8sRestExposedType);
   }
 
   @JsonIgnore
-  public ExecutionMode getExecutionModeEnum() {
-    return ExecutionMode.of(executionMode);
+  public ExecutionModeEnum getExecutionModeEnum() {
+    return ExecutionModeEnum.of(executionMode);
   }
 
   public boolean cpFailedTrigger() {
@@ -423,8 +423,8 @@ public class Application implements Serializable {
   }
 
   @JsonIgnore
-  public ApplicationType getApplicationType() {
-    return ApplicationType.of(appType);
+  public ApplicationTypeEnum getApplicationType() {
+    return ApplicationTypeEnum.of(appType);
   }
 
   @JsonIgnore
@@ -441,40 +441,40 @@ public class Application implements Serializable {
 
   @JsonIgnore
   public boolean isFlinkSqlJob() {
-    return DevelopmentMode.FLINK_SQL.getMode().equals(this.getJobType());
+    return DevelopmentModeEnum.FLINK_SQL.getMode().equals(this.getJobType());
   }
 
   @JsonIgnore
   public boolean isFlinkSqlJobOrPyFlinkJob() {
-    return DevelopmentMode.FLINK_SQL.getMode().equals(this.getJobType())
-        || DevelopmentMode.PYFLINK.getMode().equals(this.getJobType());
+    return DevelopmentModeEnum.FLINK_SQL.getMode().equals(this.getJobType())
+        || DevelopmentModeEnum.PYFLINK.getMode().equals(this.getJobType());
   }
 
   @JsonIgnore
   public boolean isCustomCodeJob() {
-    return DevelopmentMode.CUSTOM_CODE.getMode().equals(this.getJobType());
+    return DevelopmentModeEnum.CUSTOM_CODE.getMode().equals(this.getJobType());
   }
 
   @JsonIgnore
   public boolean isCustomCodeOrPyFlinkJob() {
-    return DevelopmentMode.CUSTOM_CODE.getMode().equals(this.getJobType())
-        || DevelopmentMode.PYFLINK.getMode().equals(this.getJobType());
+    return DevelopmentModeEnum.CUSTOM_CODE.getMode().equals(this.getJobType())
+        || DevelopmentModeEnum.PYFLINK.getMode().equals(this.getJobType());
   }
 
   @JsonIgnore
   public boolean isUploadJob() {
     return isCustomCodeOrPyFlinkJob()
-        && ResourceFrom.UPLOAD.getValue().equals(this.getResourceFrom());
+        && ResourceFromEnum.UPLOAD.getValue().equals(this.getResourceFrom());
   }
 
   @JsonIgnore
   public boolean isCICDJob() {
     return isCustomCodeOrPyFlinkJob()
-        && ResourceFrom.CICD.getValue().equals(this.getResourceFrom());
+        && ResourceFromEnum.CICD.getValue().equals(this.getResourceFrom());
   }
 
   public boolean isStreamParkJob() {
-    return this.getAppType() == ApplicationType.STREAMPARK_FLINK.getType();
+    return this.getAppType() == ApplicationTypeEnum.STREAMPARK_FLINK.getType();
   }
 
   @JsonIgnore
@@ -490,12 +490,12 @@ public class Application implements Serializable {
 
   @JsonIgnore
   public boolean isRunning() {
-    return FlinkAppState.RUNNING.getValue() == this.getState();
+    return FlinkAppStateEnum.RUNNING.getValue() == this.getState();
   }
 
   @JsonIgnore
   public boolean isNeedRollback() {
-    return ReleaseState.NEED_ROLLBACK.get() == this.getRelease();
+    return ReleaseStateEnum.NEED_ROLLBACK.get() == this.getRelease();
   }
 
   @JsonIgnore
@@ -507,23 +507,23 @@ public class Application implements Serializable {
   }
 
   @JsonIgnore
-  public StorageType getStorageType() {
+  public StorageTypeEnum getStorageType() {
     return getStorageType(getExecutionMode());
   }
 
-  public static StorageType getStorageType(Integer execMode) {
-    ExecutionMode executionMode = ExecutionMode.of(execMode);
-    switch (Objects.requireNonNull(executionMode)) {
+  public static StorageTypeEnum getStorageType(Integer execMode) {
+    ExecutionModeEnum executionModeEnum = ExecutionModeEnum.of(execMode);
+    switch (Objects.requireNonNull(executionModeEnum)) {
       case YARN_APPLICATION:
-        return StorageType.HDFS;
+        return StorageTypeEnum.HDFS;
       case YARN_PER_JOB:
       case YARN_SESSION:
       case KUBERNETES_NATIVE_SESSION:
       case KUBERNETES_NATIVE_APPLICATION:
       case REMOTE:
-        return StorageType.LFS;
+        return StorageTypeEnum.LFS;
       default:
-        throw new UnsupportedOperationException("Unsupported ".concat(executionMode.getName()));
+        throw new UnsupportedOperationException("Unsupported ".concat(executionModeEnum.getName()));
     }
   }
 
@@ -559,7 +559,7 @@ public class Application implements Serializable {
     if (appParam != this) {
       this.hotParams = null;
     }
-    ExecutionMode executionModeEnum = appParam.getExecutionModeEnum();
+    ExecutionModeEnum executionModeEnum = appParam.getExecutionModeEnum();
     Map<String, String> hotParams = new HashMap<>(0);
     if (needFillYarnQueueLabel(executionModeEnum)) {
       hotParams.putAll(YarnQueueLabelExpression.getQueueLabelMap(appParam.getYarnQueue()));
@@ -569,8 +569,8 @@ public class Application implements Serializable {
     }
   }
 
-  private boolean needFillYarnQueueLabel(ExecutionMode mode) {
-    return ExecutionMode.YARN_PER_JOB == mode || ExecutionMode.YARN_APPLICATION == mode;
+  private boolean needFillYarnQueueLabel(ExecutionModeEnum mode) {
+    return ExecutionModeEnum.YARN_PER_JOB == mode || ExecutionModeEnum.YARN_APPLICATION == mode;
   }
 
   @Override
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/ApplicationConfig.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/ApplicationConfig.java
index 5d9fba53d..48a2f951e 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/ApplicationConfig.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/ApplicationConfig.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.entity;
 import org.apache.streampark.common.conf.ConfigConst;
 import org.apache.streampark.common.util.DeflaterUtils;
 import org.apache.streampark.common.util.PropertiesUtils;
-import org.apache.streampark.console.core.enums.ConfigFileType;
+import org.apache.streampark.console.core.enums.ConfigFileTypeEnum;
 
 import com.baomidou.mybatisplus.annotation.FieldStrategy;
 import com.baomidou.mybatisplus.annotation.IdType;
@@ -75,7 +75,7 @@ public class ApplicationConfig {
   }
 
   public Map<String, String> readConfig() {
-    ConfigFileType fileType = ConfigFileType.of(this.format);
+    ConfigFileTypeEnum fileType = ConfigFileTypeEnum.of(this.format);
     Map<String, String> configs = null;
     if (fileType != null) {
       switch (fileType) {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Effective.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Effective.java
index 35613076d..d5befb948 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Effective.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Effective.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.entity;
 
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 
 import com.baomidou.mybatisplus.annotation.IdType;
 import com.baomidou.mybatisplus.annotation.TableId;
@@ -45,11 +45,11 @@ public class Effective {
   private Long targetId;
   private Date createTime;
 
-  private transient EffectiveType effectiveType;
+  private transient EffectiveTypeEnum effectiveTypeEnum;
 
   public Effective() {}
 
-  public Effective(Long appId, EffectiveType type, Long targetId) {
+  public Effective(Long appId, EffectiveTypeEnum type, Long targetId) {
     this.appId = appId;
     this.targetType = type.getType();
     this.targetId = targetId;
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/FlinkCluster.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/FlinkCluster.java
index a09727e2c..0c3ef209a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/FlinkCluster.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/FlinkCluster.java
@@ -18,10 +18,10 @@
 package org.apache.streampark.console.core.entity;
 
 import org.apache.streampark.common.conf.ConfigConst;
-import org.apache.streampark.common.enums.ClusterState;
-import org.apache.streampark.common.enums.ExecutionMode;
-import org.apache.streampark.common.enums.FlinkK8sRestExposedType;
-import org.apache.streampark.common.enums.ResolveOrder;
+import org.apache.streampark.common.enums.ClusterStateEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
+import org.apache.streampark.common.enums.FlinkK8sRestExposedTypeEnum;
+import org.apache.streampark.common.enums.ResolveOrderEnum;
 import org.apache.streampark.common.util.HttpClientUtils;
 import org.apache.streampark.common.util.PropertiesUtils;
 import org.apache.streampark.console.base.util.JacksonUtils;
@@ -118,16 +118,16 @@ public class FlinkCluster implements Serializable {
   private transient Integer affectedJobs = 0;
 
   @JsonIgnore
-  public FlinkK8sRestExposedType getK8sRestExposedTypeEnum() {
-    return FlinkK8sRestExposedType.of(this.k8sRestExposedType);
+  public FlinkK8sRestExposedTypeEnum getK8sRestExposedTypeEnum() {
+    return FlinkK8sRestExposedTypeEnum.of(this.k8sRestExposedType);
   }
 
-  public ExecutionMode getExecutionModeEnum() {
-    return ExecutionMode.of(this.executionMode);
+  public ExecutionModeEnum getExecutionModeEnum() {
+    return ExecutionModeEnum.of(this.executionMode);
   }
 
-  public ClusterState getClusterStateEnum() {
-    return ClusterState.of(this.clusterState);
+  public ClusterStateEnum getClusterStateEnum() {
+    return ClusterStateEnum.of(this.clusterState);
   }
 
   @JsonIgnore
@@ -137,7 +137,7 @@ public class FlinkCluster implements Serializable {
       return Collections.emptyMap();
     }
     Map<String, Object> map = JacksonUtils.read(this.options, Map.class);
-    if (ExecutionMode.YARN_SESSION == getExecutionModeEnum()) {
+    if (ExecutionModeEnum.YARN_SESSION == getExecutionModeEnum()) {
       map.put(ConfigConst.KEY_YARN_APP_NAME(), this.clusterName);
       map.putAll(YarnQueueLabelExpression.getQueueLabelMap(yarnQueue));
     }
@@ -181,9 +181,9 @@ public class FlinkCluster implements Serializable {
         PropertiesUtils.extractDynamicPropertiesAsJava(this.getDynamicProperties());
     map.putAll(this.getOptionMap());
     map.putAll(dynamicProperties);
-    ResolveOrder resolveOrder = ResolveOrder.of(this.getResolveOrder());
-    if (resolveOrder != null) {
-      map.put(CoreOptions.CLASSLOADER_RESOLVE_ORDER.key(), resolveOrder.getName());
+    ResolveOrderEnum resolveOrderEnum = ResolveOrderEnum.of(this.getResolveOrder());
+    if (resolveOrderEnum != null) {
+      map.put(CoreOptions.CLASSLOADER_RESOLVE_ORDER.key(), resolveOrderEnum.getName());
     }
     return map;
   }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Message.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Message.java
index ce0087637..e56b86c8a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Message.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Message.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.entity;
 
-import org.apache.streampark.console.core.enums.NoticeType;
+import org.apache.streampark.console.core.enums.NoticeTypeEnum;
 
 import com.baomidou.mybatisplus.annotation.IdType;
 import com.baomidou.mybatisplus.annotation.TableId;
@@ -50,12 +50,13 @@ public class Message {
 
   public Message() {}
 
-  public Message(Long userId, Long appId, String title, String context, NoticeType noticeType) {
+  public Message(
+      Long userId, Long appId, String title, String context, NoticeTypeEnum noticeTypeEnum) {
     this.userId = userId;
     this.appId = appId;
     this.title = title;
     this.context = context;
-    this.type = noticeType.get();
+    this.type = noticeTypeEnum.get();
     this.createTime = new Date();
     this.isRead = false;
   }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Project.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Project.java
index 98b2ff7f9..fabdfdcb3 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Project.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Project.java
@@ -25,7 +25,7 @@ import org.apache.streampark.console.base.exception.ApiDetailException;
 import org.apache.streampark.console.base.util.CommonUtils;
 import org.apache.streampark.console.base.util.GitUtils;
 import org.apache.streampark.console.base.util.WebUtils;
-import org.apache.streampark.console.core.enums.GitAuthorizedError;
+import org.apache.streampark.console.core.enums.GitAuthorizedErrorEnum;
 
 import org.apache.commons.io.FileUtils;
 import org.apache.commons.lang3.StringUtils;
@@ -150,18 +150,18 @@ public class Project implements Serializable {
     }
   }
 
-  public GitAuthorizedError gitCheck() {
+  public GitAuthorizedErrorEnum gitCheck() {
     try {
       GitUtils.getBranchList(this);
-      return GitAuthorizedError.SUCCESS;
+      return GitAuthorizedErrorEnum.SUCCESS;
     } catch (Exception e) {
       String err = e.getMessage();
       if (err.contains("not authorized")) {
-        return GitAuthorizedError.ERROR;
+        return GitAuthorizedErrorEnum.ERROR;
       } else if (err.contains("Authentication is required")) {
-        return GitAuthorizedError.REQUIRED;
+        return GitAuthorizedErrorEnum.REQUIRED;
       }
-      return GitAuthorizedError.UNKNOW;
+      return GitAuthorizedErrorEnum.UNKNOW;
     }
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Resource.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Resource.java
index e02661f1a..87903a486 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Resource.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/entity/Resource.java
@@ -17,8 +17,8 @@
 
 package org.apache.streampark.console.core.entity;
 
-import org.apache.streampark.console.core.enums.EngineType;
-import org.apache.streampark.console.core.enums.ResourceType;
+import org.apache.streampark.console.core.enums.EngineTypeEnum;
+import org.apache.streampark.console.core.enums.ResourceTypeEnum;
 
 import com.baomidou.mybatisplus.annotation.IdType;
 import com.baomidou.mybatisplus.annotation.TableId;
@@ -54,9 +54,9 @@ public class Resource implements Serializable {
   /** user id of creator */
   private Long creatorId;
 
-  private ResourceType resourceType;
+  private ResourceTypeEnum resourceTypeEnum;
 
-  private EngineType engineType;
+  private EngineTypeEnum engineTypeEnum;
 
   // for flink app
   private String mainClass;
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenStateEnum.java
similarity index 93%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenStateEnum.java
index b2a3735a7..823684495 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AccessTokenStateEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.enums;
 
-public enum AccessTokenState {
+public enum AccessTokenStateEnum {
 
   /** not added token */
   NULL(0),
@@ -30,7 +30,7 @@ public enum AccessTokenState {
 
   private final int value;
 
-  AccessTokenState(int value) {
+  AccessTokenStateEnum(int value) {
     this.value = value;
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertTypeEnum.java
similarity index 79%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertTypeEnum.java
index b63c8495d..6756b8aad 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AlertTypeEnum.java
@@ -36,7 +36,7 @@ import java.util.Map;
 
 /** The AlertType enum represents different types of alerts that can be used for notifications. */
 @Getter
-public enum AlertType {
+public enum AlertTypeEnum {
 
   /** Email */
   EMAIL(1, EmailAlertNotifyServiceImpl.class),
@@ -63,27 +63,27 @@ public enum AlertType {
   private final Class<? extends AlertNotifyService> clazz;
 
   /** A cache map used to quickly get the alert type from an integer code */
-  private static final Map<Integer, AlertType> CACHE_MAP = createCacheMap();
+  private static final Map<Integer, AlertTypeEnum> CACHE_MAP = createCacheMap();
 
-  private static Map<Integer, AlertType> createCacheMap() {
-    Map<Integer, AlertType> map = new HashMap<>();
-    for (AlertType notifyType : AlertType.values()) {
+  private static Map<Integer, AlertTypeEnum> createCacheMap() {
+    Map<Integer, AlertTypeEnum> map = new HashMap<>();
+    for (AlertTypeEnum notifyType : AlertTypeEnum.values()) {
       map.put(notifyType.code, notifyType);
     }
     return Collections.unmodifiableMap(map);
   }
 
-  AlertType(Integer code, Class<? extends AlertNotifyService> clazz) {
+  AlertTypeEnum(Integer code, Class<? extends AlertNotifyService> clazz) {
     this.code = code;
     this.clazz = clazz;
   }
 
-  public static List<AlertType> decode(Integer level) {
+  public static List<AlertTypeEnum> decode(Integer level) {
     if (level == null) {
       level = EMPTY_LEVEL;
     }
 
-    List<AlertType> result = new ArrayList<>(AlertType.values().length);
+    List<AlertTypeEnum> result = new ArrayList<>(AlertTypeEnum.values().length);
     while (level != 0) {
       int code = level & -level;
       result.add(getByCode(code));
@@ -92,19 +92,19 @@ public enum AlertType {
     return result;
   }
 
-  public static int encode(List<AlertType> alertTypes) {
-    if (CollectionUtils.isEmpty(alertTypes)) {
+  public static int encode(List<AlertTypeEnum> alertTypeEnums) {
+    if (CollectionUtils.isEmpty(alertTypeEnums)) {
       return EMPTY_LEVEL;
     }
 
     int result = 0;
-    for (AlertType alertType : alertTypes) {
-      result |= alertType.code;
+    for (AlertTypeEnum alertTypeEnum : alertTypeEnums) {
+      result |= alertTypeEnum.code;
     }
     return result;
   }
 
-  private static AlertType getByCode(Integer code) {
+  private static AlertTypeEnum getByCode(Integer code) {
     return CACHE_MAP.get(code);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsStateEnum.java
similarity index 94%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsStateEnum.java
index 707a60ac6..cc6f570dd 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/AppExistsStateEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.enums;
 
-public enum AppExistsState {
+public enum AppExistsStateEnum {
 
   /** no exists */
   NO(0),
@@ -36,7 +36,7 @@ public enum AppExistsState {
 
   private final int value;
 
-  AppExistsState(int value) {
+  AppExistsStateEnum(int value) {
     this.value = value;
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildStateEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildStateEnum.java
index 2229a8b36..f1de52ac5 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/BuildStateEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum BuildState {
+public enum BuildStateEnum {
 
   /** has changed, need rebuild */
   NEED_REBUILD(-2),
@@ -37,7 +37,7 @@ public enum BuildState {
 
   private final int value;
 
-  BuildState(int value) {
+  BuildStateEnum(int value) {
     this.value = value;
   }
 
@@ -45,7 +45,7 @@ public enum BuildState {
     return this.value;
   }
 
-  public static BuildState of(Integer state) {
+  public static BuildStateEnum of(Integer state) {
     return Arrays.stream(values()).filter((x) -> x.value == state).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateTypeEnum.java
similarity index 91%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateTypeEnum.java
index d4693d84d..95cfed241 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CandidateTypeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum CandidateType {
+public enum CandidateTypeEnum {
 
   /** non candidate */
   NONE(0),
@@ -32,7 +32,7 @@ public enum CandidateType {
 
   private final int value;
 
-  CandidateType(int value) {
+  CandidateTypeEnum(int value) {
     this.value = value;
   }
 
@@ -40,7 +40,7 @@ public enum CandidateType {
     return this.value;
   }
 
-  public static CandidateType of(Integer value) {
+  public static CandidateTypeEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatus.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatusEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatus.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatusEnum.java
index f065b864a..fd11e9d94 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatus.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointStatusEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum CheckPointStatus {
+public enum CheckPointStatusEnum {
   /** IN_PROGRESS */
   IN_PROGRESS(1),
   /** COMPLETED */
@@ -34,11 +34,11 @@ public enum CheckPointStatus {
     return this.value;
   }
 
-  CheckPointStatus(int value) {
+  CheckPointStatusEnum(int value) {
     this.value = value;
   }
 
-  public static CheckPointStatus of(Integer value) {
+  public static CheckPointStatusEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointTypeEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointTypeEnum.java
index e4a25e94c..15e3448f9 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/CheckPointTypeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum CheckPointType {
+public enum CheckPointTypeEnum {
   /** CHECKPOINT */
   CHECKPOINT(1),
   /** SAVEPOINT */
@@ -33,11 +33,11 @@ public enum CheckPointType {
     return this.value;
   }
 
-  CheckPointType(int value) {
+  CheckPointTypeEnum(int value) {
     this.value = value;
   }
 
-  public static CheckPointType of(Integer value) {
+  public static CheckPointTypeEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileTypeEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileTypeEnum.java
index 9cd62bf7f..aeba51f18 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ConfigFileTypeEnum.java
@@ -23,7 +23,7 @@ import java.util.Arrays;
 
 /** configFile Type enum */
 @Getter
-public enum ConfigFileType {
+public enum ConfigFileTypeEnum {
   YAML(1, "yaml"),
 
   PROPERTIES(2, "prop"),
@@ -35,12 +35,12 @@ public enum ConfigFileType {
   private final int value;
   private final String typeName;
 
-  ConfigFileType(int value, String name) {
+  ConfigFileTypeEnum(int value, String name) {
     this.value = value;
     this.typeName = name;
   }
 
-  public static ConfigFileType of(Integer value) {
+  public static ConfigFileTypeEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveTypeEnum.java
similarity index 93%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveTypeEnum.java
index 6de934c00..bba8c9d8a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EffectiveTypeEnum.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.enums;
 import lombok.Getter;
 
 @Getter
-public enum EffectiveType {
+public enum EffectiveTypeEnum {
   /** config */
   CONFIG(1),
   /** FLINKSQL */
@@ -28,7 +28,7 @@ public enum EffectiveType {
 
   private final int type;
 
-  EffectiveType(int value) {
+  EffectiveTypeEnum(int value) {
     this.type = value;
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineTypeEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineTypeEnum.java
index d0310e37b..a2eecb3d1 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/EngineTypeEnum.java
@@ -24,7 +24,7 @@ import java.util.Arrays;
 
 /** Compute engine type. */
 @Getter
-public enum EngineType {
+public enum EngineTypeEnum {
 
   /** Apache Flink: activated by default */
   FLINK(0),
@@ -34,11 +34,11 @@ public enum EngineType {
 
   @EnumValue private final int code;
 
-  EngineType(int code) {
+  EngineTypeEnum(int code) {
     this.code = code;
   }
 
-  public static EngineType of(Integer code) {
+  public static EngineTypeEnum of(Integer code) {
     return Arrays.stream(values()).filter((x) -> x.code == code).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategy.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategyEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategy.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategyEnum.java
index 9c4f8ad20..a1f2be11a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategy.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FailoverStrategyEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum FailoverStrategy {
+public enum FailoverStrategyEnum {
 
   /** send alert */
   ALERT(1),
@@ -29,7 +29,7 @@ public enum FailoverStrategy {
 
   private final int value;
 
-  FailoverStrategy(int value) {
+  FailoverStrategyEnum(int value) {
     this.value = value;
   }
 
@@ -37,7 +37,7 @@ public enum FailoverStrategy {
     return this.value;
   }
 
-  public static FailoverStrategy of(Integer value) {
+  public static FailoverStrategyEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppStateEnum.java
similarity index 74%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppStateEnum.java
index 4875d5969..77963c321 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/FlinkAppStateEnum.java
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.console.core.enums;
 
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState;
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum;
 
 import lombok.Getter;
 
 import scala.Enumeration;
 
 @Getter
-public enum FlinkAppState {
+public enum FlinkAppStateEnum {
 
   /** Added new job to database. */
   ADDED(0),
@@ -107,42 +107,42 @@ public enum FlinkAppState {
 
   private final int value;
 
-  FlinkAppState(int value) {
+  FlinkAppStateEnum(int value) {
     this.value = value;
   }
 
-  public static FlinkAppState of(Integer state) {
-    for (FlinkAppState appState : values()) {
+  public static FlinkAppStateEnum of(Integer state) {
+    for (FlinkAppStateEnum appState : values()) {
       if (appState.value == state) {
         return appState;
       }
     }
-    return FlinkAppState.OTHER;
+    return FlinkAppStateEnum.OTHER;
   }
 
-  public static FlinkAppState of(String name) {
-    for (FlinkAppState appState : values()) {
+  public static FlinkAppStateEnum of(String name) {
+    for (FlinkAppStateEnum appState : values()) {
       if (appState.name().equals(name)) {
         return appState;
       }
     }
-    return FlinkAppState.OTHER;
+    return FlinkAppStateEnum.OTHER;
   }
 
   public static boolean isEndState(Integer appState) {
-    FlinkAppState flinkAppState = FlinkAppState.of(appState);
-    return FlinkAppState.CANCELED == flinkAppState
-        || FlinkAppState.FAILED == flinkAppState
-        || FlinkAppState.KILLED == flinkAppState
-        || FlinkAppState.FINISHED == flinkAppState
-        || FlinkAppState.SUCCEEDED == flinkAppState
-        || FlinkAppState.LOST == flinkAppState
-        || FlinkAppState.TERMINATED == flinkAppState;
+    FlinkAppStateEnum flinkAppStateEnum = FlinkAppStateEnum.of(appState);
+    return FlinkAppStateEnum.CANCELED == flinkAppStateEnum
+        || FlinkAppStateEnum.FAILED == flinkAppStateEnum
+        || FlinkAppStateEnum.KILLED == flinkAppStateEnum
+        || FlinkAppStateEnum.FINISHED == flinkAppStateEnum
+        || FlinkAppStateEnum.SUCCEEDED == flinkAppStateEnum
+        || FlinkAppStateEnum.LOST == flinkAppStateEnum
+        || FlinkAppStateEnum.TERMINATED == flinkAppStateEnum;
   }
 
   public static boolean isLost(Integer appState) {
-    FlinkAppState flinkAppState = FlinkAppState.of(appState);
-    return FlinkAppState.LOST == flinkAppState;
+    FlinkAppStateEnum flinkAppStateEnum = FlinkAppStateEnum.of(appState);
+    return FlinkAppStateEnum.LOST == flinkAppStateEnum;
   }
 
   /**
@@ -152,8 +152,8 @@ public enum FlinkAppState {
   @Deprecated
   public static class Bridge {
     /** covert from org.apache.streampark.flink.k8s.enums.FlinkJobState */
-    public static FlinkAppState fromK8sFlinkJobState(Enumeration.Value flinkJobState) {
-      if (FlinkJobState.K8S_INITIALIZING() == flinkJobState) {
+    public static FlinkAppStateEnum fromK8sFlinkJobState(Enumeration.Value flinkJobState) {
+      if (FlinkJobStateEnum.K8S_INITIALIZING() == flinkJobState) {
         return INITIALIZING;
       } else {
         return of(flinkJobState.toString());
@@ -161,8 +161,8 @@ public enum FlinkAppState {
     }
 
     /** covert to org.apache.streampark.flink.k8s.enums.FlinkJobState */
-    public static Enumeration.Value toK8sFlinkJobState(FlinkAppState flinkAppState) {
-      return FlinkJobState.of(flinkAppState.name());
+    public static Enumeration.Value toK8sFlinkJobState(FlinkAppStateEnum flinkAppStateEnum) {
+      return FlinkJobStateEnum.of(flinkAppStateEnum.name());
     }
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedError.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedErrorEnum.java
similarity index 86%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedError.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedErrorEnum.java
index de45a7445..9ae2ddc05 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedError.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitAuthorizedErrorEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.enums;
 
-public enum GitAuthorizedError {
+public enum GitAuthorizedErrorEnum {
 
   /** Success. */
   SUCCESS(0),
@@ -33,12 +33,12 @@ public enum GitAuthorizedError {
 
   private final int value;
 
-  GitAuthorizedError(int value) {
+  GitAuthorizedErrorEnum(int value) {
     this.value = value;
   }
 
-  public static GitAuthorizedError of(Integer state) {
-    for (GitAuthorizedError error : values()) {
+  public static GitAuthorizedErrorEnum of(Integer state) {
+    for (GitAuthorizedErrorEnum error : values()) {
       if (error.value == state) {
         return error;
       }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredential.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredentialEnum.java
similarity index 85%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredential.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredentialEnum.java
index 748077ff3..07e892043 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredential.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/GitCredentialEnum.java
@@ -19,22 +19,22 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum GitCredential {
+public enum GitCredentialEnum {
   HTTPS(1),
   SSH(2);
 
   private final int value;
 
-  GitCredential(int value) {
+  GitCredentialEnum(int value) {
     this.value = value;
   }
 
-  public static GitCredential of(Integer value) {
+  public static GitCredentialEnum of(Integer value) {
     return Arrays.stream(values()).filter(x -> x.value == value).findFirst().orElse(null);
   }
 
   public static boolean isSSH(Integer gitCredential) {
-    return GitCredential.SSH == GitCredential.of(gitCredential);
+    return GitCredentialEnum.SSH == GitCredentialEnum.of(gitCredential);
   }
 
   public Integer getValue() {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginTypeEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginTypeEnum.java
index 3b4f7c796..20e2ae360 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/LoginTypeEnum.java
@@ -25,7 +25,7 @@ import java.util.Objects;
 
 /** The user login type. */
 @Getter
-public enum LoginType {
+public enum LoginTypeEnum {
 
   /** sign in with password */
   PASSWORD(0),
@@ -38,15 +38,15 @@ public enum LoginType {
 
   @EnumValue private final int code;
 
-  LoginType(int code) {
+  LoginTypeEnum(int code) {
     this.code = code;
   }
 
-  public static LoginType of(Integer code) {
+  public static LoginTypeEnum of(Integer code) {
     return Arrays.stream(values()).filter((x) -> x.code == code).findFirst().orElse(null);
   }
 
-  public static LoginType of(String loginType) {
+  public static LoginTypeEnum of(String loginType) {
     return Arrays.stream(values())
         .filter((x) -> Objects.equals(x.toString(), loginType))
         .findFirst()
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeTypeEnum.java
similarity index 91%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeTypeEnum.java
index 29f67cc0b..5d6b22839 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/NoticeTypeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum NoticeType {
+public enum NoticeTypeEnum {
   /** exception */
   EXCEPTION(1),
   /** message */
@@ -31,11 +31,11 @@ public enum NoticeType {
     return this.value;
   }
 
-  NoticeType(int value) {
+  NoticeTypeEnum(int value) {
     this.value = value;
   }
 
-  public static NoticeType of(Integer value) {
+  public static NoticeTypeEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/Operation.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OperationEnum.java
similarity index 91%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/Operation.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OperationEnum.java
index ba6cee3dd..fa559eb4a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/Operation.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OperationEnum.java
@@ -22,7 +22,7 @@ import lombok.Getter;
 import java.util.Arrays;
 
 @Getter
-public enum Operation {
+public enum OperationEnum {
   RELEASE(0),
   START(1),
   SAVEPOINT(2),
@@ -30,11 +30,11 @@ public enum Operation {
 
   private final int value;
 
-  Operation(int value) {
+  OperationEnum(int value) {
     this.value = value;
   }
 
-  public static Operation of(Integer option) {
+  public static OperationEnum of(Integer option) {
     return Arrays.stream(values()).filter((x) -> x.value == option).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionStateEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionStateEnum.java
index 500aefe4d..9a6b8ca1a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/OptionStateEnum.java
@@ -22,7 +22,7 @@ import lombok.Getter;
 import java.util.Arrays;
 
 @Getter
-public enum OptionState {
+public enum OptionStateEnum {
 
   /** Application which is currently action: none. */
   NONE(0),
@@ -39,11 +39,11 @@ public enum OptionState {
 
   private final int value;
 
-  OptionState(int value) {
+  OptionStateEnum(int value) {
     this.value = value;
   }
 
-  public static OptionState of(Integer state) {
+  public static OptionStateEnum of(Integer state) {
     return Arrays.stream(values()).filter((x) -> x.value == state).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionTypeEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionTypeEnum.java
index 2609e9f5f..989fd5baa 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PermissionTypeEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum PermissionType {
+public enum PermissionTypeEnum {
   USER(1),
   TEAM(2),
   APP(3);
@@ -30,11 +30,11 @@ public enum PermissionType {
     return this.value;
   }
 
-  PermissionType(int value) {
+  PermissionTypeEnum(int value) {
     this.value = value;
   }
 
-  public static PermissionType of(Integer value) {
+  public static PermissionTypeEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value == value).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderTypeEnum.java
similarity index 93%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderTypeEnum.java
index b3804124a..971e391a2 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/PlaceholderTypeEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.console.core.enums;
 
 /** configFile Type enum */
-public enum PlaceholderType {
+public enum PlaceholderTypeEnum {
   JOB_ID("job_id"),
 
   JOB_NAME("job_name"),
@@ -27,7 +27,7 @@ public enum PlaceholderType {
 
   private final String name;
 
-  PlaceholderType(String name) {
+  PlaceholderTypeEnum(String name) {
     this.name = name;
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseState.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseStateEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseState.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseStateEnum.java
index ad47e1fa0..db092156c 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseState.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ReleaseStateEnum.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.enums;
 
 import java.util.Arrays;
 
-public enum ReleaseState {
+public enum ReleaseStateEnum {
 
   /** release failed */
   FAILED(-1),
@@ -46,7 +46,7 @@ public enum ReleaseState {
 
   private final int value;
 
-  ReleaseState(int value) {
+  ReleaseStateEnum(int value) {
     this.value = value;
   }
 
@@ -54,7 +54,7 @@ public enum ReleaseState {
     return this.value;
   }
 
-  public static ReleaseState of(Integer state) {
+  public static ReleaseStateEnum of(Integer state) {
     return Arrays.stream(values()).filter((x) -> x.value == state).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFrom.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFromEnum.java
similarity index 90%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFrom.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFromEnum.java
index d2e08addb..03ce3787e 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFrom.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceFromEnum.java
@@ -22,7 +22,7 @@ import lombok.Getter;
 import java.util.Arrays;
 
 @Getter
-public enum ResourceFrom {
+public enum ResourceFromEnum {
 
   /** cicd(build from cvs) */
   CICD(1),
@@ -32,11 +32,11 @@ public enum ResourceFrom {
 
   private final Integer value;
 
-  ResourceFrom(Integer value) {
+  ResourceFromEnum(Integer value) {
     this.value = value;
   }
 
-  public static ResourceFrom of(Integer value) {
+  public static ResourceFromEnum of(Integer value) {
     return Arrays.stream(values()).filter((x) -> x.value.equals(value)).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceTypeEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceTypeEnum.java
index 2a000a448..cd8a20b81 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/ResourceTypeEnum.java
@@ -24,7 +24,7 @@ import java.util.Arrays;
 
 /** The resource type. */
 @Getter
-public enum ResourceType {
+public enum ResourceTypeEnum {
 
   /** Flink application */
   FLINK_APP(0),
@@ -43,11 +43,11 @@ public enum ResourceType {
 
   @EnumValue private final int code;
 
-  ResourceType(int code) {
+  ResourceTypeEnum(int code) {
     this.code = code;
   }
 
-  public static ResourceType of(Integer code) {
+  public static ResourceTypeEnum of(Integer code) {
     return Arrays.stream(values()).filter((x) -> x.code == code).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFrom.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFromEnum.java
similarity index 97%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFrom.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFromEnum.java
index 45e99a7b7..a77f13107 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFrom.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/StopFromEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.enums;
 
-public enum StopFrom {
+public enum StopFromEnum {
   /** None */
   NONE,
   /** StreamPark */
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserType.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserTypeEnum.java
similarity index 92%
rename from streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserType.java
rename to streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserTypeEnum.java
index 73e8a07b1..29cfe74a2 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserType.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/enums/UserTypeEnum.java
@@ -24,7 +24,7 @@ import java.util.Arrays;
 
 /** The user type. */
 @Getter
-public enum UserType {
+public enum UserTypeEnum {
 
   /** The admin of StreamPark. */
   ADMIN(1),
@@ -34,11 +34,11 @@ public enum UserType {
 
   @EnumValue private final int code;
 
-  UserType(int code) {
+  UserTypeEnum(int code) {
     this.code = code;
   }
 
-  public static UserType of(Integer code) {
+  public static UserTypeEnum of(Integer code) {
     return Arrays.stream(values()).filter((x) -> x.code == code).findFirst().orElse(null);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/metrics/flink/CheckPoints.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/metrics/flink/CheckPoints.java
index b92fc4059..812d84ea3 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/metrics/flink/CheckPoints.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/metrics/flink/CheckPoints.java
@@ -17,8 +17,8 @@
 
 package org.apache.streampark.console.core.metrics.flink;
 
-import org.apache.streampark.console.core.enums.CheckPointStatus;
-import org.apache.streampark.console.core.enums.CheckPointType;
+import org.apache.streampark.console.core.enums.CheckPointStatusEnum;
+import org.apache.streampark.console.core.enums.CheckPointTypeEnum;
 
 import com.fasterxml.jackson.annotation.JsonIgnore;
 import com.fasterxml.jackson.annotation.JsonProperty;
@@ -73,17 +73,17 @@ public class CheckPoints implements Serializable {
 
     private Boolean discarded;
 
-    public CheckPointStatus getCheckPointStatus() {
-      return CheckPointStatus.valueOf(this.status);
+    public CheckPointStatusEnum getCheckPointStatus() {
+      return CheckPointStatusEnum.valueOf(this.status);
     }
 
-    public CheckPointType getCheckPointType() {
+    public CheckPointTypeEnum getCheckPointType() {
       if ("CHECKPOINT".equals(this.checkpointType)) {
-        return CheckPointType.CHECKPOINT;
+        return CheckPointTypeEnum.CHECKPOINT;
       } else if ("SAVEPOINT".equals(this.checkpointType)) {
-        return CheckPointType.SAVEPOINT;
+        return CheckPointTypeEnum.SAVEPOINT;
       }
-      return CheckPointType.SYNC_SAVEPOINT;
+      return CheckPointTypeEnum.SYNC_SAVEPOINT;
     }
 
     public String getPath() {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/EnvInitializer.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/EnvInitializer.java
index c4cb68a27..e3f685a2f 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/EnvInitializer.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/runner/EnvInitializer.java
@@ -22,7 +22,7 @@ import org.apache.streampark.common.conf.ConfigConst;
 import org.apache.streampark.common.conf.InternalConfigHolder;
 import org.apache.streampark.common.conf.InternalOption;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.StorageType;
+import org.apache.streampark.common.enums.StorageTypeEnum;
 import org.apache.streampark.common.fs.FsOperator;
 import org.apache.streampark.common.util.SystemPropertyUtils;
 import org.apache.streampark.common.util.Utils;
@@ -53,7 +53,7 @@ import java.util.Set;
 import java.util.regex.Matcher;
 import java.util.regex.Pattern;
 
-import static org.apache.streampark.common.enums.StorageType.LFS;
+import static org.apache.streampark.common.enums.StorageTypeEnum.LFS;
 
 @Order(1)
 @Slf4j
@@ -64,7 +64,7 @@ public class EnvInitializer implements ApplicationRunner {
 
   @Autowired private SettingService settingService;
 
-  private final Set<StorageType> initialized = new HashSet<>(2);
+  private final Set<StorageTypeEnum> initialized = new HashSet<>(2);
 
   private final FileFilter fileFilter = p -> !".gitkeep".equals(p.getName());
 
@@ -126,17 +126,17 @@ public class EnvInitializer implements ApplicationRunner {
     SystemPropertyUtils.set(key, value);
   }
 
-  public synchronized void storageInitialize(StorageType storageType) {
+  public synchronized void storageInitialize(StorageTypeEnum storageTypeEnum) {
 
-    if (initialized.contains(storageType)) {
+    if (initialized.contains(storageTypeEnum)) {
       return;
     }
 
-    FsOperator fsOperator = FsOperator.of(storageType);
-    Workspace workspace = Workspace.of(storageType);
+    FsOperator fsOperator = FsOperator.of(storageTypeEnum);
+    Workspace workspace = Workspace.of(storageTypeEnum);
 
     // 1. prepare workspace dir
-    if (LFS == storageType) {
+    if (LFS == storageTypeEnum) {
       fsOperator.mkdirsIfNotExists(Workspace.APP_LOCAL_DIST());
     }
     Arrays.asList(
@@ -198,18 +198,18 @@ public class EnvInitializer implements ApplicationRunner {
       FsOperator.lfs().mkdirs(localMavenRepo);
     }
 
-    initialized.add(storageType);
+    initialized.add(storageTypeEnum);
   }
 
-  public void checkFlinkEnv(StorageType storageType, FlinkEnv flinkEnv) throws IOException {
+  public void checkFlinkEnv(StorageTypeEnum storageTypeEnum, FlinkEnv flinkEnv) throws IOException {
     String flinkLocalHome = flinkEnv.getFlinkHome();
     if (flinkLocalHome == null) {
       throw new ExceptionInInitializerError(
           "[StreamPark] FLINK_HOME is undefined,Make sure that Flink is installed.");
     }
-    Workspace workspace = Workspace.of(storageType);
+    Workspace workspace = Workspace.of(storageTypeEnum);
     String appFlink = workspace.APP_FLINK();
-    FsOperator fsOperator = FsOperator.of(storageType);
+    FsOperator fsOperator = FsOperator.of(storageTypeEnum);
     if (!fsOperator.exists(appFlink)) {
       log.info("checkFlinkEnv, now mkdir [{}] starting ...", appFlink);
       fsOperator.mkdirs(appFlink);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/AppBuildPipeService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/AppBuildPipeService.java
index 4b455b5c5..531626a39 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/AppBuildPipeService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/AppBuildPipeService.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.service;
 
 import org.apache.streampark.console.core.entity.AppBuildPipeline;
 import org.apache.streampark.flink.packer.pipeline.DockerResolvedSnapshot;
-import org.apache.streampark.flink.packer.pipeline.PipelineStatus;
+import org.apache.streampark.flink.packer.pipeline.PipelineStatusEnum;
 
 import com.baomidou.mybatisplus.extension.service.IService;
 
@@ -55,7 +55,7 @@ public interface AppBuildPipeService extends IService<AppBuildPipeline> {
   boolean allowToBuildNow(@Nonnull Long appId);
 
   /** list pipeline status on application id list */
-  Map<Long, PipelineStatus> listPipelineStatus(List<Long> appIds);
+  Map<Long, PipelineStatusEnum> listPipelineStatus(List<Long> appIds);
 
   /**
    * delete appBuildPipeline By application
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/EffectiveService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/EffectiveService.java
index c003526f7..7488be867 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/EffectiveService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/EffectiveService.java
@@ -18,17 +18,17 @@
 package org.apache.streampark.console.core.service;
 
 import org.apache.streampark.console.core.entity.Effective;
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 
 import com.baomidou.mybatisplus.extension.service.IService;
 
 public interface EffectiveService extends IService<Effective> {
 
-  void delete(Long appId, EffectiveType config);
+  void delete(Long appId, EffectiveTypeEnum config);
 
-  Effective get(Long appId, EffectiveType config);
+  Effective get(Long appId, EffectiveTypeEnum config);
 
-  void saveOrUpdate(Long appId, EffectiveType type, Long id);
+  void saveOrUpdate(Long appId, EffectiveTypeEnum type, Long id);
 
   void removeApp(Long appId);
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkClusterService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkClusterService.java
index 18393545a..2c8ed7ff2 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkClusterService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkClusterService.java
@@ -17,8 +17,8 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ClusterState;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ClusterStateEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.core.bean.ResponseResult;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 
@@ -49,7 +49,7 @@ public interface FlinkClusterService extends IService<FlinkCluster> {
 
   Boolean existsByFlinkEnvId(Long id);
 
-  List<FlinkCluster> getByExecutionModes(Collection<ExecutionMode> executionModes);
+  List<FlinkCluster> getByExecutionModes(Collection<ExecutionModeEnum> executionModeEnums);
 
-  void updateClusterState(Long id, ClusterState state);
+  void updateClusterState(Long id, ClusterStateEnum state);
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkSqlService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkSqlService.java
index 0a836821a..88b3f93c3 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkSqlService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/FlinkSqlService.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.service;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkSql;
-import org.apache.streampark.console.core.enums.CandidateType;
+import org.apache.streampark.console.core.enums.CandidateTypeEnum;
 import org.apache.streampark.flink.core.FlinkSqlValidationResult;
 
 import com.baomidou.mybatisplus.core.metadata.IPage;
@@ -32,7 +32,7 @@ public interface FlinkSqlService extends IService<FlinkSql> {
 
   void create(FlinkSql flinkSql);
 
-  void setCandidate(CandidateType candidateType, Long appId, Long sqlId);
+  void setCandidate(CandidateTypeEnum candidateTypeEnum, Long appId, Long sqlId);
 
   FlinkSql getEffective(Long appId, boolean decode);
 
@@ -40,7 +40,7 @@ public interface FlinkSqlService extends IService<FlinkSql> {
 
   List<FlinkSql> history(Application application);
 
-  FlinkSql getCandidate(Long appId, CandidateType type);
+  FlinkSql getCandidate(Long appId, CandidateTypeEnum type);
 
   void toEffective(Long appId, Long sqlId);
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/MessageService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/MessageService.java
index a26fd8be7..0d33ff2ba 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/MessageService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/MessageService.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.service;
 
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.core.entity.Message;
-import org.apache.streampark.console.core.enums.NoticeType;
+import org.apache.streampark.console.core.enums.NoticeTypeEnum;
 
 import com.baomidou.mybatisplus.core.metadata.IPage;
 import com.baomidou.mybatisplus.extension.service.IService;
@@ -28,5 +28,5 @@ public interface MessageService extends IService<Message> {
 
   void push(Message message);
 
-  IPage<Message> getUnRead(NoticeType noticeType, RestRequest request);
+  IPage<Message> getUnRead(NoticeTypeEnum noticeTypeEnum, RestRequest request);
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/YarnQueueService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/YarnQueueService.java
index 6f92456c4..eaaafd649 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/YarnQueueService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/YarnQueueService.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.core.bean.ResponseResult;
 import org.apache.streampark.console.core.entity.YarnQueue;
@@ -37,7 +37,7 @@ public interface YarnQueueService extends IService<YarnQueue> {
 
   void deleteYarnQueue(YarnQueue yarnQueue);
 
-  void checkQueueLabel(ExecutionMode executionMode, String queueLabel);
+  void checkQueueLabel(ExecutionModeEnum executionModeEnum, String queueLabel);
 
   boolean isDefaultQueue(String queueLabel);
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/alert/impl/AlertServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/alert/impl/AlertServiceImpl.java
index dfaac9873..1175096e8 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/alert/impl/AlertServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/alert/impl/AlertServiceImpl.java
@@ -22,7 +22,7 @@ import org.apache.streampark.console.base.util.SpringContextUtils;
 import org.apache.streampark.console.core.bean.AlertConfigParams;
 import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.entity.AlertConfig;
-import org.apache.streampark.console.core.enums.AlertType;
+import org.apache.streampark.console.core.enums.AlertTypeEnum;
 import org.apache.streampark.console.core.service.alert.AlertConfigService;
 import org.apache.streampark.console.core.service.alert.AlertService;
 
@@ -52,18 +52,18 @@ public class AlertServiceImpl implements AlertService {
     AlertConfig alertConfig = alertConfigService.getById(alertConfigId);
     try {
       AlertConfigParams params = AlertConfigParams.of(alertConfig);
-      List<AlertType> alertTypes = AlertType.decode(params.getAlertType());
-      if (CollectionUtils.isEmpty(alertTypes)) {
+      List<AlertTypeEnum> alertTypeEnums = AlertTypeEnum.decode(params.getAlertType());
+      if (CollectionUtils.isEmpty(alertTypeEnums)) {
         return true;
       }
       // No use thread pool, ensure that the alarm can be sent successfully
       Tuple2<Boolean, AlertException> reduce =
-          alertTypes.stream()
+          alertTypeEnums.stream()
               .map(
-                  alertType -> {
+                  alertTypeEnum -> {
                     try {
                       boolean alertRes =
-                          SpringContextUtils.getBean(alertType.getClazz())
+                          SpringContextUtils.getBean(alertTypeEnum.getClazz())
                               .doAlert(params, alertTemplate);
                       return new Tuple2<Boolean, AlertException>(alertRes, null);
                     } catch (AlertException e) {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationInfoService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationInfoService.java
index 57fbdef59..5dc9349a7 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationInfoService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationInfoService.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.service.application;
 
 import org.apache.streampark.console.base.exception.ApplicationException;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.AppExistsState;
+import org.apache.streampark.console.core.enums.AppExistsStateEnum;
 
 import com.baomidou.mybatisplus.extension.service.IService;
 
@@ -140,7 +140,7 @@ public interface ApplicationInfoService extends IService<Application> {
    * @param appParam The application to check for existence.
    * @return AppExistsState indicating the existence state of the application.
    */
-  AppExistsState checkExists(Application appParam);
+  AppExistsStateEnum checkExists(Application appParam);
 
   /**
    * Persists the metrics of the given application.
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationManageService.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationManageService.java
index c472b0cc3..4ea3708e7 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationManageService.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/ApplicationManageService.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service.application;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.core.entity.Application;
 
@@ -143,12 +143,12 @@ public interface ApplicationManageService extends IService<Application> {
    * Retrieves a list of applications by team ID and execution modes.
    *
    * @param teamId The ID of the team to filter by
-   * @param executionModes The collection of execution modes to filter by
+   * @param executionModeEnums The collection of execution modes to filter by
    * @return A list of applications that belong to the specified team and have the specified
    *     execution modes
    */
   List<Application> getByTeamIdAndExecutionModes(
-      Long teamId, Collection<ExecutionMode> executionModes);
+      Long teamId, Collection<ExecutionModeEnum> executionModeEnums);
 
   /**
    * Retrieves a list of applications be probing or need to probe.
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationActionServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationActionServiceImpl.java
index a364c1db9..7b5c63900 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationActionServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationActionServiceImpl.java
@@ -20,10 +20,10 @@ package org.apache.streampark.console.core.service.application.impl;
 import org.apache.streampark.common.conf.ConfigConst;
 import org.apache.streampark.common.conf.K8sFlinkConfig;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.DevelopmentMode;
-import org.apache.streampark.common.enums.ExecutionMode;
-import org.apache.streampark.common.enums.ResolveOrder;
-import org.apache.streampark.common.enums.RestoreMode;
+import org.apache.streampark.common.enums.DevelopmentModeEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
+import org.apache.streampark.common.enums.ResolveOrderEnum;
+import org.apache.streampark.common.enums.RestoreModeEnum;
 import org.apache.streampark.common.fs.FsOperator;
 import org.apache.streampark.common.util.CompletableFutureUtils;
 import org.apache.streampark.common.util.DeflaterUtils;
@@ -43,12 +43,12 @@ import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.FlinkSql;
 import org.apache.streampark.console.core.entity.Resource;
 import org.apache.streampark.console.core.entity.SavePoint;
-import org.apache.streampark.console.core.enums.CheckPointType;
-import org.apache.streampark.console.core.enums.ConfigFileType;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.Operation;
-import org.apache.streampark.console.core.enums.OptionState;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.CheckPointTypeEnum;
+import org.apache.streampark.console.core.enums.ConfigFileTypeEnum;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.OperationEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.mapper.ApplicationMapper;
 import org.apache.streampark.console.core.service.AppBuildPipeService;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
@@ -186,12 +186,12 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
     LambdaUpdateWrapper<Application> updateWrapper = Wrappers.lambdaUpdate();
     updateWrapper.eq(Application::getId, application.getId());
     if (application.isFlinkSqlJob()) {
-      updateWrapper.set(Application::getRelease, ReleaseState.FAILED.get());
+      updateWrapper.set(Application::getRelease, ReleaseStateEnum.FAILED.get());
     } else {
-      updateWrapper.set(Application::getRelease, ReleaseState.NEED_RELEASE.get());
+      updateWrapper.set(Application::getRelease, ReleaseStateEnum.NEED_RELEASE.get());
     }
     if (!application.isRunning()) {
-      updateWrapper.set(Application::getState, FlinkAppState.REVOKED.getValue());
+      updateWrapper.set(Application::getState, FlinkAppStateEnum.REVOKED.getValue());
     }
     baseMapper.update(null, updateWrapper);
   }
@@ -228,12 +228,12 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
   @Override
   public void cancel(Application appParam) throws Exception {
-    FlinkAppHttpWatcher.setOptionState(appParam.getId(), OptionState.CANCELLING);
+    FlinkAppHttpWatcher.setOptionState(appParam.getId(), OptionStateEnum.CANCELLING);
     Application application = getById(appParam.getId());
-    application.setState(FlinkAppState.CANCELLING.getValue());
+    application.setState(FlinkAppStateEnum.CANCELLING.getValue());
 
     ApplicationLog applicationLog = new ApplicationLog();
-    applicationLog.setOptionName(Operation.CANCEL.getValue());
+    applicationLog.setOptionName(OperationEnum.CANCEL.getValue());
     applicationLog.setAppId(application.getId());
     applicationLog.setJobManagerUrl(application.getJobManagerUrl());
     applicationLog.setOptionTime(new Date());
@@ -241,9 +241,9 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
     if (appParam.getSavePointed()) {
       FlinkAppHttpWatcher.addSavepoint(application.getId());
-      application.setOptionState(OptionState.SAVEPOINTING.getValue());
+      application.setOptionState(OptionStateEnum.SAVEPOINTING.getValue());
     } else {
-      application.setOptionState(OptionState.CANCELLING.getValue());
+      application.setOptionState(OptionStateEnum.CANCELLING.getValue());
     }
 
     application.setOptionTime(new Date());
@@ -266,10 +266,10 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
     }
 
     String clusterId = null;
-    if (ExecutionMode.isKubernetesMode(application.getExecutionMode())) {
+    if (ExecutionModeEnum.isKubernetesMode(application.getExecutionMode())) {
       clusterId = application.getClusterId();
-    } else if (ExecutionMode.isYarnMode(application.getExecutionMode())) {
-      if (ExecutionMode.YARN_SESSION == application.getExecutionModeEnum()) {
+    } else if (ExecutionModeEnum.isYarnMode(application.getExecutionMode())) {
+      if (ExecutionModeEnum.YARN_SESSION == application.getExecutionModeEnum()) {
         FlinkCluster cluster = flinkClusterService.getById(application.getFlinkClusterId());
         ApiAlertException.throwIfNull(
             cluster,
@@ -284,7 +284,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
     Map<String, Object> properties = new HashMap<>();
 
-    if (ExecutionMode.isRemoteMode(application.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isRemoteMode(application.getExecutionModeEnum())) {
       FlinkCluster cluster = flinkClusterService.getById(application.getFlinkClusterId());
       ApiAlertException.throwIfNull(
           cluster,
@@ -301,7 +301,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
         new CancelRequest(
             application.getId(),
             flinkEnv.getFlinkVersion(),
-            ExecutionMode.of(application.getExecutionMode()),
+            ExecutionModeEnum.of(application.getExecutionMode()),
             properties,
             clusterId,
             application.getJobId(),
@@ -330,7 +330,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
                 savePoint.setPath(savePointDir);
                 savePoint.setAppId(application.getId());
                 savePoint.setLatest(true);
-                savePoint.setType(CheckPointType.SAVEPOINT.get());
+                savePoint.setType(CheckPointTypeEnum.SAVEPOINT.get());
                 savePoint.setCreateTime(new Date());
                 savePoint.setTriggerTime(triggerTime);
                 savePointService.save(savePoint);
@@ -344,8 +344,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
                 updateToStopped(application);
               } else {
                 log.error("stop flink job fail.", e);
-                application.setOptionState(OptionState.NONE.getValue());
-                application.setState(FlinkAppState.FAILED.getValue());
+                application.setOptionState(OptionStateEnum.NONE.getValue());
+                application.setState(FlinkAppStateEnum.FAILED.getValue());
                 updateById(application);
 
                 if (appParam.getSavePointed()) {
@@ -407,7 +407,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
     String jobId = new JobID().toHexString();
     ApplicationLog applicationLog = new ApplicationLog();
-    applicationLog.setOptionName(Operation.START.getValue());
+    applicationLog.setOptionName(OperationEnum.START.getValue());
     applicationLog.setAppId(application.getId());
     applicationLog.setOptionTime(new Date());
 
@@ -439,7 +439,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
     String appConf = userJarAndAppConf.f1;
 
     BuildResult buildResult = buildPipeline.getBuildResult();
-    if (ExecutionMode.YARN_APPLICATION == application.getExecutionModeEnum()) {
+    if (ExecutionModeEnum.YARN_APPLICATION == application.getExecutionModeEnum()) {
       buildResult = new ShadedBuildResponse(null, flinkUserJar, true);
     }
 
@@ -450,17 +450,19 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
     SubmitRequest submitRequest =
         new SubmitRequest(
             flinkEnv.getFlinkVersion(),
-            ExecutionMode.of(application.getExecutionMode()),
+            ExecutionModeEnum.of(application.getExecutionMode()),
             getProperties(application),
             flinkEnv.getFlinkConf(),
-            DevelopmentMode.of(application.getJobType()),
+            DevelopmentModeEnum.of(application.getJobType()),
             application.getId(),
             jobId,
             application.getJobName(),
             appConf,
             application.getApplicationType(),
             getSavePointed(appParam),
-            appParam.getRestoreMode() == null ? null : RestoreMode.of(appParam.getRestoreMode()),
+            appParam.getRestoreMode() == null
+                ? null
+                : RestoreModeEnum.of(appParam.getRestoreMode()),
             applicationArgs,
             buildResult,
             kubernetesSubmitParam,
@@ -501,7 +503,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
               application.setStartTime(new Date());
               application.setEndTime(null);
               if (isKubernetesApp(application)) {
-                application.setRelease(ReleaseState.DONE.get());
+                application.setRelease(ReleaseStateEnum.DONE.get());
               }
               updateById(application);
 
@@ -509,7 +511,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
               if (isKubernetesApp(application)) {
                 k8SFlinkTrackMonitor.doWatching(toTrackId(application));
               } else {
-                FlinkAppHttpWatcher.setOptionState(appParam.getId(), OptionState.STARTING);
+                FlinkAppHttpWatcher.setOptionState(appParam.getId(), OptionStateEnum.STARTING);
                 FlinkAppHttpWatcher.doWatching(application);
               }
 
@@ -525,8 +527,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
                 applicationLog.setException(exception);
                 applicationLog.setSuccess(false);
                 Application app = getById(appParam.getId());
-                app.setState(FlinkAppState.FAILED.getValue());
-                app.setOptionState(OptionState.NONE.getValue());
+                app.setState(FlinkAppStateEnum.FAILED.getValue());
+                app.setOptionState(OptionStateEnum.NONE.getValue());
                 updateById(app);
                 if (isKubernetesApp(app)) {
                   k8SFlinkTrackMonitor.unWatching(toTrackId(app));
@@ -538,7 +540,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
         .whenComplete(
             (t, e) -> {
               if (!K8sFlinkConfig.isV2Enabled()
-                  && ExecutionMode.isKubernetesApplicationMode(application.getExecutionMode())) {
+                  && ExecutionModeEnum.isKubernetesApplicationMode(
+                      application.getExecutionMode())) {
                 String domainName = settingService.getIngressModeDefault();
                 if (StringUtils.isNotBlank(domainName)) {
                   try {
@@ -551,8 +554,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
                     applicationLog.setException(e.getMessage());
                     applicationLog.setSuccess(false);
                     applicationLogService.save(applicationLog);
-                    application.setState(FlinkAppState.FAILED.getValue());
-                    application.setOptionState(OptionState.NONE.getValue());
+                    application.setState(FlinkAppStateEnum.FAILED.getValue());
+                    application.setOptionState(OptionStateEnum.NONE.getValue());
                     updateById(application);
                     return;
                   }
@@ -565,17 +568,17 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
   }
 
   private void starting(Application application) {
-    application.setState(FlinkAppState.STARTING.getValue());
+    application.setState(FlinkAppStateEnum.STARTING.getValue());
     application.setOptionTime(new Date());
     updateById(application);
   }
 
   private Tuple2<String, String> getUserJarAndAppConf(FlinkEnv flinkEnv, Application application) {
-    ExecutionMode executionMode = application.getExecutionModeEnum();
+    ExecutionModeEnum executionModeEnum = application.getExecutionModeEnum();
     ApplicationConfig applicationConfig = configService.getEffective(application.getId());
 
     ApiAlertException.throwIfNull(
-        executionMode, "ExecutionMode can't be null, start application failed.");
+        executionModeEnum, "ExecutionMode can't be null, start application failed.");
 
     String flinkUserJar = null;
     String appConf = null;
@@ -592,7 +595,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
                 ? null
                 : String.format("yaml://%s", applicationConfig.getContent());
         // 3) client
-        if (ExecutionMode.YARN_APPLICATION == executionMode) {
+        if (ExecutionModeEnum.YARN_APPLICATION == executionModeEnum) {
           String clientPath = Workspace.remote().APP_CLIENT();
           flinkUserJar = String.format("%s/%s", clientPath, sqlDistJar);
         }
@@ -624,8 +627,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
         } else {
           switch (application.getApplicationType()) {
             case STREAMPARK_FLINK:
-              ConfigFileType fileType = ConfigFileType.of(applicationConfig.getFormat());
-              if (fileType != null && ConfigFileType.UNKNOWN != fileType) {
+              ConfigFileTypeEnum fileType = ConfigFileTypeEnum.of(applicationConfig.getFormat());
+              if (fileType != null && ConfigFileTypeEnum.UNKNOWN != fileType) {
                 appConf =
                     String.format(
                         "%s://%s", fileType.getTypeName(), applicationConfig.getContent());
@@ -646,7 +649,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
           }
         }
 
-        if (ExecutionMode.YARN_APPLICATION == executionMode) {
+        if (ExecutionModeEnum.YARN_APPLICATION == executionModeEnum) {
           switch (application.getApplicationType()) {
             case STREAMPARK_FLINK:
               flinkUserJar =
@@ -679,7 +682,7 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
   private Map<String, Object> getProperties(Application application) {
     Map<String, Object> properties = new HashMap<>(application.getOptionMap());
-    if (ExecutionMode.isRemoteMode(application.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isRemoteMode(application.getExecutionModeEnum())) {
       FlinkCluster cluster = flinkClusterService.getById(application.getFlinkClusterId());
       ApiAlertException.throwIfNull(
           cluster,
@@ -690,8 +693,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
       URI activeAddress = cluster.getRemoteURI();
       properties.put(RestOptions.ADDRESS.key(), activeAddress.getHost());
       properties.put(RestOptions.PORT.key(), activeAddress.getPort());
-    } else if (ExecutionMode.isYarnMode(application.getExecutionModeEnum())) {
-      if (ExecutionMode.YARN_SESSION == application.getExecutionModeEnum()) {
+    } else if (ExecutionModeEnum.isYarnMode(application.getExecutionModeEnum())) {
+      if (ExecutionModeEnum.YARN_SESSION == application.getExecutionModeEnum()) {
         FlinkCluster cluster = flinkClusterService.getById(application.getFlinkClusterId());
         ApiAlertException.throwIfNull(
             cluster,
@@ -710,11 +713,11 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
         Optional.ofNullable(yarnLabelExpr)
             .ifPresent(yLabel -> properties.put(ConfigConst.KEY_YARN_APP_NODE_LABEL(), yLabel));
       }
-    } else if (ExecutionMode.isKubernetesMode(application.getExecutionModeEnum())) {
+    } else if (ExecutionModeEnum.isKubernetesMode(application.getExecutionModeEnum())) {
       properties.put(ConfigConst.KEY_K8S_IMAGE_PULL_POLICY(), "Always");
     }
 
-    if (ExecutionMode.isKubernetesApplicationMode(application.getExecutionMode())) {
+    if (ExecutionModeEnum.isKubernetesApplicationMode(application.getExecutionMode())) {
       try {
         HadoopUtils.yarnClient();
         properties.put(JobManagerOptions.ARCHIVE_DIR.key(), Workspace.ARCHIVES_FILE_PATH());
@@ -730,9 +733,9 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
     Map<String, String> dynamicProperties =
         PropertiesUtils.extractDynamicPropertiesAsJava(application.getDynamicProperties());
     properties.putAll(dynamicProperties);
-    ResolveOrder resolveOrder = ResolveOrder.of(application.getResolveOrder());
-    if (resolveOrder != null) {
-      properties.put(CoreOptions.CLASSLOADER_RESOLVE_ORDER.key(), resolveOrder.getName());
+    ResolveOrderEnum resolveOrderEnum = ResolveOrderEnum.of(application.getResolveOrder());
+    if (resolveOrderEnum != null) {
+      properties.put(CoreOptions.CLASSLOADER_RESOLVE_ORDER.key(), resolveOrderEnum.getName());
     }
 
     return properties;
@@ -740,8 +743,8 @@ public class ApplicationActionServiceImpl extends ServiceImpl<ApplicationMapper,
 
   private void updateToStopped(Application app) {
     Application application = getById(app);
-    application.setOptionState(OptionState.NONE.getValue());
-    application.setState(FlinkAppState.CANCELED.getValue());
+    application.setOptionState(OptionStateEnum.NONE.getValue());
+    application.setState(FlinkAppStateEnum.CANCELED.getValue());
     application.setOptionTime(new Date());
     updateById(application);
     savePointService.expire(application.getId());
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationInfoServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationInfoServiceImpl.java
index 528f31169..1be928ea6 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationInfoServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationInfoServiceImpl.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.core.service.application.impl;
 
 import org.apache.streampark.common.conf.K8sFlinkConfig;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.fs.LfsOperator;
 import org.apache.streampark.common.util.ExceptionUtils;
 import org.apache.streampark.common.util.Utils;
@@ -32,8 +32,8 @@ import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.Project;
-import org.apache.streampark.console.core.enums.AppExistsState;
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.AppExistsStateEnum;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 import org.apache.streampark.console.core.mapper.ApplicationMapper;
 import org.apache.streampark.console.core.metrics.flink.JobsOverview;
 import org.apache.streampark.console.core.runner.EnvInitializer;
@@ -73,7 +73,7 @@ import java.util.jar.Manifest;
 import java.util.regex.Pattern;
 import java.util.stream.Collectors;
 
-import static org.apache.streampark.common.enums.StorageType.LFS;
+import static org.apache.streampark.common.enums.StorageTypeEnum.LFS;
 import static org.apache.streampark.console.core.task.FlinkK8sWatcherWrapper.Bridge.toTrackId;
 import static org.apache.streampark.console.core.task.FlinkK8sWatcherWrapper.isKubernetesApp;
 
@@ -135,7 +135,7 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
       if (app.getAvailableSlot() != null) {
         availableSlot += app.getAvailableSlot();
       }
-      if (app.getState() == FlinkAppState.RUNNING.getValue()) {
+      if (app.getState() == FlinkAppStateEnum.RUNNING.getValue()) {
         runningJob++;
       }
       JobsOverview.Task task = app.getOverview();
@@ -203,8 +203,8 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
       envInitializer.checkFlinkEnv(application.getStorageType(), flinkEnv);
       envInitializer.storageInitialize(application.getStorageType());
 
-      if (ExecutionMode.YARN_SESSION == application.getExecutionModeEnum()
-          || ExecutionMode.REMOTE == application.getExecutionModeEnum()) {
+      if (ExecutionModeEnum.YARN_SESSION == application.getExecutionModeEnum()
+          || ExecutionModeEnum.REMOTE == application.getExecutionModeEnum()) {
         FlinkCluster flinkCluster = flinkClusterService.getById(application.getFlinkClusterId());
         boolean conned = flinkClusterWatcher.verifyClusterConnection(flinkCluster);
         if (!conned) {
@@ -221,7 +221,7 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
   @Override
   public boolean checkAlter(Application appParam) {
     Long appId = appParam.getId();
-    if (FlinkAppState.CANCELED != appParam.getStateEnum()) {
+    if (FlinkAppStateEnum.CANCELED != appParam.getStateEnum()) {
       return false;
     }
     long cancelUserId = FlinkAppHttpWatcher.getCanceledJobUserId(appId);
@@ -248,7 +248,7 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
             .anyMatch(
                 application ->
                     clusterId.equals(application.getFlinkClusterId())
-                        && FlinkAppState.RUNNING == application.getStateEnum());
+                        && FlinkAppStateEnum.RUNNING == application.getStateEnum());
   }
 
   @Override
@@ -322,7 +322,7 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
     Application application = getById(id);
     ApiAlertException.throwIfNull(
         application, String.format("The application id=%s can't be found.", id));
-    if (ExecutionMode.isKubernetesMode(application.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isKubernetesMode(application.getExecutionModeEnum())) {
       CompletableFuture<String> future =
           CompletableFuture.supplyAsync(
               () ->
@@ -378,10 +378,10 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
    * @return The state of the application's existence.
    */
   @Override
-  public AppExistsState checkExists(Application appParam) {
+  public AppExistsStateEnum checkExists(Application appParam) {
 
     if (!checkJobName(appParam.getJobName())) {
-      return AppExistsState.INVALID;
+      return AppExistsStateEnum.INVALID;
     }
 
     boolean existsByJobName = this.existsByJobName(appParam.getJobName());
@@ -389,43 +389,43 @@ public class ApplicationInfoServiceImpl extends ServiceImpl<ApplicationMapper, A
     if (appParam.getId() != null) {
       Application app = getById(appParam.getId());
       if (app.getJobName().equals(appParam.getJobName())) {
-        return AppExistsState.NO;
+        return AppExistsStateEnum.NO;
       }
 
       if (existsByJobName) {
-        return AppExistsState.IN_DB;
+        return AppExistsStateEnum.IN_DB;
       }
 
       // has stopped status
-      if (FlinkAppState.isEndState(app.getState())) {
+      if (FlinkAppStateEnum.isEndState(app.getState())) {
         // check whether jobName exists on yarn
-        if (ExecutionMode.isYarnMode(appParam.getExecutionMode())
+        if (ExecutionModeEnum.isYarnMode(appParam.getExecutionMode())
             && YarnUtils.isContains(appParam.getJobName())) {
-          return AppExistsState.IN_YARN;
+          return AppExistsStateEnum.IN_YARN;
         }
         // check whether clusterId, namespace, jobId on kubernetes
-        else if (ExecutionMode.isKubernetesMode(appParam.getExecutionMode())
+        else if (ExecutionModeEnum.isKubernetesMode(appParam.getExecutionMode())
             && k8SFlinkTrackMonitor.checkIsInRemoteCluster(toTrackId(appParam))) {
-          return AppExistsState.IN_KUBERNETES;
+          return AppExistsStateEnum.IN_KUBERNETES;
         }
       }
     } else {
       if (existsByJobName) {
-        return AppExistsState.IN_DB;
+        return AppExistsStateEnum.IN_DB;
       }
 
       // check whether jobName exists on yarn
-      if (ExecutionMode.isYarnMode(appParam.getExecutionMode())
+      if (ExecutionModeEnum.isYarnMode(appParam.getExecutionMode())
           && YarnUtils.isContains(appParam.getJobName())) {
-        return AppExistsState.IN_YARN;
+        return AppExistsStateEnum.IN_YARN;
       }
       // check whether clusterId, namespace, jobId on kubernetes
-      else if (ExecutionMode.isKubernetesMode(appParam.getExecutionMode())
+      else if (ExecutionModeEnum.isKubernetesMode(appParam.getExecutionMode())
           && k8SFlinkTrackMonitor.checkIsInRemoteCluster(toTrackId(appParam))) {
-        return AppExistsState.IN_KUBERNETES;
+        return AppExistsStateEnum.IN_KUBERNETES;
       }
     }
-    return AppExistsState.NO;
+    return AppExistsStateEnum.NO;
   }
 
   private boolean existsByJobName(String jobName) {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationManageServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationManageServiceImpl.java
index 4a46462ff..48d3638ed 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationManageServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/application/impl/ApplicationManageServiceImpl.java
@@ -19,8 +19,8 @@ package org.apache.streampark.console.core.service.application.impl;
 
 import org.apache.streampark.common.conf.K8sFlinkConfig;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.ExecutionMode;
-import org.apache.streampark.common.enums.StorageType;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
+import org.apache.streampark.common.enums.StorageTypeEnum;
 import org.apache.streampark.common.fs.HdfsOperator;
 import org.apache.streampark.common.util.DeflaterUtils;
 import org.apache.streampark.console.base.domain.RestRequest;
@@ -34,11 +34,11 @@ import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ApplicationConfig;
 import org.apache.streampark.console.core.entity.FlinkSql;
 import org.apache.streampark.console.core.entity.Resource;
-import org.apache.streampark.console.core.enums.CandidateType;
+import org.apache.streampark.console.core.enums.CandidateTypeEnum;
 import org.apache.streampark.console.core.enums.ChangeTypeEnum;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.OptionState;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.mapper.ApplicationMapper;
 import org.apache.streampark.console.core.service.AppBuildPipeService;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
@@ -57,7 +57,7 @@ import org.apache.streampark.console.core.task.FlinkAppHttpWatcher;
 import org.apache.streampark.console.core.task.FlinkK8sObserverStub;
 import org.apache.streampark.console.core.utils.FlinkK8sDataTypeConverterStub;
 import org.apache.streampark.flink.kubernetes.FlinkK8sWatcher;
-import org.apache.streampark.flink.packer.pipeline.PipelineStatus;
+import org.apache.streampark.flink.packer.pipeline.PipelineStatusEnum;
 
 import org.apache.commons.lang3.StringUtils;
 
@@ -203,7 +203,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
           .delete(application.getWorkspace().APP_WORKSPACE().concat("/").concat(appId.toString()));
       // try to delete yarn-application, and leave no trouble.
       String path =
-          Workspace.of(StorageType.HDFS).APP_WORKSPACE().concat("/").concat(appId.toString());
+          Workspace.of(StorageTypeEnum.HDFS).APP_WORKSPACE().concat("/").concat(appId.toString());
       if (HdfsOperator.exists(path)) {
         HdfsOperator.delete(path);
       }
@@ -220,12 +220,12 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     Page<Application> page = new MybatisPager<Application>().getDefaultPage(request);
     if (CommonUtils.notEmpty(appParam.getStateArray())) {
       if (Arrays.stream(appParam.getStateArray())
-          .anyMatch(x -> x == FlinkAppState.FINISHED.getValue())) {
+          .anyMatch(x -> x == FlinkAppStateEnum.FINISHED.getValue())) {
         Integer[] newArray =
             CommonUtils.arrayInsertIndex(
                 appParam.getStateArray(),
                 appParam.getStateArray().length,
-                FlinkAppState.POS_TERMINATED.getValue());
+                FlinkAppStateEnum.POS_TERMINATED.getValue());
         appParam.setStateArray(newArray);
       }
     }
@@ -234,7 +234,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     long now = System.currentTimeMillis();
 
     List<Long> appIds = records.stream().map(Application::getId).collect(Collectors.toList());
-    Map<Long, PipelineStatus> pipeStates = appBuildPipeService.listPipelineStatus(appIds);
+    Map<Long, PipelineStatusEnum> pipeStates = appBuildPipeService.listPipelineStatus(appIds);
 
     List<Application> newRecords =
         records.stream()
@@ -260,12 +260,12 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
                       new AppControl()
                           .setAllowBuild(
                               record.getBuildStatus() == null
-                                  || !PipelineStatus.running
+                                  || !PipelineStatusEnum.running
                                       .getCode()
                                       .equals(record.getBuildStatus()))
                           .setAllowStart(
                               !record.shouldTracking()
-                                  && PipelineStatus.success
+                                  && PipelineStatusEnum.success
                                       .getCode()
                                       .equals(record.getBuildStatus()))
                           .setAllowStop(record.isRunning());
@@ -292,9 +292,9 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     ApiAlertException.throwIfNull(
         appParam.getTeamId(), "The teamId can't be null. Create application failed.");
     appParam.setUserId(commonService.getUserId());
-    appParam.setState(FlinkAppState.ADDED.getValue());
-    appParam.setRelease(ReleaseState.NEED_RELEASE.get());
-    appParam.setOptionState(OptionState.NONE.getValue());
+    appParam.setState(FlinkAppStateEnum.ADDED.getValue());
+    appParam.setRelease(ReleaseStateEnum.NEED_RELEASE.get());
+    appParam.setOptionState(OptionStateEnum.NONE.getValue());
     appParam.setCreateTime(new Date());
     appParam.setDefaultModeIngress(settingService.getIngressModeDefault());
 
@@ -345,7 +345,8 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
   }
 
   private boolean shouldHandleK8sName(Application app) {
-    return K8sFlinkConfig.isV2Enabled() && ExecutionMode.isKubernetesMode(app.getExecutionMode());
+    return K8sFlinkConfig.isV2Enabled()
+        && ExecutionModeEnum.isKubernetesMode(app.getExecutionMode());
   }
 
   private boolean existsByJobName(String jobName) {
@@ -369,7 +370,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
 
     newApp.setJobName(jobName);
     newApp.setClusterId(
-        ExecutionMode.isSessionMode(oldApp.getExecutionModeEnum())
+        ExecutionModeEnum.isSessionMode(oldApp.getExecutionModeEnum())
             ? oldApp.getClusterId()
             : jobName);
     newApp.setArgs(appParam.getArgs() != null ? appParam.getArgs() : oldApp.getArgs());
@@ -400,9 +401,9 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     newApp.setProjectId(oldApp.getProjectId());
     newApp.setModule(oldApp.getModule());
     newApp.setUserId(commonService.getUserId());
-    newApp.setState(FlinkAppState.ADDED.getValue());
-    newApp.setRelease(ReleaseState.NEED_RELEASE.get());
-    newApp.setOptionState(OptionState.NONE.getValue());
+    newApp.setState(FlinkAppStateEnum.ADDED.getValue());
+    newApp.setRelease(ReleaseStateEnum.NEED_RELEASE.get());
+    newApp.setOptionState(OptionStateEnum.NONE.getValue());
     newApp.setCreateTime(new Date());
     newApp.setHotParams(oldApp.getHotParams());
 
@@ -449,7 +450,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
         success,
         String.format(ERROR_APP_QUEUE_HINT, appParam.getYarnQueue(), appParam.getTeamId()));
 
-    application.setRelease(ReleaseState.NEED_RELEASE.get());
+    application.setRelease(ReleaseStateEnum.NEED_RELEASE.get());
 
     // 1) jar job jar file changed
     if (application.isUploadJob()) {
@@ -472,7 +473,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     }
 
     // 2) k8s podTemplate changed.
-    if (application.getBuild() && ExecutionMode.isKubernetesMode(appParam.getExecutionMode())) {
+    if (application.getBuild() && ExecutionModeEnum.isKubernetesMode(appParam.getExecutionMode())) {
       if (ObjectUtils.trimNoEquals(
               application.getK8sRestExposedType(), appParam.getK8sRestExposedType())
           || ObjectUtils.trimNoEquals(
@@ -497,8 +498,8 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     // 4) yarn application mode change
     if (!application.getBuild()) {
       if (!application.getExecutionMode().equals(appParam.getExecutionMode())) {
-        if (ExecutionMode.YARN_APPLICATION == appParam.getExecutionModeEnum()
-            || ExecutionMode.YARN_APPLICATION == application.getExecutionModeEnum()) {
+        if (ExecutionModeEnum.YARN_APPLICATION == appParam.getExecutionModeEnum()
+            || ExecutionModeEnum.YARN_APPLICATION == application.getExecutionModeEnum()) {
           application.setBuild(true);
         }
       }
@@ -590,7 +591,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
   private void updateFlinkSqlJob(Application application, Application appParam) {
     FlinkSql effectiveFlinkSql = flinkSqlService.getEffective(application.getId(), true);
     if (effectiveFlinkSql == null) {
-      effectiveFlinkSql = flinkSqlService.getCandidate(application.getId(), CandidateType.NEW);
+      effectiveFlinkSql = flinkSqlService.getCandidate(application.getId(), CandidateTypeEnum.NEW);
       flinkSqlService.removeById(effectiveFlinkSql.getId());
       FlinkSql sql = new FlinkSql(appParam);
       flinkSqlService.create(sql);
@@ -613,7 +614,8 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
       // if has been changed
       if (changeTypeEnum.hasChanged()) {
         // check if there is a candidate version for the newly added record
-        FlinkSql newFlinkSql = flinkSqlService.getCandidate(application.getId(), CandidateType.NEW);
+        FlinkSql newFlinkSql =
+            flinkSqlService.getCandidate(application.getId(), CandidateTypeEnum.NEW);
         // If the candidate version of the new record exists, it will be deleted directly,
         // and only one candidate version will be retained. If the new candidate version is not
         // effective,
@@ -624,7 +626,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
           flinkSqlService.removeById(newFlinkSql.getId());
         }
         FlinkSql historyFlinkSql =
-            flinkSqlService.getCandidate(application.getId(), CandidateType.HISTORY);
+            flinkSqlService.getCandidate(application.getId(), CandidateTypeEnum.HISTORY);
         // remove candidate flags that already exist but are set as candidates
         if (historyFlinkSql != null) {
           flinkSqlService.cleanCandidate(historyFlinkSql.getId());
@@ -639,9 +641,9 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
         boolean versionChanged = !effectiveFlinkSql.getId().equals(appParam.getSqlId());
         if (versionChanged) {
           // sql and dependency not changed, but version changed, means that rollback to the version
-          CandidateType type = CandidateType.HISTORY;
+          CandidateTypeEnum type = CandidateTypeEnum.HISTORY;
           flinkSqlService.setCandidate(type, appParam.getId(), appParam.getSqlId());
-          application.setRelease(ReleaseState.NEED_ROLLBACK.get());
+          application.setRelease(ReleaseStateEnum.NEED_ROLLBACK.get());
           application.setBuild(true);
         }
       }
@@ -674,15 +676,15 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
 
   @Override
   public List<Application> getByTeamIdAndExecutionModes(
-      Long teamId, @Nonnull Collection<ExecutionMode> executionModes) {
+      Long teamId, @Nonnull Collection<ExecutionModeEnum> executionModeEnums) {
     return getBaseMapper()
         .selectList(
             new LambdaQueryWrapper<Application>()
                 .eq((SFunction<Application, Long>) Application::getTeamId, teamId)
                 .in(
                     Application::getExecutionMode,
-                    executionModes.stream()
-                        .map(ExecutionMode::getMode)
+                    executionModeEnums.stream()
+                        .map(ExecutionModeEnum::getMode)
                         .collect(Collectors.toSet())));
   }
 
@@ -697,16 +699,17 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
       LambdaUpdateWrapper<Application> updateWrapper = Wrappers.lambdaUpdate();
       updateWrapper.eq(Application::getId, appParam.getId());
       if (appParam.isRunning()) {
-        updateWrapper.set(Application::getRelease, ReleaseState.NEED_RESTART.get());
+        updateWrapper.set(Application::getRelease, ReleaseStateEnum.NEED_RESTART.get());
       } else {
-        updateWrapper.set(Application::getRelease, ReleaseState.DONE.get());
-        updateWrapper.set(Application::getOptionState, OptionState.NONE.getValue());
+        updateWrapper.set(Application::getRelease, ReleaseStateEnum.DONE.get());
+        updateWrapper.set(Application::getOptionState, OptionStateEnum.NONE.getValue());
       }
       this.update(updateWrapper);
 
       // backup
       if (appParam.isFlinkSqlJob()) {
-        FlinkSql newFlinkSql = flinkSqlService.getCandidate(appParam.getId(), CandidateType.NEW);
+        FlinkSql newFlinkSql =
+            flinkSqlService.getCandidate(appParam.getId(), CandidateTypeEnum.NEW);
         if (!appParam.isNeedRollback() && newFlinkSql != null) {
           backUpService.backup(appParam, newFlinkSql);
         }
@@ -724,7 +727,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
 
   @Override
   public void clean(Application appParam) {
-    appParam.setRelease(ReleaseState.DONE.get());
+    appParam.setRelease(ReleaseStateEnum.DONE.get());
     this.updateRelease(appParam);
   }
 
@@ -739,7 +742,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     if (application.isFlinkSqlJob()) {
       FlinkSql flinkSql = flinkSqlService.getEffective(application.getId(), true);
       if (flinkSql == null) {
-        flinkSql = flinkSqlService.getCandidate(application.getId(), CandidateType.NEW);
+        flinkSql = flinkSqlService.getCandidate(application.getId(), CandidateTypeEnum.NEW);
         flinkSql.setSql(DeflaterUtils.unzipString(flinkSql.getSql()));
       }
       flinkSql.setToApplication(application);
@@ -799,7 +802,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
     }
 
     oldApp.setYarnQueueByHotParams();
-    if (ExecutionMode.isYarnPerJobOrAppMode(newApp.getExecutionModeEnum())
+    if (ExecutionModeEnum.isYarnPerJobOrAppMode(newApp.getExecutionModeEnum())
         && StringUtils.equals(oldApp.getYarnQueue(), newApp.getYarnQueue())) {
       return true;
     }
@@ -815,7 +818,7 @@ public class ApplicationManageServiceImpl extends ServiceImpl<ApplicationMapper,
    *     (empty or default), return true, false else.
    */
   private boolean isYarnNotDefaultQueue(Application application) {
-    return ExecutionMode.isYarnPerJobOrAppMode(application.getExecutionModeEnum())
+    return ExecutionModeEnum.isYarnPerJobOrAppMode(application.getExecutionModeEnum())
         && !yarnQueueService.isDefaultQueue(application.getYarnQueue());
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/AppBuildPipeServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/AppBuildPipeServiceImpl.java
index 8c3120638..c4716db2d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/AppBuildPipeServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/AppBuildPipeServiceImpl.java
@@ -20,9 +20,9 @@ package org.apache.streampark.console.core.service.impl;
 import org.apache.streampark.common.conf.ConfigConst;
 import org.apache.streampark.common.conf.K8sFlinkConfig;
 import org.apache.streampark.common.conf.Workspace;
-import org.apache.streampark.common.enums.ApplicationType;
-import org.apache.streampark.common.enums.DevelopmentMode;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ApplicationTypeEnum;
+import org.apache.streampark.common.enums.DevelopmentModeEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.fs.FsOperator;
 import org.apache.streampark.common.util.ExceptionUtils;
 import org.apache.streampark.common.util.FileUtils;
@@ -41,11 +41,11 @@ import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.FlinkSql;
 import org.apache.streampark.console.core.entity.Message;
 import org.apache.streampark.console.core.entity.Resource;
-import org.apache.streampark.console.core.enums.CandidateType;
-import org.apache.streampark.console.core.enums.NoticeType;
-import org.apache.streampark.console.core.enums.OptionState;
-import org.apache.streampark.console.core.enums.ReleaseState;
-import org.apache.streampark.console.core.enums.ResourceType;
+import org.apache.streampark.console.core.enums.CandidateTypeEnum;
+import org.apache.streampark.console.core.enums.NoticeTypeEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
+import org.apache.streampark.console.core.enums.ResourceTypeEnum;
 import org.apache.streampark.console.core.mapper.ApplicationBuildPipelineMapper;
 import org.apache.streampark.console.core.service.AppBuildPipeService;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
@@ -77,8 +77,8 @@ import org.apache.streampark.flink.packer.pipeline.FlinkRemotePerJobBuildRequest
 import org.apache.streampark.flink.packer.pipeline.FlinkYarnApplicationBuildRequest;
 import org.apache.streampark.flink.packer.pipeline.PipeSnapshot;
 import org.apache.streampark.flink.packer.pipeline.PipeWatcher;
-import org.apache.streampark.flink.packer.pipeline.PipelineStatus;
-import org.apache.streampark.flink.packer.pipeline.PipelineType;
+import org.apache.streampark.flink.packer.pipeline.PipelineStatusEnum;
+import org.apache.streampark.flink.packer.pipeline.PipelineTypeEnum;
 import org.apache.streampark.flink.packer.pipeline.impl.FlinkK8sApplicationBuildPipeline;
 import org.apache.streampark.flink.packer.pipeline.impl.FlinkK8sApplicationBuildPipelineV2;
 import org.apache.streampark.flink.packer.pipeline.impl.FlinkK8sSessionBuildPipeline;
@@ -115,7 +115,7 @@ import java.util.concurrent.ThreadPoolExecutor;
 import java.util.concurrent.TimeUnit;
 import java.util.stream.Collectors;
 
-import static org.apache.streampark.console.core.enums.Operation.RELEASE;
+import static org.apache.streampark.console.core.enums.OperationEnum.RELEASE;
 
 @Service
 @Slf4j
@@ -200,7 +200,7 @@ public class AppBuildPipeServiceImpl
     }
 
     // 1) flink sql setDependency
-    FlinkSql newFlinkSql = flinkSqlService.getCandidate(app.getId(), CandidateType.NEW);
+    FlinkSql newFlinkSql = flinkSqlService.getCandidate(app.getId(), CandidateTypeEnum.NEW);
     FlinkSql effectiveFlinkSql = flinkSqlService.getEffective(app.getId(), false);
     if (app.isFlinkSqlJobOrPyFlinkJob()) {
       FlinkSql flinkSql = newFlinkSql == null ? effectiveFlinkSql : newFlinkSql;
@@ -224,7 +224,7 @@ public class AppBuildPipeServiceImpl
                 AppBuildPipeline.fromPipeSnapshot(snapshot).setAppId(app.getId());
             saveEntity(buildPipeline);
 
-            app.setRelease(ReleaseState.RELEASING.get());
+            app.setRelease(ReleaseStateEnum.RELEASING.get());
             applicationManageService.updateRelease(app);
 
             if (flinkAppHttpWatcher.isWatchingApp(app.getId())) {
@@ -313,10 +313,10 @@ public class AppBuildPipeServiceImpl
             if (result.pass()) {
               // running job ...
               if (app.isRunning()) {
-                app.setRelease(ReleaseState.NEED_RESTART.get());
+                app.setRelease(ReleaseStateEnum.NEED_RESTART.get());
               } else {
-                app.setOptionState(OptionState.NONE.getValue());
-                app.setRelease(ReleaseState.DONE.get());
+                app.setOptionState(OptionStateEnum.NONE.getValue());
+                app.setRelease(ReleaseStateEnum.DONE.get());
                 // If the current task is not running, or the task has just been added, directly set
                 // the candidate version to the official version
                 if (app.isFlinkSqlJob()) {
@@ -349,10 +349,10 @@ public class AppBuildPipeServiceImpl
                       app.getId(),
                       app.getJobName().concat(" release failed"),
                       ExceptionUtils.stringifyException(snapshot.error().exception()),
-                      NoticeType.EXCEPTION);
+                      NoticeTypeEnum.EXCEPTION);
               messageService.push(message);
-              app.setRelease(ReleaseState.FAILED.get());
-              app.setOptionState(OptionState.NONE.getValue());
+              app.setRelease(ReleaseStateEnum.FAILED.get());
+              app.setOptionState(OptionStateEnum.NONE.getValue());
               app.setBuild(true);
               applicationLog.setException(
                   ExceptionUtils.stringifyException(snapshot.error().exception()));
@@ -366,7 +366,7 @@ public class AppBuildPipeServiceImpl
           }
         });
     // save docker resolve progress detail to cache, only for flink-k8s application mode.
-    if (PipelineType.FLINK_NATIVE_K8S_APPLICATION == pipeline.pipeType()) {
+    if (PipelineTypeEnum.FLINK_NATIVE_K8S_APPLICATION == pipeline.pipeType()) {
       pipeline
           .as(FlinkK8sApplicationBuildPipeline.class)
           .registerDockerProgressWatcher(
@@ -439,14 +439,14 @@ public class AppBuildPipeServiceImpl
       }
     }
 
-    ExecutionMode executionMode = app.getExecutionModeEnum();
+    ExecutionModeEnum executionModeEnum = app.getExecutionModeEnum();
     String mainClass = ConfigConst.STREAMPARK_FLINKSQL_CLIENT_CLASS();
-    switch (executionMode) {
+    switch (executionModeEnum) {
       case YARN_APPLICATION:
         String yarnProvidedPath = app.getAppLib();
         String localWorkspace = app.getLocalAppHome().concat("/lib");
-        if (DevelopmentMode.CUSTOM_CODE == app.getDevelopmentMode()
-            && ApplicationType.APACHE_FLINK == app.getApplicationType()) {
+        if (DevelopmentModeEnum.CUSTOM_CODE == app.getDevelopmentMode()
+            && ApplicationTypeEnum.APACHE_FLINK == app.getApplicationType()) {
           yarnProvidedPath = app.getAppHome();
           localWorkspace = app.getLocalAppHome();
         }
@@ -544,7 +544,7 @@ public class AppBuildPipeServiceImpl
         return String.format("%s/%s", app.getAppHome(), app.getJar());
       case FLINK_SQL:
         String sqlDistJar = commonService.getSqlClientJar(flinkEnv);
-        if (app.getExecutionModeEnum() == ExecutionMode.YARN_APPLICATION) {
+        if (app.getExecutionModeEnum() == ExecutionModeEnum.YARN_APPLICATION) {
           String clientPath = Workspace.remote().APP_CLIENT();
           return String.format("%s/%s", clientPath, sqlDistJar);
         }
@@ -571,12 +571,12 @@ public class AppBuildPipeServiceImpl
   @Override
   public boolean allowToBuildNow(@Nonnull Long appId) {
     return getCurrentBuildPipeline(appId)
-        .map(pipeline -> PipelineStatus.running != pipeline.getPipelineStatus())
+        .map(pipeline -> PipelineStatusEnum.running != pipeline.getPipelineStatus())
         .orElse(true);
   }
 
   @Override
-  public Map<Long, PipelineStatus> listPipelineStatus(List<Long> appIds) {
+  public Map<Long, PipelineStatusEnum> listPipelineStatus(List<Long> appIds) {
     if (CollectionUtils.isEmpty(appIds)) {
       return Collections.emptyMap();
     }
@@ -635,7 +635,7 @@ public class AppBuildPipeServiceImpl
               resourceId -> {
                 Resource resource = resourceService.getById(resourceId);
 
-                if (resource.getResourceType() != ResourceType.GROUP) {
+                if (resource.getResourceTypeEnum() != ResourceTypeEnum.GROUP) {
                   mergeDependency(application, mvnArtifacts, jarLibs, resource);
                 } else {
                   try {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationBackUpServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationBackUpServiceImpl.java
index 6afe5b45e..d7ae18d6e 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationBackUpServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationBackUpServiceImpl.java
@@ -26,8 +26,8 @@ import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ApplicationBackUp;
 import org.apache.streampark.console.core.entity.ApplicationConfig;
 import org.apache.streampark.console.core.entity.FlinkSql;
-import org.apache.streampark.console.core.enums.EffectiveType;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.mapper.ApplicationBackUpMapper;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
 import org.apache.streampark.console.core.service.ApplicationConfigService;
@@ -103,11 +103,12 @@ public class ApplicationBackUpServiceImpl
       // rollback to back up config
       configService.setLatestOrEffective(true, bakParam.getId(), bakParam.getAppId());
     } else {
-      effectiveService.saveOrUpdate(bakParam.getAppId(), EffectiveType.CONFIG, bakParam.getId());
+      effectiveService.saveOrUpdate(
+          bakParam.getAppId(), EffectiveTypeEnum.CONFIG, bakParam.getId());
       // if flink sql task, will be rollback sql and dependencies
       if (application.isFlinkSqlJob()) {
         effectiveService.saveOrUpdate(
-            bakParam.getAppId(), EffectiveType.FLINKSQL, bakParam.getSqlId());
+            bakParam.getAppId(), EffectiveTypeEnum.FLINKSQL, bakParam.getSqlId());
       }
     }
 
@@ -123,7 +124,7 @@ public class ApplicationBackUpServiceImpl
         new UpdateWrapper<Application>()
             .lambda()
             .eq(Application::getId, application.getId())
-            .set(Application::getRelease, ReleaseState.NEED_RESTART.get()));
+            .set(Application::getRelease, ReleaseStateEnum.NEED_RESTART.get()));
   }
 
   @Override
@@ -173,8 +174,8 @@ public class ApplicationBackUpServiceImpl
     ApiAlertException.throwIfNull(
         backUp, "Application backup can't be null. Rollback flink sql failed.");
     // rollback config and sql
-    effectiveService.saveOrUpdate(backUp.getAppId(), EffectiveType.CONFIG, backUp.getId());
-    effectiveService.saveOrUpdate(backUp.getAppId(), EffectiveType.FLINKSQL, backUp.getSqlId());
+    effectiveService.saveOrUpdate(backUp.getAppId(), EffectiveTypeEnum.CONFIG, backUp.getId());
+    effectiveService.saveOrUpdate(backUp.getAppId(), EffectiveTypeEnum.FLINKSQL, backUp.getSqlId());
   }
 
   @Override
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationConfigServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationConfigServiceImpl.java
index 56f778b1c..d5f5da1f2 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationConfigServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ApplicationConfigServiceImpl.java
@@ -25,8 +25,8 @@ import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.base.mybatis.pager.MybatisPager;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ApplicationConfig;
-import org.apache.streampark.console.core.enums.ConfigFileType;
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.ConfigFileTypeEnum;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 import org.apache.streampark.console.core.mapper.ApplicationConfigMapper;
 import org.apache.streampark.console.core.service.ApplicationConfigService;
 import org.apache.streampark.console.core.service.EffectiveService;
@@ -73,8 +73,8 @@ public class ApplicationConfigServiceImpl
     applicationConfig.setAppId(application.getId());
 
     if (application.getFormat() != null) {
-      ConfigFileType fileType = ConfigFileType.of(application.getFormat());
-      if (fileType == null || ConfigFileType.UNKNOWN == fileType) {
+      ConfigFileTypeEnum fileType = ConfigFileTypeEnum.of(application.getFormat());
+      if (fileType == null || ConfigFileTypeEnum.UNKNOWN == fileType) {
         throw new ApiAlertException(
             "application' config error. must be (.properties|.yaml|.yml |.conf)");
       }
@@ -110,7 +110,7 @@ public class ApplicationConfigServiceImpl
       ApplicationConfig effectiveConfig = getEffective(application.getId());
       if (Utils.isEmpty(application.getConfig())) {
         if (effectiveConfig != null) {
-          effectiveService.delete(application.getId(), EffectiveType.CONFIG);
+          effectiveService.delete(application.getId(), EffectiveTypeEnum.CONFIG);
         }
       } else {
         // there was no configuration before, is a new configuration
@@ -180,7 +180,7 @@ public class ApplicationConfigServiceImpl
     LambdaUpdateWrapper<ApplicationConfig> updateWrapper = Wrappers.lambdaUpdate();
     updateWrapper.eq(ApplicationConfig::getAppId, appId).set(ApplicationConfig::getLatest, false);
     this.update(updateWrapper);
-    effectiveService.saveOrUpdate(appId, EffectiveType.CONFIG, configId);
+    effectiveService.saveOrUpdate(appId, EffectiveTypeEnum.CONFIG, configId);
   }
 
   @Override
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/EffectiveServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/EffectiveServiceImpl.java
index f721693ea..12afa4ec8 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/EffectiveServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/EffectiveServiceImpl.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.console.core.service.impl;
 
 import org.apache.streampark.console.core.entity.Effective;
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 import org.apache.streampark.console.core.mapper.EffectiveMapper;
 import org.apache.streampark.console.core.service.EffectiveService;
 
@@ -39,25 +39,25 @@ public class EffectiveServiceImpl extends ServiceImpl<EffectiveMapper, Effective
     implements EffectiveService {
 
   @Override
-  public void delete(Long appId, EffectiveType effectiveType) {
+  public void delete(Long appId, EffectiveTypeEnum effectiveTypeEnum) {
     LambdaQueryWrapper<Effective> queryWrapper =
         new LambdaQueryWrapper<Effective>()
             .eq(Effective::getAppId, appId)
-            .eq(Effective::getTargetType, effectiveType.getType());
+            .eq(Effective::getTargetType, effectiveTypeEnum.getType());
     baseMapper.delete(queryWrapper);
   }
 
   @Override
-  public Effective get(Long appId, EffectiveType effectiveType) {
+  public Effective get(Long appId, EffectiveTypeEnum effectiveTypeEnum) {
     LambdaQueryWrapper<Effective> queryWrapper =
         new LambdaQueryWrapper<Effective>()
             .eq(Effective::getAppId, appId)
-            .eq(Effective::getTargetType, effectiveType.getType());
+            .eq(Effective::getTargetType, effectiveTypeEnum.getType());
     return this.getOne(queryWrapper);
   }
 
   @Override
-  public void saveOrUpdate(Long appId, EffectiveType type, Long id) {
+  public void saveOrUpdate(Long appId, EffectiveTypeEnum type, Long id) {
     LambdaQueryWrapper<Effective> queryWrapper =
         new LambdaQueryWrapper<Effective>()
             .eq(Effective::getAppId, appId)
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ExternalLinkServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ExternalLinkServiceImpl.java
index 4fc211d49..2d78edd46 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ExternalLinkServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ExternalLinkServiceImpl.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.service.impl;
 import org.apache.streampark.common.util.Utils;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ExternalLink;
-import org.apache.streampark.console.core.enums.PlaceholderType;
+import org.apache.streampark.console.core.enums.PlaceholderTypeEnum;
 import org.apache.streampark.console.core.mapper.ExternalLinkMapper;
 import org.apache.streampark.console.core.service.ExternalLinkService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
@@ -86,9 +86,9 @@ public class ExternalLinkServiceImpl extends ServiceImpl<ExternalLinkMapper, Ext
 
   private void renderLinkUrl(ExternalLink link, Application app) {
     HashMap<String, String> map = new HashMap();
-    map.put(PlaceholderType.JOB_ID.get(), app.getJobId());
-    map.put(PlaceholderType.JOB_NAME.get(), app.getJobName());
-    map.put(PlaceholderType.YARN_ID.get(), app.getAppId());
+    map.put(PlaceholderTypeEnum.JOB_ID.get(), app.getJobId());
+    map.put(PlaceholderTypeEnum.JOB_NAME.get(), app.getJobName());
+    map.put(PlaceholderTypeEnum.YARN_ID.get(), app.getAppId());
     PropertyPlaceholderHelper propertyPlaceholderHelper = new PropertyPlaceholderHelper("{", "}");
     link.setRenderedLinkUrl(
         propertyPlaceholderHelper.replacePlaceholders(link.getLinkUrl().trim(), map::get));
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkClusterServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkClusterServiceImpl.java
index d7f62ab08..affe3c246 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkClusterServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkClusterServiceImpl.java
@@ -18,8 +18,8 @@
 package org.apache.streampark.console.core.service.impl;
 
 import org.apache.streampark.common.conf.K8sFlinkConfig;
-import org.apache.streampark.common.enums.ClusterState;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ClusterStateEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.ThreadUtils;
 import org.apache.streampark.common.util.YarnUtils;
 import org.apache.streampark.console.base.exception.ApiAlertException;
@@ -123,13 +123,13 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
     }
 
     // 3) Check connection
-    if (ExecutionMode.isRemoteMode(cluster.getExecutionModeEnum())
+    if (ExecutionModeEnum.isRemoteMode(cluster.getExecutionModeEnum())
         && !flinkClusterWatcher.verifyClusterConnection(cluster)) {
       result.setMsg("The remote cluster connection failed, please check!");
       result.setStatus(3);
       return result;
     }
-    if (ExecutionMode.isYarnMode(cluster.getExecutionModeEnum())
+    if (ExecutionModeEnum.isYarnMode(cluster.getExecutionModeEnum())
         && cluster.getClusterId() != null
         && !flinkClusterWatcher.verifyClusterConnection(cluster)) {
       result.setMsg("The flink cluster connection failed, please check!");
@@ -152,15 +152,15 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
     ApiAlertException.throwIfFalse(
         successful, String.format(ERROR_CLUSTER_QUEUE_HINT, flinkCluster.getYarnQueue()));
     flinkCluster.setCreateTime(new Date());
-    if (ExecutionMode.isRemoteMode(flinkCluster.getExecutionModeEnum())) {
-      flinkCluster.setClusterState(ClusterState.RUNNING.getState());
+    if (ExecutionModeEnum.isRemoteMode(flinkCluster.getExecutionModeEnum())) {
+      flinkCluster.setClusterState(ClusterStateEnum.RUNNING.getState());
       flinkCluster.setStartTime(new Date());
       flinkCluster.setEndTime(null);
     } else {
-      flinkCluster.setClusterState(ClusterState.CREATED.getState());
+      flinkCluster.setClusterState(ClusterStateEnum.CREATED.getState());
     }
     boolean ret = save(flinkCluster);
-    if (ret && ExecutionMode.isRemoteMode(flinkCluster.getExecutionMode())) {
+    if (ret && ExecutionModeEnum.isRemoteMode(flinkCluster.getExecutionMode())) {
       FlinkClusterWatcher.addWatching(flinkCluster);
     }
     if (shouldWatchForK8s(flinkCluster)) {
@@ -178,7 +178,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
       ApiAlertException.throwIfNull(
           deployResponse,
           "Deploy cluster failed, unknown reason,please check you params or StreamPark error log");
-      if (ExecutionMode.isYarnSessionMode(flinkCluster.getExecutionModeEnum())) {
+      if (ExecutionModeEnum.isYarnSessionMode(flinkCluster.getExecutionModeEnum())) {
         String address =
             String.format(
                 "%s/proxy/%s/", YarnUtils.getRMWebAppURL(true), deployResponse.clusterId());
@@ -188,7 +188,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
         flinkCluster.setAddress(deployResponse.address());
       }
       flinkCluster.setClusterId(deployResponse.clusterId());
-      flinkCluster.setClusterState(ClusterState.RUNNING.getState());
+      flinkCluster.setClusterState(ClusterStateEnum.RUNNING.getState());
       flinkCluster.setException(null);
       flinkCluster.setEndTime(null);
       updateById(flinkCluster);
@@ -198,7 +198,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
       }
     } catch (Exception e) {
       log.error(e.getMessage(), e);
-      flinkCluster.setClusterState(ClusterState.FAILED.getState());
+      flinkCluster.setClusterState(ClusterStateEnum.FAILED.getState());
       flinkCluster.setException(e.toString());
       updateById(flinkCluster);
       throw new ApiDetailException(e);
@@ -215,9 +215,9 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
     flinkCluster.setClusterName(paramOfCluster.getClusterName());
     flinkCluster.setAlertId(paramOfCluster.getAlertId());
     flinkCluster.setDescription(paramOfCluster.getDescription());
-    if (ExecutionMode.isRemoteMode(flinkCluster.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isRemoteMode(flinkCluster.getExecutionModeEnum())) {
       flinkCluster.setAddress(paramOfCluster.getAddress());
-      flinkCluster.setClusterState(ClusterState.RUNNING.getState());
+      flinkCluster.setClusterState(ClusterStateEnum.RUNNING.getState());
       flinkCluster.setStartTime(new Date());
       flinkCluster.setEndTime(null);
       FlinkClusterWatcher.addWatching(flinkCluster);
@@ -249,7 +249,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
       ShutDownResponse shutDownResponse =
           shutdownInternal(flinkCluster, flinkCluster.getClusterId());
       ApiAlertException.throwIfNull(shutDownResponse, "Get shutdown response failed");
-      flinkCluster.setClusterState(ClusterState.CANCELED.getState());
+      flinkCluster.setClusterState(ClusterStateEnum.CANCELED.getState());
       flinkCluster.setEndTime(new Date());
       updateById(flinkCluster);
       FlinkClusterWatcher.unWatching(flinkCluster);
@@ -305,19 +305,19 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
   }
 
   @Override
-  public List<FlinkCluster> getByExecutionModes(Collection<ExecutionMode> executionModes) {
+  public List<FlinkCluster> getByExecutionModes(Collection<ExecutionModeEnum> executionModeEnums) {
     return getBaseMapper()
         .selectList(
             new LambdaQueryWrapper<FlinkCluster>()
                 .in(
                     FlinkCluster::getExecutionMode,
-                    executionModes.stream()
-                        .map(ExecutionMode::getMode)
+                    executionModeEnums.stream()
+                        .map(ExecutionModeEnum::getMode)
                         .collect(Collectors.toSet())));
   }
 
   @Override
-  public void updateClusterState(Long id, ClusterState state) {
+  public void updateClusterState(Long id, ClusterStateEnum state) {
     LambdaUpdateWrapper<FlinkCluster> updateWrapper =
         new LambdaUpdateWrapper<FlinkCluster>()
             .eq(FlinkCluster::getId, id)
@@ -347,10 +347,10 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
     FlinkCluster flinkCluster = getById(id);
     ApiAlertException.throwIfNull(flinkCluster, "Flink cluster not exist, please check.");
 
-    if (ExecutionMode.isYarnSessionMode(flinkCluster.getExecutionModeEnum())
-        || ExecutionMode.isKubernetesSessionMode(flinkCluster.getExecutionMode())) {
+    if (ExecutionModeEnum.isYarnSessionMode(flinkCluster.getExecutionModeEnum())
+        || ExecutionModeEnum.isKubernetesSessionMode(flinkCluster.getExecutionMode())) {
       ApiAlertException.throwIfTrue(
-          ClusterState.isRunning(flinkCluster.getClusterStateEnum()),
+          ClusterStateEnum.isRunning(flinkCluster.getClusterStateEnum()),
           "Flink cluster is running, cannot be delete, please check.");
     }
     if (shouldWatchForK8s(flinkCluster)) {
@@ -393,7 +393,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
       return true;
     }
 
-    if (ExecutionMode.isYarnSessionMode(newCluster.getExecutionModeEnum())
+    if (ExecutionModeEnum.isYarnSessionMode(newCluster.getExecutionModeEnum())
         && StringUtils.equals(oldCluster.getYarnQueue(), newCluster.getYarnQueue())) {
       return true;
     }
@@ -409,7 +409,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
    *     default), return true, false else.
    */
   private boolean isYarnNotDefaultQueue(FlinkCluster cluster) {
-    return ExecutionMode.isYarnSessionMode(cluster.getExecutionModeEnum())
+    return ExecutionModeEnum.isYarnSessionMode(cluster.getExecutionModeEnum())
         && !yarnQueueService.isDefaultQueue(cluster.getYarnQueue());
   }
 
@@ -442,12 +442,12 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
   }
 
   private void checkActiveIfNeeded(FlinkCluster flinkCluster) {
-    if (ExecutionMode.isYarnSessionMode(flinkCluster.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isYarnSessionMode(flinkCluster.getExecutionModeEnum())) {
       ApiAlertException.throwIfFalse(
-          ClusterState.isRunning(flinkCluster.getClusterStateEnum()),
+          ClusterStateEnum.isRunning(flinkCluster.getClusterStateEnum()),
           "Current cluster is not active, please check!");
       if (!flinkClusterWatcher.verifyClusterConnection(flinkCluster)) {
-        flinkCluster.setClusterState(ClusterState.LOST.getState());
+        flinkCluster.setClusterState(ClusterStateEnum.LOST.getState());
         updateById(flinkCluster);
         throw new ApiAlertException("Current cluster is not active, please check!");
       }
@@ -457,7 +457,7 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
   @Nullable
   private KubernetesDeployParam getKubernetesDeployDesc(
       @Nonnull FlinkCluster flinkCluster, String action) {
-    ExecutionMode executionModeEnum = flinkCluster.getExecutionModeEnum();
+    ExecutionModeEnum executionModeEnum = flinkCluster.getExecutionModeEnum();
     switch (executionModeEnum) {
       case YARN_SESSION:
         break;
@@ -479,6 +479,6 @@ public class FlinkClusterServiceImpl extends ServiceImpl<FlinkClusterMapper, Fli
 
   private boolean shouldWatchForK8s(FlinkCluster flinkCluster) {
     return K8sFlinkConfig.isV2Enabled()
-        && ExecutionMode.isKubernetesSessionMode(flinkCluster.getExecutionMode());
+        && ExecutionModeEnum.isKubernetesSessionMode(flinkCluster.getExecutionMode());
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkSqlServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkSqlServiceImpl.java
index 0c4abe195..ded40a9a1 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkSqlServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/FlinkSqlServiceImpl.java
@@ -26,8 +26,8 @@ import org.apache.streampark.console.base.mybatis.pager.MybatisPager;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.FlinkSql;
-import org.apache.streampark.console.core.enums.CandidateType;
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.CandidateTypeEnum;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 import org.apache.streampark.console.core.mapper.FlinkSqlMapper;
 import org.apache.streampark.console.core.service.ApplicationBackUpService;
 import org.apache.streampark.console.core.service.EffectiveService;
@@ -102,20 +102,20 @@ public class FlinkSqlServiceImpl extends ServiceImpl<FlinkSqlMapper, FlinkSql>
     String sql = DeflaterUtils.zipString(flinkSql.getSql());
     flinkSql.setSql(sql);
     this.save(flinkSql);
-    this.setCandidate(CandidateType.NEW, flinkSql.getAppId(), flinkSql.getId());
+    this.setCandidate(CandidateTypeEnum.NEW, flinkSql.getAppId(), flinkSql.getId());
   }
 
   @Override
-  public void setCandidate(CandidateType candidateType, Long appId, Long sqlId) {
+  public void setCandidate(CandidateTypeEnum candidateTypeEnum, Long appId, Long sqlId) {
     this.update(
         new LambdaUpdateWrapper<FlinkSql>()
             .eq(FlinkSql::getAppId, appId)
-            .set(FlinkSql::getCandidate, CandidateType.NONE.get()));
+            .set(FlinkSql::getCandidate, CandidateTypeEnum.NONE.get()));
 
     this.update(
         new LambdaUpdateWrapper<FlinkSql>()
             .eq(FlinkSql::getId, sqlId)
-            .set(FlinkSql::getCandidate, candidateType.get()));
+            .set(FlinkSql::getCandidate, candidateTypeEnum.get()));
   }
 
   @Override
@@ -139,20 +139,20 @@ public class FlinkSqlServiceImpl extends ServiceImpl<FlinkSqlMapper, FlinkSql>
   }
 
   @Override
-  public FlinkSql getCandidate(Long appId, CandidateType candidateType) {
+  public FlinkSql getCandidate(Long appId, CandidateTypeEnum candidateTypeEnum) {
     LambdaQueryWrapper<FlinkSql> queryWrapper =
         new LambdaQueryWrapper<FlinkSql>().eq(FlinkSql::getAppId, appId);
-    if (candidateType == null) {
-      queryWrapper.gt(FlinkSql::getCandidate, CandidateType.NONE.get());
+    if (candidateTypeEnum == null) {
+      queryWrapper.gt(FlinkSql::getCandidate, CandidateTypeEnum.NONE.get());
     } else {
-      queryWrapper.eq(FlinkSql::getCandidate, candidateType.get());
+      queryWrapper.eq(FlinkSql::getCandidate, candidateTypeEnum.get());
     }
     return baseMapper.selectOne(queryWrapper);
   }
 
   @Override
   public void toEffective(Long appId, Long sqlId) {
-    effectiveService.saveOrUpdate(appId, EffectiveType.FLINKSQL, sqlId);
+    effectiveService.saveOrUpdate(appId, EffectiveTypeEnum.FLINKSQL, sqlId);
   }
 
   @Override
@@ -160,7 +160,7 @@ public class FlinkSqlServiceImpl extends ServiceImpl<FlinkSqlMapper, FlinkSql>
     this.update(
         new LambdaUpdateWrapper<FlinkSql>()
             .eq(FlinkSql::getId, id)
-            .set(FlinkSql::getCandidate, CandidateType.NONE.get()));
+            .set(FlinkSql::getCandidate, CandidateTypeEnum.NONE.get()));
   }
 
   @Override
@@ -173,7 +173,7 @@ public class FlinkSqlServiceImpl extends ServiceImpl<FlinkSqlMapper, FlinkSql>
   @Override
   @Transactional(propagation = Propagation.REQUIRES_NEW, rollbackFor = Exception.class)
   public void rollback(Application application) {
-    FlinkSql sql = getCandidate(application.getId(), CandidateType.HISTORY);
+    FlinkSql sql = getCandidate(application.getId(), CandidateTypeEnum.HISTORY);
     Utils.notNull(sql);
     try {
       // check and backup current job
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/MessageServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/MessageServiceImpl.java
index 5a60fd968..128c625a5 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/MessageServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/MessageServiceImpl.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.service.impl;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.mybatis.pager.MybatisPager;
 import org.apache.streampark.console.core.entity.Message;
-import org.apache.streampark.console.core.enums.NoticeType;
+import org.apache.streampark.console.core.enums.NoticeTypeEnum;
 import org.apache.streampark.console.core.mapper.MessageMapper;
 import org.apache.streampark.console.core.service.MessageService;
 import org.apache.streampark.console.core.websocket.WebSocketEndpoint;
@@ -47,13 +47,13 @@ public class MessageServiceImpl extends ServiceImpl<MessageMapper, Message>
   }
 
   @Override
-  public IPage<Message> getUnRead(NoticeType noticeType, RestRequest request) {
+  public IPage<Message> getUnRead(NoticeTypeEnum noticeTypeEnum, RestRequest request) {
     Page<Message> page = new MybatisPager<Message>().getDefaultPage(request);
     LambdaQueryWrapper<Message> queryWrapper =
         new LambdaQueryWrapper<Message>()
             .eq(Message::getIsRead, false)
             .orderByDesc(Message::getCreateTime)
-            .eq(Message::getType, noticeType.get());
+            .eq(Message::getType, noticeTypeEnum.get());
     return this.baseMapper.selectPage(page, queryWrapper);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ProjectServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ProjectServiceImpl.java
index 416e162b8..0970b769d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ProjectServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ProjectServiceImpl.java
@@ -32,9 +32,9 @@ import org.apache.streampark.console.base.mybatis.pager.MybatisPager;
 import org.apache.streampark.console.base.util.GZipUtils;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.Project;
-import org.apache.streampark.console.core.enums.BuildState;
-import org.apache.streampark.console.core.enums.GitCredential;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.BuildStateEnum;
+import org.apache.streampark.console.core.enums.GitCredentialEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.mapper.ProjectMapper;
 import org.apache.streampark.console.core.service.ProjectService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
@@ -122,7 +122,7 @@ public class ProjectServiceImpl extends ServiceImpl<ProjectMapper, Project>
         project.getTeamId().equals(projectParam.getTeamId()),
         "TeamId can't be changed, update project failed.");
     ApiAlertException.throwIfFalse(
-        !project.getBuildState().equals(BuildState.BUILDING.get()),
+        !project.getBuildState().equals(BuildStateEnum.BUILDING.get()),
         "The project is being built, update project failed.");
     project.setName(projectParam.getName());
     project.setUrl(projectParam.getUrl());
@@ -134,21 +134,21 @@ public class ProjectServiceImpl extends ServiceImpl<ProjectMapper, Project>
     project.setPom(projectParam.getPom());
     project.setDescription(projectParam.getDescription());
     project.setBuildArgs(projectParam.getBuildArgs());
-    if (GitCredential.isSSH(project.getGitCredential())) {
+    if (GitCredentialEnum.isSSH(project.getGitCredential())) {
       project.setUserName(null);
     } else {
       project.setPrvkeyPath(null);
     }
     if (projectParam.getBuildState() != null) {
       project.setBuildState(projectParam.getBuildState());
-      if (BuildState.NEED_REBUILD == BuildState.of(projectParam.getBuildState())) {
+      if (BuildStateEnum.NEED_REBUILD == BuildStateEnum.of(projectParam.getBuildState())) {
         List<Application> applications = getApplications(project);
         // Update deployment status
         applications.forEach(
             (app) -> {
               log.info(
                   "update deploy by project: {}, appName:{}", project.getName(), app.getJobName());
-              app.setRelease(ReleaseState.NEED_CHECK.get());
+              app.setRelease(ReleaseStateEnum.NEED_CHECK.get());
               applicationManageService.updateRelease(app);
             });
       }
@@ -196,15 +196,15 @@ public class ProjectServiceImpl extends ServiceImpl<ProjectMapper, Project>
   @Override
   public void build(Long id) throws Exception {
     Project project = getById(id);
-    this.baseMapper.updateBuildState(project.getId(), BuildState.BUILDING.get());
+    this.baseMapper.updateBuildState(project.getId(), BuildStateEnum.BUILDING.get());
     String logPath = getBuildLogPath(id);
     ProjectBuildTask projectBuildTask =
         new ProjectBuildTask(
             logPath,
             project,
-            buildState -> {
-              baseMapper.updateBuildState(id, buildState.get());
-              if (buildState == BuildState.SUCCESSFUL) {
+            buildStateEnum -> {
+              baseMapper.updateBuildState(id, buildStateEnum.get());
+              if (buildStateEnum == BuildStateEnum.SUCCESSFUL) {
                 baseMapper.updateBuildTime(id);
               }
               flinkAppHttpWatcher.init();
@@ -218,7 +218,7 @@ public class ProjectServiceImpl extends ServiceImpl<ProjectMapper, Project>
                         "update deploy by project: {}, appName:{}",
                         project.getName(),
                         app.getJobName());
-                    app.setRelease(ReleaseState.NEED_RELEASE.get());
+                    app.setRelease(ReleaseStateEnum.NEED_RELEASE.get());
                     app.setBuild(true);
                     this.applicationManageService.updateRelease(app);
                   });
@@ -235,7 +235,7 @@ public class ProjectServiceImpl extends ServiceImpl<ProjectMapper, Project>
     Project project = getById(id);
     Utils.notNull(project);
 
-    if (BuildState.SUCCESSFUL != BuildState.of(project.getBuildState())
+    if (BuildStateEnum.SUCCESSFUL != BuildStateEnum.of(project.getBuildState())
         || !project.getDistHome().exists()) {
       return Collections.emptyList();
     }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ResourceServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ResourceServiceImpl.java
index 30fb4fdf8..3a5e885ff 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ResourceServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/ResourceServiceImpl.java
@@ -35,7 +35,7 @@ import org.apache.streampark.console.core.bean.MavenPom;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkSql;
 import org.apache.streampark.console.core.entity.Resource;
-import org.apache.streampark.console.core.enums.ResourceType;
+import org.apache.streampark.console.core.enums.ResourceTypeEnum;
 import org.apache.streampark.console.core.mapper.ResourceMapper;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.core.service.FlinkSqlService;
@@ -131,13 +131,13 @@ public class ResourceServiceImpl extends ServiceImpl<ResourceMapper, Resource>
         jars.isEmpty() && poms.isEmpty(), "Please add pom or jar resource.");
 
     ApiAlertException.throwIfTrue(
-        resource.getResourceType() == ResourceType.FLINK_APP && jars.isEmpty(),
+        resource.getResourceTypeEnum() == ResourceTypeEnum.FLINK_APP && jars.isEmpty(),
         "Please upload jar for Flink_App resource");
 
     ApiAlertException.throwIfTrue(
         jars.size() + poms.size() > 1, "Please do not add multi dependency at one time.");
 
-    if (resource.getResourceType() != ResourceType.CONNECTOR) {
+    if (resource.getResourceTypeEnum() != ResourceTypeEnum.CONNECTOR) {
       ApiAlertException.throwIfNull(resource.getResourceName(), "The resourceName is required.");
     } else {
       String connector = resource.getConnector();
@@ -273,7 +273,7 @@ public class ResourceServiceImpl extends ServiceImpl<ResourceMapper, Resource>
 
   @Override
   public RestResponse checkResource(Resource resourceParam) throws JsonProcessingException {
-    ResourceType type = resourceParam.getResourceType();
+    ResourceTypeEnum type = resourceParam.getResourceTypeEnum();
     Map<String, Serializable> resp = new HashMap<>(0);
     resp.put("state", 0);
     switch (type) {
@@ -305,7 +305,7 @@ public class ResourceServiceImpl extends ServiceImpl<ResourceMapper, Resource>
         FlinkConnector connectorResource;
 
         ApiAlertException.throwIfFalse(
-            ResourceType.CONNECTOR == resourceParam.getResourceType(),
+            ResourceTypeEnum.CONNECTOR == resourceParam.getResourceTypeEnum(),
             "getConnectorId method error, resource not flink connector.");
 
         List<File> jars;
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SavePointServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SavePointServiceImpl.java
index 526ef2565..d8aa71a9c 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SavePointServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SavePointServiceImpl.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service.impl;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.CompletableFutureUtils;
 import org.apache.streampark.common.util.ExceptionUtils;
 import org.apache.streampark.common.util.ThreadUtils;
@@ -34,9 +34,9 @@ import org.apache.streampark.console.core.entity.ApplicationLog;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.SavePoint;
-import org.apache.streampark.console.core.enums.CheckPointType;
-import org.apache.streampark.console.core.enums.Operation;
-import org.apache.streampark.console.core.enums.OptionState;
+import org.apache.streampark.console.core.enums.CheckPointTypeEnum;
+import org.apache.streampark.console.core.enums.OperationEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
 import org.apache.streampark.console.core.mapper.SavePointMapper;
 import org.apache.streampark.console.core.service.ApplicationConfigService;
 import org.apache.streampark.console.core.service.ApplicationLogService;
@@ -84,7 +84,7 @@ import java.util.concurrent.TimeoutException;
 import static org.apache.flink.configuration.CheckpointingOptions.MAX_RETAINED_CHECKPOINTS;
 import static org.apache.flink.configuration.CheckpointingOptions.SAVEPOINT_DIRECTORY;
 import static org.apache.streampark.common.util.PropertiesUtils.extractDynamicPropertiesAsJava;
-import static org.apache.streampark.console.core.enums.CheckPointType.CHECKPOINT;
+import static org.apache.streampark.console.core.enums.CheckPointTypeEnum.CHECKPOINT;
 
 @Slf4j
 @Service
@@ -171,7 +171,7 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
     Application application = applicationManageService.getById(appId);
 
     ApplicationLog applicationLog = new ApplicationLog();
-    applicationLog.setOptionName(Operation.SAVEPOINT.getValue());
+    applicationLog.setOptionName(OperationEnum.SAVEPOINT.getValue());
     applicationLog.setAppId(application.getId());
     applicationLog.setJobManagerUrl(application.getJobManagerUrl());
     applicationLog.setOptionTime(new Date());
@@ -179,7 +179,7 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
 
     FlinkAppHttpWatcher.addSavepoint(application.getId());
 
-    application.setOptionState(OptionState.SAVEPOINTING.getValue());
+    application.setOptionState(OptionStateEnum.SAVEPOINTING.getValue());
     application.setOptionTime(new Date());
     this.applicationManageService.updateById(application);
     flinkAppHttpWatcher.init();
@@ -264,7 +264,7 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
         .whenComplete(
             (t, e) -> {
               applicationLogService.save(applicationLog);
-              application.setOptionState(OptionState.NONE.getValue());
+              application.setOptionState(OptionStateEnum.NONE.getValue());
               application.setOptionTime(new Date());
               applicationManageService.update(application);
               flinkAppHttpWatcher.init();
@@ -290,7 +290,7 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
   private Map<String, Object> tryGetRestProps(Application application, FlinkCluster cluster) {
     Map<String, Object> properties = new HashMap<>();
 
-    if (ExecutionMode.isRemoteMode(application.getExecutionModeEnum())) {
+    if (ExecutionModeEnum.isRemoteMode(application.getExecutionModeEnum())) {
       Utils.notNull(
           cluster,
           String.format(
@@ -304,10 +304,10 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
   }
 
   private String getClusterId(Application application, FlinkCluster cluster) {
-    if (ExecutionMode.isKubernetesMode(application.getExecutionMode())) {
+    if (ExecutionModeEnum.isKubernetesMode(application.getExecutionMode())) {
       return application.getClusterId();
-    } else if (ExecutionMode.isYarnMode(application.getExecutionMode())) {
-      if (ExecutionMode.YARN_SESSION == application.getExecutionModeEnum()) {
+    } else if (ExecutionModeEnum.isYarnMode(application.getExecutionMode())) {
+      if (ExecutionModeEnum.YARN_SESSION == application.getExecutionModeEnum()) {
         Utils.notNull(
             cluster,
             String.format(
@@ -368,7 +368,7 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
   public String getSavepointFromDeployLayer(Application application)
       throws JsonProcessingException {
     // At the yarn or k8s mode, then read the savepoint in flink-conf.yml in the bound flink
-    if (!ExecutionMode.isRemoteMode(application.getExecutionMode())) {
+    if (!ExecutionModeEnum.isRemoteMode(application.getExecutionMode())) {
       FlinkEnv flinkEnv = flinkEnvService.getById(application.getVersionId());
       return flinkEnv.convertFlinkYamlAsMap().get(SAVEPOINT_DIRECTORY.key());
     }
@@ -446,7 +446,8 @@ public class SavePointServiceImpl extends ServiceImpl<SavePointMapper, SavePoint
     int cpThreshold =
         tryGetChkNumRetainedFromDynamicProps(application.getDynamicProperties())
             .orElse(getChkNumRetainedFromFlinkEnv(flinkEnv, application));
-    cpThreshold = CHECKPOINT == CheckPointType.of(entity.getType()) ? cpThreshold - 1 : cpThreshold;
+    cpThreshold =
+        CHECKPOINT == CheckPointTypeEnum.of(entity.getType()) ? cpThreshold - 1 : cpThreshold;
 
     if (cpThreshold == 0) {
       LambdaQueryWrapper<SavePoint> queryWrapper =
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SqlWorkBenchServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SqlWorkBenchServiceImpl.java
index 3b7be04ae..79ac65ca1 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SqlWorkBenchServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/SqlWorkBenchServiceImpl.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service.impl;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.HadoopConfigUtils;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.FlinkEnv;
@@ -27,7 +27,7 @@ import org.apache.streampark.console.core.service.FlinkEnvService;
 import org.apache.streampark.console.core.service.FlinkGateWayService;
 import org.apache.streampark.console.core.service.SqlWorkBenchService;
 import org.apache.streampark.flink.kubernetes.KubernetesRetriever;
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode;
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum;
 import org.apache.streampark.flink.kubernetes.ingress.IngressController;
 import org.apache.streampark.gateway.OperationHandle;
 import org.apache.streampark.gateway.factories.FactoryUtil;
@@ -55,10 +55,6 @@ import java.util.Map;
 import java.util.Objects;
 import java.util.UUID;
 
-import static org.apache.streampark.common.enums.ExecutionMode.KUBERNETES_NATIVE_SESSION;
-import static org.apache.streampark.common.enums.ExecutionMode.REMOTE;
-import static org.apache.streampark.common.enums.ExecutionMode.YARN_SESSION;
-
 @Slf4j
 @Service
 @RequiredArgsConstructor
@@ -103,13 +99,13 @@ public class SqlWorkBenchServiceImpl implements SqlWorkBenchService {
     String port = String.valueOf(remoteURI.getPort());
     String clusterId = flinkCluster.getClusterId();
 
-    ExecutionMode executionMode = ExecutionMode.of(flinkCluster.getExecutionMode());
-    if (executionMode == null) {
+    ExecutionModeEnum executionModeEnum = ExecutionModeEnum.of(flinkCluster.getExecutionMode());
+    if (executionModeEnum == null) {
       throw new IllegalArgumentException("executionMode is null");
     }
 
-    streamParkConf.put("execution.target", executionMode.getName());
-    switch (Objects.requireNonNull(executionMode)) {
+    streamParkConf.put("execution.target", executionModeEnum.getName());
+    switch (Objects.requireNonNull(executionModeEnum)) {
       case REMOTE:
         streamParkConf.put("rest.address", host);
         streamParkConf.put("rest.port", port);
@@ -125,7 +121,7 @@ public class SqlWorkBenchServiceImpl implements SqlWorkBenchService {
         try (ClusterClient<?> clusterClient =
             (ClusterClient<?>)
                 KubernetesRetriever.newFinkClusterClient(
-                    clusterId, k8sNamespace, FlinkK8sExecuteMode.of(executionMode))) {
+                    clusterId, k8sNamespace, FlinkK8sExecuteModeEnum.of(executionModeEnum))) {
           restAddress = IngressController.ingressUrlAddress(k8sNamespace, clusterId, clusterClient);
         } catch (Exception e) {
           throw new IllegalArgumentException("get k8s rest address error", e);
@@ -137,7 +133,7 @@ public class SqlWorkBenchServiceImpl implements SqlWorkBenchService {
         streamParkConf.put("rest.address", restAddress);
         break;
       default:
-        throw new IllegalArgumentException("Unsupported execution mode: " + executionMode);
+        throw new IllegalArgumentException("Unsupported execution mode: " + executionModeEnum);
     }
 
     return sqlGateWayService.openSession(
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/VariableServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/VariableServiceImpl.java
index 830fb3de1..e4de5d58f 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/VariableServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/VariableServiceImpl.java
@@ -24,7 +24,7 @@ import org.apache.streampark.console.base.mybatis.pager.MybatisPager;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkSql;
 import org.apache.streampark.console.core.entity.Variable;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.mapper.VariableMapper;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.core.service.FlinkSqlService;
@@ -142,7 +142,7 @@ public class VariableServiceImpl extends ServiceImpl<VariableMapper, Variable>
               .in(
                   Application::getId,
                   applications.stream().map(Application::getId).collect(Collectors.toList()))
-              .set(Application::getRelease, ReleaseState.NEED_RESTART.get()));
+              .set(Application::getRelease, ReleaseStateEnum.NEED_RESTART.get()));
     }
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/YarnQueueServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/YarnQueueServiceImpl.java
index ef9dd4f18..1eed75625 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/YarnQueueServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/service/impl/YarnQueueServiceImpl.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service.impl;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.Utils;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.exception.ApiAlertException;
@@ -172,12 +172,12 @@ public class YarnQueueServiceImpl extends ServiceImpl<YarnQueueMapper, YarnQueue
    * Only check the validation of queue-labelExpression when using yarn application or yarn-session
    * mode or yarn-perjob mode.
    *
-   * @param executionMode execution mode.
+   * @param executionModeEnum execution mode.
    * @param queueLabel queueLabel expression.
    */
   @Override
-  public void checkQueueLabel(ExecutionMode executionMode, String queueLabel) {
-    if (ExecutionMode.isYarnMode(executionMode)) {
+  public void checkQueueLabel(ExecutionModeEnum executionModeEnum, String queueLabel) {
+    if (ExecutionModeEnum.isYarnMode(executionModeEnum)) {
       ApiAlertException.throwIfFalse(isValid(queueLabel, true), ERR_FORMAT_HINTS);
     }
   }
@@ -217,7 +217,7 @@ public class YarnQueueServiceImpl extends ServiceImpl<YarnQueueMapper, YarnQueue
   public void checkNotReferencedByFlinkClusters(
       @Nonnull String queueLabel, @Nonnull String operation) {
     List<FlinkCluster> clustersReferenceYarnQueueLabel =
-        flinkClusterService.getByExecutionModes(Sets.newHashSet(ExecutionMode.YARN_SESSION))
+        flinkClusterService.getByExecutionModes(Sets.newHashSet(ExecutionModeEnum.YARN_SESSION))
             .stream()
             .filter(flinkCluster -> StringUtils.equals(flinkCluster.getYarnQueue(), queueLabel))
             .collect(Collectors.toList());
@@ -232,7 +232,8 @@ public class YarnQueueServiceImpl extends ServiceImpl<YarnQueueMapper, YarnQueue
     List<Application> appsReferenceQueueLabel =
         applicationManageService
             .getByTeamIdAndExecutionModes(
-                teamId, Sets.newHashSet(ExecutionMode.YARN_APPLICATION, ExecutionMode.YARN_PER_JOB))
+                teamId,
+                Sets.newHashSet(ExecutionModeEnum.YARN_APPLICATION, ExecutionModeEnum.YARN_PER_JOB))
             .stream()
             .filter(
                 application -> {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppHttpWatcher.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppHttpWatcher.java
index ce034f404..ff334915d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppHttpWatcher.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppHttpWatcher.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.task;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.HttpClientUtils;
 import org.apache.streampark.common.util.ThreadUtils;
 import org.apache.streampark.common.util.YarnUtils;
@@ -25,10 +25,10 @@ import org.apache.streampark.console.base.util.JacksonUtils;
 import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkCluster;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.OptionState;
-import org.apache.streampark.console.core.enums.ReleaseState;
-import org.apache.streampark.console.core.enums.StopFrom;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
+import org.apache.streampark.console.core.enums.StopFromEnum;
 import org.apache.streampark.console.core.metrics.flink.CheckPoints;
 import org.apache.streampark.console.core.metrics.flink.JobsOverview;
 import org.apache.streampark.console.core.metrics.flink.Overview;
@@ -132,7 +132,7 @@ public class FlinkAppHttpWatcher {
    *    and needs to be manually specified when started again.
    * </pre>
    */
-  private static final Map<Long, StopFrom> STOP_FROM_MAP = new ConcurrentHashMap<>(0);
+  private static final Map<Long, StopFromEnum> STOP_FROM_MAP = new ConcurrentHashMap<>(0);
 
   /**
    * Cancelling tasks are placed in this cache with an expiration time of 10 seconds (the time of 2
@@ -148,7 +148,7 @@ public class FlinkAppHttpWatcher {
 
   private static final Map<Long, FlinkCluster> FLINK_CLUSTER_MAP = new ConcurrentHashMap<>(0);
 
-  private static final Map<Long, OptionState> OPTIONING = new ConcurrentHashMap<>(0);
+  private static final Map<Long, OptionStateEnum> OPTIONING = new ConcurrentHashMap<>(0);
 
   private Long lastWatchTime = 0L;
 
@@ -172,8 +172,8 @@ public class FlinkAppHttpWatcher {
         applicationManageService.list(
             new LambdaQueryWrapper<Application>()
                 .eq(Application::getTracking, 1)
-                .ne(Application::getState, FlinkAppState.LOST.getValue())
-                .notIn(Application::getExecutionMode, ExecutionMode.getKubernetesMode()));
+                .ne(Application::getState, FlinkAppStateEnum.LOST.getValue())
+                .notIn(Application::getExecutionMode, ExecutionModeEnum.getKubernetesMode()));
     applications.forEach(
         (app) -> {
           WATCHING_APPS.put(app.getId(), app);
@@ -209,7 +209,7 @@ public class FlinkAppHttpWatcher {
   }
 
   @VisibleForTesting
-  public @Nullable FlinkAppState tryQueryFlinkAppState(@Nonnull Long appId) {
+  public @Nullable FlinkAppStateEnum tryQueryFlinkAppState(@Nonnull Long appId) {
     Application app = WATCHING_APPS.get(appId);
     return (app == null || app.getState() == null) ? null : app.getStateEnum();
   }
@@ -231,8 +231,8 @@ public class FlinkAppHttpWatcher {
         });
   }
 
-  private StopFrom getAppStopFrom(Long appId) {
-    return STOP_FROM_MAP.getOrDefault(appId, StopFrom.NONE);
+  private StopFromEnum getAppStopFrom(Long appId) {
+    return STOP_FROM_MAP.getOrDefault(appId, StopFromEnum.NONE);
   }
 
   /**
@@ -244,8 +244,9 @@ public class FlinkAppHttpWatcher {
   private void getStateFromFlink(Application application) throws Exception {
     JobsOverview jobsOverview = httpJobsOverview(application);
     Optional<JobsOverview.Job> optional;
-    ExecutionMode execMode = application.getExecutionModeEnum();
-    if (ExecutionMode.YARN_APPLICATION == execMode || ExecutionMode.YARN_PER_JOB == execMode) {
+    ExecutionModeEnum execMode = application.getExecutionModeEnum();
+    if (ExecutionModeEnum.YARN_APPLICATION == execMode
+        || ExecutionModeEnum.YARN_PER_JOB == execMode) {
       optional =
           !jobsOverview.getJobs().isEmpty()
               ? jobsOverview.getJobs().stream()
@@ -261,9 +262,9 @@ public class FlinkAppHttpWatcher {
     if (optional.isPresent()) {
 
       JobsOverview.Job jobOverview = optional.get();
-      FlinkAppState currentState = FlinkAppState.of(jobOverview.getState());
+      FlinkAppStateEnum currentState = FlinkAppStateEnum.of(jobOverview.getState());
 
-      if (FlinkAppState.OTHER != currentState) {
+      if (FlinkAppStateEnum.OTHER != currentState) {
         try {
           // 1) set info from JobOverview
           handleJobOverview(application, jobOverview);
@@ -277,11 +278,11 @@ public class FlinkAppHttpWatcher {
           log.error("get flink jobOverview error: {}", e.getMessage(), e);
         }
         // 3) savePoint obsolete check and NEED_START check
-        OptionState optionState = OPTIONING.get(application.getId());
-        if (FlinkAppState.RUNNING == currentState) {
-          handleRunningState(application, optionState, currentState);
+        OptionStateEnum optionStateEnum = OPTIONING.get(application.getId());
+        if (FlinkAppStateEnum.RUNNING == currentState) {
+          handleRunningState(application, optionStateEnum, currentState);
         } else {
-          handleNotRunState(application, optionState, currentState);
+          handleNotRunState(application, optionStateEnum, currentState);
         }
       }
     }
@@ -294,68 +295,68 @@ public class FlinkAppHttpWatcher {
    * @param application application
    */
   private void getStateFromYarn(Application application) throws Exception {
-    OptionState optionState = OPTIONING.get(application.getId());
+    OptionStateEnum optionStateEnum = OPTIONING.get(application.getId());
     /*
      If the status of the last time is CANCELING (flink rest server is not closed at the time of getting information)
      and the status is not obtained this time (flink rest server is closed),
      the task is considered CANCELED
     */
     Byte flag = CANCELING_CACHE.getIfPresent(application.getId());
-    StopFrom stopFrom = getAppStopFrom(application.getId());
+    StopFromEnum stopFromEnum = getAppStopFrom(application.getId());
 
     if (flag != null) {
       log.info("FlinkAppHttpWatcher previous state: canceling.");
-      if (stopFrom.isNone()) {
+      if (stopFromEnum.isNone()) {
         log.error(
             "FlinkAppHttpWatcher query previous state was canceling and stopFrom NotFound,savePoint expired!");
         savePointService.expire(application.getId());
       }
-      application.setState(FlinkAppState.CANCELED.getValue());
+      application.setState(FlinkAppStateEnum.CANCELED.getValue());
       cleanSavepoint(application);
-      cleanOptioning(optionState, application.getId());
+      cleanOptioning(optionStateEnum, application.getId());
       doPersistMetrics(application, true);
     } else {
       // query the status from the yarn rest Api
       YarnAppInfo yarnAppInfo = httpYarnAppInfo(application);
       if (yarnAppInfo == null) {
-        if (ExecutionMode.REMOTE != application.getExecutionModeEnum()) {
+        if (ExecutionModeEnum.REMOTE != application.getExecutionModeEnum()) {
           throw new RuntimeException("FlinkAppHttpWatcher getStateFromYarn failed ");
         }
       } else {
         try {
           String state = yarnAppInfo.getApp().getFinalStatus();
-          FlinkAppState flinkAppState = FlinkAppState.of(state);
-          if (FlinkAppState.OTHER == flinkAppState) {
+          FlinkAppStateEnum flinkAppStateEnum = FlinkAppStateEnum.of(state);
+          if (FlinkAppStateEnum.OTHER == flinkAppStateEnum) {
             return;
           }
-          if (FlinkAppState.KILLED == flinkAppState) {
-            if (stopFrom.isNone()) {
+          if (FlinkAppStateEnum.KILLED == flinkAppStateEnum) {
+            if (stopFromEnum.isNone()) {
               log.error(
                   "FlinkAppHttpWatcher getStateFromYarn,job was killed and stopFrom NotFound,savePoint expired!");
               savePointService.expire(application.getId());
             }
-            flinkAppState = FlinkAppState.CANCELED;
+            flinkAppStateEnum = FlinkAppStateEnum.CANCELED;
             cleanSavepoint(application);
             application.setEndTime(new Date());
           }
-          if (FlinkAppState.SUCCEEDED == flinkAppState) {
-            flinkAppState = FlinkAppState.FINISHED;
+          if (FlinkAppStateEnum.SUCCEEDED == flinkAppStateEnum) {
+            flinkAppStateEnum = FlinkAppStateEnum.FINISHED;
           }
-          application.setState(flinkAppState.getValue());
-          cleanOptioning(optionState, application.getId());
+          application.setState(flinkAppStateEnum.getValue());
+          cleanOptioning(optionStateEnum, application.getId());
           doPersistMetrics(application, true);
-          if (FlinkAppState.FAILED == flinkAppState
-              || FlinkAppState.LOST == flinkAppState
-              || (FlinkAppState.CANCELED == flinkAppState && stopFrom.isNone())
+          if (FlinkAppStateEnum.FAILED == flinkAppStateEnum
+              || FlinkAppStateEnum.LOST == flinkAppStateEnum
+              || (FlinkAppStateEnum.CANCELED == flinkAppStateEnum && stopFromEnum.isNone())
               || applicationInfoService.checkAlter(application)) {
-            doAlert(application, flinkAppState);
+            doAlert(application, flinkAppStateEnum);
             stopCanceledJob(application.getId());
-            if (FlinkAppState.FAILED == flinkAppState) {
+            if (FlinkAppStateEnum.FAILED == flinkAppStateEnum) {
               applicationActionService.start(application, true);
             }
           }
         } catch (Exception e) {
-          if (ExecutionMode.REMOTE != application.getExecutionModeEnum()) {
+          if (ExecutionModeEnum.REMOTE != application.getExecutionModeEnum()) {
             throw new RuntimeException("FlinkAppHttpWatcher getStateFromYarn error,", e);
           }
         }
@@ -368,19 +369,19 @@ public class FlinkAppHttpWatcher {
      Query from flink's restAPI and yarn's restAPI both failed.
      In this case, it is necessary to decide whether to return to the final state depending on the state being operated
     */
-    final OptionState optionState = OPTIONING.get(application.getId());
-    if (OptionState.STARTING != optionState) {
+    final OptionStateEnum optionStateEnum = OPTIONING.get(application.getId());
+    if (OptionStateEnum.STARTING != optionStateEnum) {
       // non-mapping
-      if (application.getStateEnum() != FlinkAppState.MAPPING) {
+      if (application.getStateEnum() != FlinkAppStateEnum.MAPPING) {
         log.error(
             "FlinkAppHttpWatcher getStateFromFlink and getStateFromYARN error,job failed, savePoint expired!");
-        StopFrom stopFrom = getAppStopFrom(application.getId());
-        if (stopFrom.isNone()) {
+        StopFromEnum stopFromEnum = getAppStopFrom(application.getId());
+        if (stopFromEnum.isNone()) {
           savePointService.expire(application.getId());
-          application.setState(FlinkAppState.LOST.getValue());
-          doAlert(application, FlinkAppState.LOST);
+          application.setState(FlinkAppStateEnum.LOST.getValue());
+          doAlert(application, FlinkAppStateEnum.LOST);
         } else {
-          application.setState(FlinkAppState.CANCELED.getValue());
+          application.setState(FlinkAppStateEnum.CANCELED.getValue());
         }
       }
       /*
@@ -390,12 +391,12 @@ public class FlinkAppHttpWatcher {
       */
       application.setEndTime(new Date());
       cleanSavepoint(application);
-      cleanOptioning(optionState, application.getId());
+      cleanOptioning(optionStateEnum, application.getId());
       doPersistMetrics(application, true);
-      FlinkAppState appState = application.getStateEnum();
-      if (FlinkAppState.FAILED == appState || FlinkAppState.LOST == appState) {
+      FlinkAppStateEnum appState = application.getStateEnum();
+      if (FlinkAppStateEnum.FAILED == appState || FlinkAppStateEnum.LOST == appState) {
         doAlert(application, application.getStateEnum());
-        if (FlinkAppState.FAILED == appState) {
+        if (FlinkAppStateEnum.FAILED == appState) {
           try {
             applicationActionService.start(application, true);
           } catch (Exception e) {
@@ -455,11 +456,11 @@ public class FlinkAppHttpWatcher {
    * Handle running task
    *
    * @param application application
-   * @param optionState optionState
+   * @param optionStateEnum optionState
    * @param currentState currentState
    */
   private void handleRunningState(
-      Application application, OptionState optionState, FlinkAppState currentState) {
+      Application application, OptionStateEnum optionStateEnum, FlinkAppStateEnum currentState) {
     /*
      if the last recorded state is STARTING and the latest state obtained this time is RUNNING,
      which means it is the first tracking after restart.
@@ -469,16 +470,16 @@ public class FlinkAppHttpWatcher {
      NEED_RESTART_AFTER_ROLLBACK (Need to restart after rollback)
      NEED_RESTART_AFTER_DEPLOY (Need to rollback after deploy)
     */
-    if (OptionState.STARTING == optionState) {
+    if (OptionStateEnum.STARTING == optionStateEnum) {
       Application latestApp = WATCHING_APPS.get(application.getId());
-      ReleaseState releaseState = latestApp.getReleaseState();
-      switch (releaseState) {
+      ReleaseStateEnum releaseStateEnum = latestApp.getReleaseState();
+      switch (releaseStateEnum) {
         case NEED_RESTART:
         case NEED_ROLLBACK:
           LambdaUpdateWrapper<Application> updateWrapper =
               new LambdaUpdateWrapper<Application>()
                   .eq(Application::getId, application.getId())
-                  .set(Application::getRelease, ReleaseState.DONE.get());
+                  .set(Application::getRelease, ReleaseStateEnum.DONE.get());
           applicationManageService.update(updateWrapper);
           break;
         default:
@@ -489,17 +490,17 @@ public class FlinkAppHttpWatcher {
     // The current state is running, and there is a current task in the savePointCache,
     // indicating that the task is doing savepoint
     if (SAVEPOINT_CACHE.getIfPresent(application.getId()) != null) {
-      application.setOptionState(OptionState.SAVEPOINTING.getValue());
+      application.setOptionState(OptionStateEnum.SAVEPOINTING.getValue());
     } else {
-      application.setOptionState(OptionState.NONE.getValue());
+      application.setOptionState(OptionStateEnum.NONE.getValue());
     }
     application.setState(currentState.getValue());
     doPersistMetrics(application, false);
-    cleanOptioning(optionState, application.getId());
+    cleanOptioning(optionStateEnum, application.getId());
   }
 
   private void doPersistMetrics(Application application, boolean stopWatch) {
-    if (FlinkAppState.isEndState(application.getState())) {
+    if (FlinkAppStateEnum.isEndState(application.getState())) {
       application.setOverview(null);
       application.setTotalTM(null);
       application.setTotalSlot(null);
@@ -520,11 +521,11 @@ public class FlinkAppHttpWatcher {
    * Handle not running task
    *
    * @param application application
-   * @param optionState optionState
+   * @param optionStateEnum optionState
    * @param currentState currentState
    */
   private void handleNotRunState(
-      Application application, OptionState optionState, FlinkAppState currentState)
+      Application application, OptionStateEnum optionStateEnum, FlinkAppStateEnum currentState)
       throws Exception {
 
     switch (currentState) {
@@ -540,26 +541,26 @@ public class FlinkAppHttpWatcher {
             currentState.name());
         cleanSavepoint(application);
         application.setState(currentState.getValue());
-        StopFrom stopFrom = getAppStopFrom(application.getId());
-        if (stopFrom.isNone() || applicationInfoService.checkAlter(application)) {
-          if (stopFrom.isNone()) {
+        StopFromEnum stopFromEnum = getAppStopFrom(application.getId());
+        if (stopFromEnum.isNone() || applicationInfoService.checkAlter(application)) {
+          if (stopFromEnum.isNone()) {
             log.info(
                 "FlinkAppHttpWatcher getFromFlinkRestApi, job cancel is not form StreamPark,savePoint expired!");
             savePointService.expire(application.getId());
           }
           stopCanceledJob(application.getId());
-          doAlert(application, FlinkAppState.CANCELED);
+          doAlert(application, FlinkAppStateEnum.CANCELED);
         }
         STOP_FROM_MAP.remove(application.getId());
         doPersistMetrics(application, true);
-        cleanOptioning(optionState, application.getId());
+        cleanOptioning(optionStateEnum, application.getId());
         break;
       case FAILED:
         cleanSavepoint(application);
         STOP_FROM_MAP.remove(application.getId());
-        application.setState(FlinkAppState.FAILED.getValue());
+        application.setState(FlinkAppStateEnum.FAILED.getValue());
         doPersistMetrics(application, true);
-        doAlert(application, FlinkAppState.FAILED);
+        doAlert(application, FlinkAppStateEnum.FAILED);
         applicationActionService.start(application, true);
         break;
       case RESTARTING:
@@ -574,8 +575,8 @@ public class FlinkAppHttpWatcher {
     }
   }
 
-  private void cleanOptioning(OptionState optionState, Long key) {
-    if (optionState != null) {
+  private void cleanOptioning(OptionStateEnum optionStateEnum, Long key) {
+    if (optionStateEnum != null) {
       lastOptionTime = System.currentTimeMillis();
       OPTIONING.remove(key);
     }
@@ -583,18 +584,18 @@ public class FlinkAppHttpWatcher {
 
   public void cleanSavepoint(Application application) {
     SAVEPOINT_CACHE.invalidate(application.getId());
-    application.setOptionState(OptionState.NONE.getValue());
+    application.setOptionState(OptionStateEnum.NONE.getValue());
   }
 
   /** set current option state */
-  public static void setOptionState(Long appId, OptionState state) {
+  public static void setOptionState(Long appId, OptionStateEnum state) {
     if (isKubernetesApp(appId)) {
       return;
     }
     log.info("FlinkAppHttpWatcher setOptioning");
     OPTIONING.put(appId, state);
-    if (OptionState.CANCELLING == state) {
-      STOP_FROM_MAP.put(appId, StopFrom.STREAMPARK);
+    if (OptionStateEnum.CANCELLING == state) {
+      STOP_FROM_MAP.put(appId, StopFromEnum.STREAMPARK);
     }
   }
 
@@ -661,8 +662,8 @@ public class FlinkAppHttpWatcher {
   private Overview httpOverview(Application application) throws IOException {
     String appId = application.getAppId();
     if (appId != null) {
-      if (ExecutionMode.YARN_APPLICATION == application.getExecutionModeEnum()
-          || ExecutionMode.YARN_PER_JOB == application.getExecutionModeEnum()) {
+      if (ExecutionModeEnum.YARN_APPLICATION == application.getExecutionModeEnum()
+          || ExecutionModeEnum.YARN_PER_JOB == application.getExecutionModeEnum()) {
         String reqURL;
         if (StringUtils.isBlank(application.getJobManagerUrl())) {
           String format = "proxy/%s/overview";
@@ -679,8 +680,8 @@ public class FlinkAppHttpWatcher {
 
   private JobsOverview httpJobsOverview(Application application) throws Exception {
     final String flinkUrl = "jobs/overview";
-    ExecutionMode execMode = application.getExecutionModeEnum();
-    if (ExecutionMode.isYarnMode(execMode)) {
+    ExecutionModeEnum execMode = application.getExecutionModeEnum();
+    if (ExecutionModeEnum.isYarnMode(execMode)) {
       String reqURL;
       if (StringUtils.isBlank(application.getJobManagerUrl())) {
         String format = "proxy/%s/" + flinkUrl;
@@ -692,7 +693,7 @@ public class FlinkAppHttpWatcher {
       return yarnRestRequest(reqURL, JobsOverview.class);
     }
 
-    if (application.getJobId() != null && ExecutionMode.isRemoteMode(execMode)) {
+    if (application.getJobId() != null && ExecutionModeEnum.isRemoteMode(execMode)) {
       return httpRemoteCluster(
           application.getFlinkClusterId(),
           cluster -> {
@@ -713,8 +714,8 @@ public class FlinkAppHttpWatcher {
 
   private CheckPoints httpCheckpoints(Application application) throws Exception {
     final String flinkUrl = "jobs/%s/checkpoints";
-    ExecutionMode execMode = application.getExecutionModeEnum();
-    if (ExecutionMode.isYarnMode(execMode)) {
+    ExecutionModeEnum execMode = application.getExecutionModeEnum();
+    if (ExecutionModeEnum.isYarnMode(execMode)) {
       String reqURL;
       if (StringUtils.isBlank(application.getJobManagerUrl())) {
         String format = "proxy/%s/" + flinkUrl;
@@ -726,7 +727,7 @@ public class FlinkAppHttpWatcher {
       return yarnRestRequest(reqURL, CheckPoints.class);
     }
 
-    if (application.getJobId() != null && ExecutionMode.isRemoteMode(execMode)) {
+    if (application.getJobId() != null && ExecutionModeEnum.isRemoteMode(execMode)) {
       return httpRemoteCluster(
           application.getFlinkClusterId(),
           cluster -> {
@@ -805,7 +806,7 @@ public class FlinkAppHttpWatcher {
    * @param app application
    * @param appState application state
    */
-  private void doAlert(Application app, FlinkAppState appState) {
+  private void doAlert(Application app, FlinkAppStateEnum appState) {
     if (app.getProbing()) {
       log.info("application with id {} is probing, don't send alert", app.getId());
       return;
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppLostWatcher.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppLostWatcher.java
index ebb3ab76c..b0c0e29bb 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppLostWatcher.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkAppLostWatcher.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.core.task;
 import org.apache.streampark.console.core.bean.AlertProbeMsg;
 import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 import org.apache.streampark.console.core.service.alert.AlertService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.flink.kubernetes.FlinkK8sWatcher;
@@ -39,7 +39,7 @@ import java.util.Set;
 import java.util.concurrent.atomic.AtomicBoolean;
 import java.util.stream.Collectors;
 
-import static org.apache.streampark.console.core.enums.FlinkAppState.LOST;
+import static org.apache.streampark.console.core.enums.FlinkAppStateEnum.LOST;
 import static org.apache.streampark.console.core.task.FlinkK8sWatcherWrapper.Bridge.toTrackId;
 import static org.apache.streampark.console.core.task.FlinkK8sWatcherWrapper.isKubernetesApp;
 
@@ -95,7 +95,7 @@ public class FlinkAppLostWatcher {
     isProbing.set(true);
     probeApplication =
         probeApplication.stream()
-            .filter(application -> FlinkAppState.isLost(application.getState()))
+            .filter(application -> FlinkAppStateEnum.isLost(application.getState()))
             .collect(Collectors.toList());
     updateState(probeApplication);
     probeApplication.stream().forEach(this::monitorApplication);
@@ -103,10 +103,10 @@ public class FlinkAppLostWatcher {
 
   private void updateState(List<Application> applications) {
     applications.stream()
-        .filter(application -> FlinkAppState.isLost(application.getState()))
+        .filter(application -> FlinkAppStateEnum.isLost(application.getState()))
         .forEach(
             application -> {
-              application.setState(FlinkAppState.PROBING.getValue());
+              application.setState(FlinkAppStateEnum.PROBING.getValue());
               application.setProbing(true);
             });
     applicationManageService.updateBatchById(applications);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkCheckpointProcessor.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkCheckpointProcessor.java
index 440170c42..4d134b27a 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkCheckpointProcessor.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkCheckpointProcessor.java
@@ -20,8 +20,8 @@ package org.apache.streampark.console.core.task;
 import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.SavePoint;
-import org.apache.streampark.console.core.enums.CheckPointStatus;
-import org.apache.streampark.console.core.enums.FailoverStrategy;
+import org.apache.streampark.console.core.enums.CheckPointStatusEnum;
+import org.apache.streampark.console.core.enums.FailoverStrategyEnum;
 import org.apache.streampark.console.core.metrics.flink.CheckPoints;
 import org.apache.streampark.console.core.service.SavePointService;
 import org.apache.streampark.console.core.service.alert.AlertService;
@@ -79,10 +79,10 @@ public class FlinkCheckpointProcessor {
   private void process(Application application, @Nonnull CheckPoints.CheckPoint checkPoint) {
     String jobID = application.getJobId();
     Long appId = application.getId();
-    CheckPointStatus status = checkPoint.getCheckPointStatus();
+    CheckPointStatusEnum status = checkPoint.getCheckPointStatus();
     CheckPointKey checkPointKey = new CheckPointKey(appId, jobID, checkPoint.getId());
 
-    if (CheckPointStatus.COMPLETED == status) {
+    if (CheckPointStatusEnum.COMPLETED == status) {
       if (shouldStoreAsSavepoint(checkPointKey, checkPoint)) {
         savepointedCache.put(checkPointKey.getSavePointId(), DEFAULT_FLAG_BYTE);
         saveSavepoint(checkPoint, application.getId());
@@ -104,15 +104,17 @@ public class FlinkCheckpointProcessor {
         if (minute <= application.getCpFailureRateInterval()
             && counter.getCount() >= application.getCpMaxFailureInterval()) {
           checkPointFailedCache.remove(appId);
-          FailoverStrategy failoverStrategy = FailoverStrategy.of(application.getCpFailureAction());
-          if (failoverStrategy == null) {
+          FailoverStrategyEnum failoverStrategyEnum =
+              FailoverStrategyEnum.of(application.getCpFailureAction());
+          if (failoverStrategyEnum == null) {
             throw new IllegalArgumentException(
                 "Unexpected cpFailureAction: " + application.getCpFailureAction());
           }
-          switch (failoverStrategy) {
+          switch (failoverStrategyEnum) {
             case ALERT:
               alertService.alert(
-                  application.getAlertId(), AlertTemplate.of(application, CheckPointStatus.FAILED));
+                  application.getAlertId(),
+                  AlertTemplate.of(application, CheckPointStatusEnum.FAILED));
               break;
             case RESTART:
               try {
@@ -160,8 +162,8 @@ public class FlinkCheckpointProcessor {
   }
 
   private boolean shouldProcessFailedTrigger(
-      CheckPoints.CheckPoint checkPoint, boolean cpFailedTrigger, CheckPointStatus status) {
-    return CheckPointStatus.FAILED == status && !checkPoint.getIsSavepoint() && cpFailedTrigger;
+      CheckPoints.CheckPoint checkPoint, boolean cpFailedTrigger, CheckPointStatusEnum status) {
+    return CheckPointStatusEnum.FAILED == status && !checkPoint.getIsSavepoint() && cpFailedTrigger;
   }
 
   private void saveSavepoint(CheckPoints.CheckPoint checkPoint, Long appId) {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkClusterWatcher.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkClusterWatcher.java
index f94df947b..0d697319e 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkClusterWatcher.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkClusterWatcher.java
@@ -19,8 +19,8 @@ package org.apache.streampark.console.core.task;
 
 import org.apache.streampark.common.conf.CommonConfig;
 import org.apache.streampark.common.conf.InternalConfigHolder;
-import org.apache.streampark.common.enums.ClusterState;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ClusterStateEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.HadoopUtils;
 import org.apache.streampark.common.util.HttpClientUtils;
 import org.apache.streampark.common.util.ThreadUtils;
@@ -77,7 +77,7 @@ public class FlinkClusterWatcher {
   /** Watcher cluster lists */
   private static final Map<Long, FlinkCluster> WATCHER_CLUSTERS = new ConcurrentHashMap<>(8);
 
-  private static final Cache<Long, ClusterState> FAILED_STATES =
+  private static final Cache<Long, ClusterStateEnum> FAILED_STATES =
       Caffeine.newBuilder().expireAfterWrite(WATCHER_INTERVAL).build();
 
   private boolean immediateWatch = false;
@@ -99,9 +99,9 @@ public class FlinkClusterWatcher {
     List<FlinkCluster> flinkClusters =
         flinkClusterService.list(
             new LambdaQueryWrapper<FlinkCluster>()
-                .eq(FlinkCluster::getClusterState, ClusterState.RUNNING.getState())
+                .eq(FlinkCluster::getClusterState, ClusterStateEnum.RUNNING.getState())
                 // excluding flink clusters on kubernetes
-                .notIn(FlinkCluster::getExecutionMode, ExecutionMode.getKubernetesMode()));
+                .notIn(FlinkCluster::getExecutionMode, ExecutionModeEnum.getKubernetesMode()));
     flinkClusters.forEach(cluster -> WATCHER_CLUSTERS.put(cluster.getId(), cluster));
   }
 
@@ -115,7 +115,7 @@ public class FlinkClusterWatcher {
           (aLong, flinkCluster) ->
               EXECUTOR.execute(
                   () -> {
-                    ClusterState state = getClusterState(flinkCluster);
+                    ClusterStateEnum state = getClusterState(flinkCluster);
                     switch (state) {
                       case FAILED:
                       case LOST:
@@ -132,7 +132,7 @@ public class FlinkClusterWatcher {
     }
   }
 
-  private void alert(FlinkCluster cluster, ClusterState state) {
+  private void alert(FlinkCluster cluster, ClusterStateEnum state) {
     if (cluster.getAlertId() != null) {
       cluster.setAllJobs(applicationInfoService.countByClusterId(cluster.getId()));
       cluster.setAffectedJobs(
@@ -150,13 +150,13 @@ public class FlinkClusterWatcher {
    * @param flinkCluster The FlinkCluster object representing the cluster.
    * @return The ClusterState object representing the state of the cluster.
    */
-  public ClusterState getClusterState(FlinkCluster flinkCluster) {
-    ClusterState state = FAILED_STATES.getIfPresent(flinkCluster.getId());
+  public ClusterStateEnum getClusterState(FlinkCluster flinkCluster) {
+    ClusterStateEnum state = FAILED_STATES.getIfPresent(flinkCluster.getId());
     if (state != null) {
       return state;
     }
     state = httpClusterState(flinkCluster);
-    if (ClusterState.isRunning(state)) {
+    if (ClusterStateEnum.isRunning(state)) {
       FAILED_STATES.invalidate(flinkCluster.getId());
     } else {
       immediateWatch = true;
@@ -171,7 +171,7 @@ public class FlinkClusterWatcher {
    * @param flinkCluster The FlinkCluster object representing the cluster.
    * @return The ClusterState object representing the state of the cluster.
    */
-  private ClusterState httpRemoteClusterState(FlinkCluster flinkCluster) {
+  private ClusterStateEnum httpRemoteClusterState(FlinkCluster flinkCluster) {
     return getStateFromFlinkRestApi(flinkCluster);
   }
 
@@ -181,9 +181,9 @@ public class FlinkClusterWatcher {
    * @param flinkCluster
    * @return
    */
-  private ClusterState httpYarnSessionClusterState(FlinkCluster flinkCluster) {
-    ClusterState state = getStateFromFlinkRestApi(flinkCluster);
-    if (ClusterState.LOST == state) {
+  private ClusterStateEnum httpYarnSessionClusterState(FlinkCluster flinkCluster) {
+    ClusterStateEnum state = getStateFromFlinkRestApi(flinkCluster);
+    if (ClusterStateEnum.LOST == state) {
       return getStateFromYarnRestApi(flinkCluster);
     }
     return state;
@@ -195,14 +195,14 @@ public class FlinkClusterWatcher {
    * @param flinkCluster
    * @return
    */
-  private ClusterState httpClusterState(FlinkCluster flinkCluster) {
+  private ClusterStateEnum httpClusterState(FlinkCluster flinkCluster) {
     switch (flinkCluster.getExecutionModeEnum()) {
       case REMOTE:
         return httpRemoteClusterState(flinkCluster);
       case YARN_SESSION:
         return httpYarnSessionClusterState(flinkCluster);
       default:
-        return ClusterState.UNKNOWN;
+        return ClusterStateEnum.UNKNOWN;
     }
   }
 
@@ -212,7 +212,7 @@ public class FlinkClusterWatcher {
    * @param flinkCluster
    * @return
    */
-  private ClusterState getStateFromFlinkRestApi(FlinkCluster flinkCluster) {
+  private ClusterStateEnum getStateFromFlinkRestApi(FlinkCluster flinkCluster) {
     String address = flinkCluster.getAddress();
     String jobManagerUrl = flinkCluster.getJobManagerUrl();
     String flinkUrl =
@@ -225,11 +225,11 @@ public class FlinkClusterWatcher {
               flinkUrl,
               RequestConfig.custom().setConnectTimeout(5000, TimeUnit.MILLISECONDS).build());
       JacksonUtils.read(res, Overview.class);
-      return ClusterState.RUNNING;
+      return ClusterStateEnum.RUNNING;
     } catch (Exception ignored) {
       log.error("cluster id:{} get state from flink api failed", flinkCluster.getId());
     }
-    return ClusterState.LOST;
+    return ClusterStateEnum.LOST;
   }
 
   /**
@@ -238,12 +238,12 @@ public class FlinkClusterWatcher {
    * @param flinkCluster
    * @return
    */
-  private ClusterState getStateFromYarnRestApi(FlinkCluster flinkCluster) {
+  private ClusterStateEnum getStateFromYarnRestApi(FlinkCluster flinkCluster) {
     String yarnUrl = "ws/v1/cluster/apps/".concat(flinkCluster.getClusterId());
     try {
       String result = YarnUtils.restRequest(yarnUrl);
       if (null == result) {
-        return ClusterState.UNKNOWN;
+        return ClusterStateEnum.UNKNOWN;
       }
       YarnAppInfo yarnAppInfo = JacksonUtils.read(result, YarnAppInfo.class);
       YarnApplicationState status = HadoopUtils.toYarnState(yarnAppInfo.getApp().getState());
@@ -251,11 +251,11 @@ public class FlinkClusterWatcher {
         log.error(
             "cluster id:{} final application status convert failed, invalid string ",
             flinkCluster.getId());
-        return ClusterState.UNKNOWN;
+        return ClusterStateEnum.UNKNOWN;
       }
       return yarnStateConvertClusterState(status);
     } catch (Exception e) {
-      return ClusterState.LOST;
+      return ClusterStateEnum.LOST;
     }
   }
 
@@ -265,7 +265,7 @@ public class FlinkClusterWatcher {
    * @param flinkCluster
    */
   public static void addWatching(FlinkCluster flinkCluster) {
-    if (!ExecutionMode.isKubernetesMode(flinkCluster.getExecutionModeEnum())
+    if (!ExecutionModeEnum.isKubernetesMode(flinkCluster.getExecutionModeEnum())
         && !WATCHER_CLUSTERS.containsKey(flinkCluster.getId())) {
       log.info("add the cluster with id:{} to watcher cluster cache", flinkCluster.getId());
       WATCHER_CLUSTERS.put(flinkCluster.getId(), flinkCluster);
@@ -286,10 +286,10 @@ public class FlinkClusterWatcher {
    * @param state
    * @return
    */
-  private ClusterState yarnStateConvertClusterState(YarnApplicationState state) {
+  private ClusterStateEnum yarnStateConvertClusterState(YarnApplicationState state) {
     return state == YarnApplicationState.FINISHED
-        ? ClusterState.CANCELED
-        : ClusterState.of(state.toString());
+        ? ClusterStateEnum.CANCELED
+        : ClusterStateEnum.of(state.toString());
   }
 
   /**
@@ -298,7 +298,7 @@ public class FlinkClusterWatcher {
    * @return <code>false</code> if the connection of the cluster is invalid, <code>true</code> else.
    */
   public Boolean verifyClusterConnection(FlinkCluster flinkCluster) {
-    ClusterState clusterState = httpClusterState(flinkCluster);
-    return ClusterState.isRunning(clusterState);
+    ClusterStateEnum clusterStateEnum = httpClusterState(flinkCluster);
+    return ClusterStateEnum.isRunning(clusterStateEnum);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sChangeEventListener.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sChangeEventListener.java
index c097dfb5b..c24bcd1d7 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sChangeEventListener.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sChangeEventListener.java
@@ -17,18 +17,18 @@
 
 package org.apache.streampark.console.core.task;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.ThreadUtils;
 import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.OptionState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.OptionStateEnum;
 import org.apache.streampark.console.core.metrics.flink.CheckPoints;
 import org.apache.streampark.console.core.service.alert.AlertService;
 import org.apache.streampark.console.core.service.application.ApplicationInfoService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState;
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode;
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum;
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum;
 import org.apache.streampark.flink.kubernetes.event.FlinkClusterMetricChangeEvent;
 import org.apache.streampark.flink.kubernetes.event.FlinkJobCheckpointChangeEvent;
 import org.apache.streampark.flink.kubernetes.event.FlinkJobStatusChangeEvent;
@@ -52,8 +52,8 @@ import java.util.concurrent.TimeUnit;
 
 import scala.Enumeration;
 
-import static org.apache.streampark.console.core.enums.FlinkAppState.Bridge.fromK8sFlinkJobState;
-import static org.apache.streampark.console.core.enums.FlinkAppState.Bridge.toK8sFlinkJobState;
+import static org.apache.streampark.console.core.enums.FlinkAppStateEnum.Bridge.fromK8sFlinkJobState;
+import static org.apache.streampark.console.core.enums.FlinkAppStateEnum.Bridge.toK8sFlinkJobState;
 
 /**
  * Event Listener for K8sFlinkTrackMonitor。
@@ -104,11 +104,11 @@ public class FlinkK8sChangeEventListener {
     applicationInfoService.persistMetrics(app);
 
     // email alerts when necessary
-    FlinkAppState state = app.getStateEnum();
-    if (FlinkAppState.FAILED == state
-        || FlinkAppState.LOST == state
-        || FlinkAppState.RESTARTING == state
-        || FlinkAppState.FINISHED == state) {
+    FlinkAppStateEnum state = app.getStateEnum();
+    if (FlinkAppStateEnum.FAILED == state
+        || FlinkAppStateEnum.LOST == state
+        || FlinkAppStateEnum.RESTARTING == state
+        || FlinkAppStateEnum.FINISHED == state) {
       executor.execute(
           () -> {
             if (app.getProbing()) {
@@ -129,9 +129,9 @@ public class FlinkK8sChangeEventListener {
   @Subscribe
   public void subscribeMetricsChange(FlinkClusterMetricChangeEvent event) {
     TrackId trackId = event.trackId();
-    ExecutionMode mode = FlinkK8sExecuteMode.toExecutionMode(trackId.executeMode());
+    ExecutionModeEnum mode = FlinkK8sExecuteModeEnum.toExecutionMode(trackId.executeMode());
     // discard session mode change
-    if (ExecutionMode.KUBERNETES_NATIVE_SESSION == mode) {
+    if (ExecutionModeEnum.KUBERNETES_NATIVE_SESSION == mode) {
       return;
     }
 
@@ -184,7 +184,7 @@ public class FlinkK8sChangeEventListener {
     long endTime = Math.max(jobStatus.jobEndTime(), preEndTime);
     long duration = jobStatus.duration();
 
-    if (FlinkJobState.isEndState(state)) {
+    if (FlinkJobStateEnum.isEndState(state)) {
       if (endTime < startTime) {
         endTime = System.currentTimeMillis();
       }
@@ -202,6 +202,6 @@ public class FlinkK8sChangeEventListener {
     app.setDuration(duration > 0 ? duration : 0);
     // when a flink job status change event can be received, it means
     // that the operation command sent by streampark has been completed.
-    app.setOptionState(OptionState.NONE.getValue());
+    app.setOptionState(OptionStateEnum.NONE.getValue());
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sWatcherWrapper.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sWatcherWrapper.java
index 6c82fe142..ca6dd12a8 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sWatcherWrapper.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/FlinkK8sWatcherWrapper.java
@@ -18,15 +18,15 @@
 package org.apache.streampark.console.core.task;
 
 import org.apache.streampark.common.conf.K8sFlinkConfig;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.flink.kubernetes.FlinkK8sWatcher;
 import org.apache.streampark.flink.kubernetes.FlinkK8sWatcherFactory;
 import org.apache.streampark.flink.kubernetes.FlinkTrackConfig;
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState;
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode;
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum;
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum;
 import org.apache.streampark.flink.kubernetes.model.TrackId;
 
 import org.apache.commons.collections.CollectionUtils;
@@ -45,7 +45,7 @@ import java.util.stream.Collectors;
 
 import scala.Enumeration;
 
-import static org.apache.streampark.console.core.enums.FlinkAppState.Bridge.toK8sFlinkJobState;
+import static org.apache.streampark.console.core.enums.FlinkAppStateEnum.Bridge.toK8sFlinkJobState;
 
 /**
  * Flink K8s Tracking Monitor Wrapper.
@@ -99,8 +99,8 @@ public class FlinkK8sWatcherWrapper {
     final LambdaQueryWrapper<Application> queryWrapper = new LambdaQueryWrapper<>();
     queryWrapper
         .eq(Application::getTracking, 1)
-        .ne(Application::getState, FlinkAppState.LOST.getValue())
-        .in(Application::getExecutionMode, ExecutionMode.getKubernetesMode());
+        .ne(Application::getState, FlinkAppStateEnum.LOST.getValue())
+        .in(Application::getExecutionMode, ExecutionModeEnum.getKubernetesMode());
 
     List<Application> k8sApplication = applicationManageService.list(queryWrapper);
     if (CollectionUtils.isEmpty(k8sApplication)) {
@@ -116,7 +116,7 @@ public class FlinkK8sWatcherWrapper {
     }
     // filter out the application that should be tracking
     return k8sApplication.stream()
-        .filter(app -> !FlinkJobState.isEndState(toK8sFlinkJobState(app.getStateEnum())))
+        .filter(app -> !FlinkJobStateEnum.isEndState(toK8sFlinkJobState(app.getStateEnum())))
         .map(Bridge::toTrackId)
         .collect(Collectors.toList());
   }
@@ -126,15 +126,15 @@ public class FlinkK8sWatcherWrapper {
 
     // covert Application to TrackId
     public static TrackId toTrackId(@Nonnull Application app) {
-      Enumeration.Value mode = FlinkK8sExecuteMode.of(app.getExecutionModeEnum());
-      if (FlinkK8sExecuteMode.APPLICATION() == mode) {
+      Enumeration.Value mode = FlinkK8sExecuteModeEnum.of(app.getExecutionModeEnum());
+      if (FlinkK8sExecuteModeEnum.APPLICATION() == mode) {
         return TrackId.onApplication(
             app.getK8sNamespace(),
             app.getClusterId(),
             app.getId(),
             app.getJobId(),
             app.getTeamId().toString());
-      } else if (FlinkK8sExecuteMode.SESSION() == mode) {
+      } else if (FlinkK8sExecuteModeEnum.SESSION() == mode) {
         return TrackId.onSession(
             app.getK8sNamespace(),
             app.getClusterId(),
@@ -153,6 +153,6 @@ public class FlinkK8sWatcherWrapper {
     if (application == null) {
       return false;
     }
-    return ExecutionMode.isKubernetesMode(application.getExecutionMode());
+    return ExecutionModeEnum.isKubernetesMode(application.getExecutionMode());
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/ProjectBuildTask.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/ProjectBuildTask.java
index b423b0527..ecabb95ad 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/ProjectBuildTask.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/core/task/ProjectBuildTask.java
@@ -21,7 +21,7 @@ import org.apache.streampark.common.util.CommandUtils;
 import org.apache.streampark.common.util.Utils;
 import org.apache.streampark.console.base.util.GitUtils;
 import org.apache.streampark.console.core.entity.Project;
-import org.apache.streampark.console.core.enums.BuildState;
+import org.apache.streampark.console.core.enums.BuildStateEnum;
 
 import ch.qos.logback.classic.Logger;
 import lombok.extern.slf4j.Slf4j;
@@ -40,14 +40,14 @@ public class ProjectBuildTask extends AbstractLogFileTask {
 
   final Project project;
 
-  final Consumer<BuildState> stateUpdateConsumer;
+  final Consumer<BuildStateEnum> stateUpdateConsumer;
 
   final Consumer<Logger> notifyReleaseConsumer;
 
   public ProjectBuildTask(
       String logPath,
       Project project,
-      Consumer<BuildState> stateUpdateConsumer,
+      Consumer<BuildStateEnum> stateUpdateConsumer,
       Consumer<Logger> notifyReleaseConsumer) {
     super(logPath, true);
     this.project = project;
@@ -62,23 +62,23 @@ public class ProjectBuildTask extends AbstractLogFileTask {
     boolean cloneSuccess = cloneSourceCode(project);
     if (!cloneSuccess) {
       fileLogger.error("[StreamPark] clone or pull error.");
-      stateUpdateConsumer.accept(BuildState.FAILED);
+      stateUpdateConsumer.accept(BuildStateEnum.FAILED);
       return;
     }
     boolean build = projectBuild(project);
     if (!build) {
-      stateUpdateConsumer.accept(BuildState.FAILED);
+      stateUpdateConsumer.accept(BuildStateEnum.FAILED);
       fileLogger.error("build error, project name: {} ", project.getName());
       return;
     }
-    stateUpdateConsumer.accept(BuildState.SUCCESSFUL);
+    stateUpdateConsumer.accept(BuildStateEnum.SUCCESSFUL);
     this.deploy(project);
     notifyReleaseConsumer.accept(fileLogger);
   }
 
   @Override
   protected void processException(Throwable t) {
-    stateUpdateConsumer.accept(BuildState.FAILED);
+    stateUpdateConsumer.accept(BuildStateEnum.FAILED);
     fileLogger.error("Build error, project name: {}", project.getName(), t);
   }
 
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/AccessTokenController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/AccessTokenController.java
index 9a58bd8b6..7deea1c23 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/AccessTokenController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/AccessTokenController.java
@@ -21,7 +21,7 @@ import org.apache.streampark.common.util.CURLBuilder;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.base.exception.InternalException;
-import org.apache.streampark.console.core.enums.AccessTokenState;
+import org.apache.streampark.console.core.enums.AccessTokenStateEnum;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.system.entity.AccessToken;
 import org.apache.streampark.console.system.service.AccessTokenService;
@@ -92,14 +92,14 @@ public class AccessTokenController {
     if (userId != null) {
       AccessToken accessToken = accessTokenService.getByUserId(userId);
       if (accessToken == null) {
-        restResponse.data(AccessTokenState.NULL.get());
+        restResponse.data(AccessTokenStateEnum.NULL.get());
       } else if (AccessToken.STATUS_DISABLE.equals(accessToken.getFinalStatus())) {
-        restResponse.data(AccessTokenState.INVALID.get());
+        restResponse.data(AccessTokenStateEnum.INVALID.get());
       } else {
-        restResponse.data(AccessTokenState.OK.get());
+        restResponse.data(AccessTokenStateEnum.OK.get());
       }
     } else {
-      restResponse.data(AccessTokenState.INVALID.get());
+      restResponse.data(AccessTokenStateEnum.INVALID.get());
     }
     return restResponse;
   }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/MemberController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/MemberController.java
index bcbfd2046..809e1689d 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/MemberController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/MemberController.java
@@ -20,7 +20,7 @@ package org.apache.streampark.console.system.controller;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.core.annotation.PermissionAction;
-import org.apache.streampark.console.core.enums.PermissionType;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
 import org.apache.streampark.console.system.entity.Member;
 import org.apache.streampark.console.system.entity.Team;
 import org.apache.streampark.console.system.entity.User;
@@ -83,7 +83,7 @@ public class MemberController {
   }
 
   @Operation(summary = "Create member")
-  @PermissionAction(id = "#member.teamId", type = PermissionType.TEAM)
+  @PermissionAction(id = "#member.teamId", type = PermissionTypeEnum.TEAM)
   @PostMapping("post")
   @RequiresPermissions("member:add")
   public RestResponse create(@Valid Member member) {
@@ -92,7 +92,7 @@ public class MemberController {
   }
 
   @Operation(summary = "Delete member")
-  @PermissionAction(id = "#member.teamId", type = PermissionType.TEAM)
+  @PermissionAction(id = "#member.teamId", type = PermissionTypeEnum.TEAM)
   @DeleteMapping("delete")
   @RequiresPermissions("member:delete")
   public RestResponse delete(Member member) {
@@ -101,7 +101,7 @@ public class MemberController {
   }
 
   @Operation(summary = "Update member")
-  @PermissionAction(id = "#member.teamId", type = PermissionType.TEAM)
+  @PermissionAction(id = "#member.teamId", type = PermissionTypeEnum.TEAM)
   @PutMapping("update")
   @RequiresPermissions("member:update")
   public RestResponse update(Member member) {
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/SsoController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/SsoController.java
index 06956d107..b12791771 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/SsoController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/SsoController.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.system.controller;
 
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.base.exception.ApiAlertException;
-import org.apache.streampark.console.core.enums.LoginType;
+import org.apache.streampark.console.core.enums.LoginTypeEnum;
 import org.apache.streampark.console.system.entity.User;
 import org.apache.streampark.console.system.security.Authenticator;
 import org.apache.streampark.console.system.service.UserService;
@@ -83,7 +83,7 @@ public class SsoController {
       throw new ApiAlertException("Please configure the correct Principal Name Attribute");
     }
 
-    User user = authenticator.authenticate(principal.getName(), null, LoginType.SSO.toString());
+    User user = authenticator.authenticate(principal.getName(), null, LoginTypeEnum.SSO.toString());
     return userService.getLoginUserInfo(user);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/UserController.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/UserController.java
index b0e3aa26b..4f24d8fd8 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/UserController.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/controller/UserController.java
@@ -22,8 +22,8 @@ import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.core.annotation.PermissionAction;
-import org.apache.streampark.console.core.enums.LoginType;
-import org.apache.streampark.console.core.enums.PermissionType;
+import org.apache.streampark.console.core.enums.LoginTypeEnum;
+import org.apache.streampark.console.core.enums.PermissionTypeEnum;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.system.entity.Team;
 import org.apache.streampark.console.system.entity.User;
@@ -77,7 +77,7 @@ public class UserController {
   @PostMapping("post")
   @RequiresPermissions("user:add")
   public RestResponse addUser(@Valid User user) throws Exception {
-    user.setLoginType(LoginType.PASSWORD);
+    user.setLoginTypeEnum(LoginTypeEnum.PASSWORD);
     this.userService.createUser(user);
     return RestResponse.success();
   }
@@ -112,7 +112,7 @@ public class UserController {
   }
 
   @Operation(summary = "Update password")
-  @PermissionAction(id = "#user.userId", type = PermissionType.USER)
+  @PermissionAction(id = "#user.userId", type = PermissionTypeEnum.USER)
   @PutMapping("password")
   public RestResponse updatePassword(User user) throws Exception {
     userService.updatePassword(user);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/entity/User.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/entity/User.java
index 7a328a1f9..9259fc98f 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/entity/User.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/entity/User.java
@@ -18,8 +18,8 @@
 package org.apache.streampark.console.system.entity;
 
 import org.apache.streampark.common.conf.ConfigConst;
-import org.apache.streampark.console.core.enums.LoginType;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.LoginTypeEnum;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 
 import com.baomidou.mybatisplus.annotation.IdType;
 import com.baomidou.mybatisplus.annotation.TableId;
@@ -63,9 +63,9 @@ public class User implements Serializable {
   @Email(message = "{email}")
   private String email;
 
-  private UserType userType;
+  private UserTypeEnum userTypeEnum;
 
-  private LoginType loginType;
+  private LoginTypeEnum loginTypeEnum;
 
   @NotBlank(message = "{required}")
   private String status;
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/security/impl/AuthenticatorImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/security/impl/AuthenticatorImpl.java
index 2b81a5dd0..9ff76aa8b 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/security/impl/AuthenticatorImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/security/impl/AuthenticatorImpl.java
@@ -19,8 +19,8 @@ package org.apache.streampark.console.system.security.impl;
 
 import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.base.util.ShaHashUtils;
-import org.apache.streampark.console.core.enums.LoginType;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.LoginTypeEnum;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 import org.apache.streampark.console.system.entity.User;
 import org.apache.streampark.console.system.security.Authenticator;
 import org.apache.streampark.console.system.service.UserService;
@@ -39,7 +39,7 @@ public class AuthenticatorImpl implements Authenticator {
 
   @Override
   public User authenticate(String username, String password, String loginType) throws Exception {
-    LoginType loginTypeEnum = LoginType.of(loginType);
+    LoginTypeEnum loginTypeEnum = LoginTypeEnum.of(loginType);
     if (loginTypeEnum == null) {
       throw new ApiAlertException(
           String.format("the login type [%s] is not supported.", loginType));
@@ -62,7 +62,7 @@ public class AuthenticatorImpl implements Authenticator {
     if (user == null) {
       throw new ApiAlertException(String.format("user [%s] does not exist", username));
     }
-    if (user.getLoginType() != LoginType.PASSWORD) {
+    if (user.getLoginTypeEnum() != LoginTypeEnum.PASSWORD) {
       throw new ApiAlertException(String.format("user [%s] can not login with PASSWORD", username));
     }
     String salt = user.getSalt();
@@ -82,35 +82,35 @@ public class AuthenticatorImpl implements Authenticator {
     User user = usersService.findByName(username);
 
     if (user != null) {
-      if (user.getLoginType() != LoginType.LDAP) {
+      if (user.getLoginTypeEnum() != LoginTypeEnum.LDAP) {
         throw new ApiAlertException(
-            String.format("user [%s] can only sign in with %s", username, user.getLoginType()));
+            String.format("user [%s] can only sign in with %s", username, user.getLoginTypeEnum()));
       }
       return user;
     }
-    return this.newUserCreate(LoginType.LDAP, username);
+    return this.newUserCreate(LoginTypeEnum.LDAP, username);
   }
 
   private User ssoAuthenticate(String username) throws Exception {
     // check if user exist
     User user = usersService.findByName(username);
     if (user != null) {
-      if (user.getLoginType() != LoginType.SSO) {
+      if (user.getLoginTypeEnum() != LoginTypeEnum.SSO) {
         throw new ApiAlertException(
-            String.format("user [%s] can only sign in with %s", username, user.getLoginType()));
+            String.format("user [%s] can only sign in with %s", username, user.getLoginTypeEnum()));
       }
       return user;
     }
-    return this.newUserCreate(LoginType.SSO, username);
+    return this.newUserCreate(LoginTypeEnum.SSO, username);
   }
 
-  private User newUserCreate(LoginType loginType, String username) throws Exception {
+  private User newUserCreate(LoginTypeEnum loginTypeEnum, String username) throws Exception {
     User newUser = new User();
     newUser.setCreateTime(new Date());
     newUser.setUsername(username);
     newUser.setNickName(username);
-    newUser.setLoginType(loginType);
-    newUser.setUserType(UserType.USER);
+    newUser.setLoginTypeEnum(loginTypeEnum);
+    newUser.setUserTypeEnum(UserTypeEnum.USER);
     newUser.setStatus(User.STATUS_VALID);
     newUser.setSex(User.SEX_UNKNOWN);
     usersService.createUser(newUser);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/MenuServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/MenuServiceImpl.java
index 4ce0aaca9..15a38a745 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/MenuServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/MenuServiceImpl.java
@@ -21,7 +21,7 @@ import org.apache.streampark.console.base.domain.router.RouterMeta;
 import org.apache.streampark.console.base.domain.router.RouterTree;
 import org.apache.streampark.console.base.domain.router.VueRouter;
 import org.apache.streampark.console.base.util.VueRouterUtils;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 import org.apache.streampark.console.system.entity.Menu;
 import org.apache.streampark.console.system.entity.User;
 import org.apache.streampark.console.system.mapper.MenuMapper;
@@ -66,7 +66,7 @@ public class MenuServiceImpl extends ServiceImpl<MenuMapper, Menu> implements Me
                     new IllegalArgumentException(
                         String.format("The userId [%s] not found", userId)));
     // Admin has the permission for all menus.
-    if (UserType.ADMIN == user.getUserType()) {
+    if (UserTypeEnum.ADMIN == user.getUserTypeEnum()) {
       return this.list().stream().map(Menu::getPerms).collect(Collectors.toList());
     }
     return this.baseMapper.findUserPermissions(userId, teamId);
@@ -81,7 +81,7 @@ public class MenuServiceImpl extends ServiceImpl<MenuMapper, Menu> implements Me
                     new IllegalArgumentException(
                         String.format("The userId:[%s] not found", userId)));
     // Admin has the permission for all menus.
-    if (UserType.ADMIN == user.getUserType()) {
+    if (UserTypeEnum.ADMIN == user.getUserTypeEnum()) {
       LambdaQueryWrapper<Menu> queryWrapper =
           new LambdaQueryWrapper<Menu>().eq(Menu::getType, "0").orderByAsc(Menu::getOrderNum);
       return this.list(queryWrapper);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/TeamServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/TeamServiceImpl.java
index f89c3b398..676f1be80 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/TeamServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/TeamServiceImpl.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console.system.service.impl;
 
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.exception.ApiAlertException;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 import org.apache.streampark.console.core.service.CommonService;
 import org.apache.streampark.console.core.service.ProjectService;
 import org.apache.streampark.console.core.service.VariableService;
@@ -143,7 +143,7 @@ public class TeamServiceImpl extends ServiceImpl<TeamMapper, Team> implements Te
             .orElseThrow(
                 () -> new ApiAlertException(String.format("The userId [%s] not found.", userId)));
     // Admin has the permission for all teams.
-    if (UserType.ADMIN == user.getUserType()) {
+    if (UserTypeEnum.ADMIN == user.getUserTypeEnum()) {
       return this.list();
     }
     return baseMapper.findUserTeams(userId);
diff --git a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/UserServiceImpl.java b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/UserServiceImpl.java
index a817f0c61..8c0a5a174 100644
--- a/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/UserServiceImpl.java
+++ b/streampark-console/streampark-console-service/src/main/java/org/apache/streampark/console/system/service/impl/UserServiceImpl.java
@@ -26,7 +26,7 @@ import org.apache.streampark.console.base.exception.ApiAlertException;
 import org.apache.streampark.console.base.properties.ShiroProperties;
 import org.apache.streampark.console.base.util.ShaHashUtils;
 import org.apache.streampark.console.base.util.WebUtils;
-import org.apache.streampark.console.core.enums.LoginType;
+import org.apache.streampark.console.core.enums.LoginTypeEnum;
 import org.apache.streampark.console.core.service.ResourceService;
 import org.apache.streampark.console.core.service.application.ApplicationInfoService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
@@ -129,7 +129,7 @@ public class UserServiceImpl extends ServiceImpl<UserMapper, User> implements Us
   @Transactional(rollbackFor = Exception.class)
   public RestResponse updateUser(User user) {
     User existsUser = getById(user.getUserId());
-    user.setLoginType(null);
+    user.setLoginTypeEnum(null);
     user.setPassword(null);
     user.setModifyTime(new Date());
     if (needTransferResource(existsUser, user)) {
@@ -154,7 +154,7 @@ public class UserServiceImpl extends ServiceImpl<UserMapper, User> implements Us
     User user = getById(userParam.getUserId());
     ApiAlertException.throwIfNull(user, "User is null. Update password failed.");
     ApiAlertException.throwIfFalse(
-        user.getLoginType() == LoginType.PASSWORD,
+        user.getLoginTypeEnum() == LoginTypeEnum.PASSWORD,
         "Can only update password for user who sign in with PASSWORD");
 
     String saltPassword = ShaHashUtils.encrypt(user.getSalt(), userParam.getOldPassword());
diff --git a/streampark-console/streampark-console-service/src/main/resources/mapper/core/AlertConfigMapper.xml b/streampark-console/streampark-console-service/src/main/resources/mapper/core/AlertConfigMapper.xml
index 935da7b96..eb962865e 100644
--- a/streampark-console/streampark-console-service/src/main/resources/mapper/core/AlertConfigMapper.xml
+++ b/streampark-console/streampark-console-service/src/main/resources/mapper/core/AlertConfigMapper.xml
@@ -22,7 +22,7 @@
         <id column="id" jdbcType="BIGINT" property="id"/>
         <result column="user_id" jdbcType="BIGINT" property="userId"/>
         <result column="alert_name" jdbcType="VARCHAR" property="alertName"/>
-        <result column="alert_type" jdbcType="INTEGER" property="alertType"/>
+        <result column="alert_type" jdbcType="INTEGER" property="alertTypeEnum"/>
         <result column="email_params" jdbcType="VARCHAR" property="emailParams"/>
         <result column="ding_talk_params" jdbcType="VARCHAR" property="dingTalkParams"/>
         <result column="we_com_params" jdbcType="VARCHAR" property="weComParams"/>
diff --git a/streampark-console/streampark-console-service/src/main/resources/mapper/core/ApplicationMapper.xml b/streampark-console/streampark-console-service/src/main/resources/mapper/core/ApplicationMapper.xml
index f57b8efe7..d0b1e9c47 100644
--- a/streampark-console/streampark-console-service/src/main/resources/mapper/core/ApplicationMapper.xml
+++ b/streampark-console/streampark-console-service/src/main/resources/mapper/core/ApplicationMapper.xml
@@ -37,7 +37,7 @@
         <result column="app_type" jdbcType="INTEGER" property="appType"/>
         <result column="job_type" jdbcType="INTEGER" property="jobType"/>
         <result column="resource_from" jdbcType="INTEGER" property="resourceFrom"/>
-        <result column="execution_mode" jdbcType="INTEGER" property="executionMode"/>
+        <result column="execution_mode" jdbcType="INTEGER" property="executionModeEnum"/>
         <result column="tracking" jdbcType="INTEGER" property="tracking"/>
         <result column="jar" jdbcType="VARCHAR" property="jar"/>
         <result column="jar_check_sum" jdbcType="VARCHAR" property="jarCheckSum"/>
@@ -56,14 +56,14 @@
         <result column="restart_count" jdbcType="INTEGER" property="restartCount"/>
         <result column="release" jdbcType="INTEGER" property="release"/>
         <result column="build" jdbcType="BOOLEAN" property="build"/>
-        <result column="resolve_order" jdbcType="INTEGER" property="resolveOrder"/>
+        <result column="resolve_order" jdbcType="INTEGER" property="resolveOrderEnum"/>
         <result column="total_tm" jdbcType="INTEGER" property="totalTM"/>
         <result column="total_slot" jdbcType="INTEGER" property="totalSlot"/>
         <result column="available_slot" jdbcType="INTEGER" property="availableSlot"/>
         <result column="total_task" jdbcType="INTEGER" property="totalTask"/>
         <result column="jm_memory" jdbcType="INTEGER" property="jmMemory"/>
         <result column="tm_memory" jdbcType="INTEGER" property="tmMemory"/>
-        <result column="option_state" jdbcType="INTEGER" property="optionState"/>
+        <result column="option_state" jdbcType="INTEGER" property="optionStateEnum"/>
         <result column="alert_id" jdbcType="BIGINT" property="alertId"/>
         <result column="description" jdbcType="VARCHAR" property="description"/>
         <result column="create_time" jdbcType="DATE" property="createTime"/>
@@ -146,8 +146,8 @@
                     #{item}
                 </foreach>
             </if>
-            <if test="application.executionMode != null and application.executionMode != ''">
-                and t.execution_mode = #{application.executionMode}
+            <if test="application.executionModeEnum != null and application.executionModeEnum != ''">
+                and t.execution_mode = #{application.executionModeEnum}
             </if>
             <if test="application.jobName != null and application.jobName != ''">
                 and t.job_name like '%${application.jobName}%'
@@ -193,8 +193,8 @@
             <if test="application.tracking != null">
                 tracking=#{application.tracking},
             </if>
-            <if test="application.optionState != null">
-                option_state=#{application.optionState},
+            <if test="application.optionStateEnum != null">
+                option_state=#{application.optionStateEnum},
             </if>
             <if test="application.startTime != null">
                 start_time=#{application.startTime},
@@ -304,7 +304,7 @@
             select cluster_id, max(create_time) as ct
             from t_flink_app
             where cluster_id is not null
-            and execution_mode = #{executionMode}
+            and execution_mode = #{executionModeEnum}
             group by cluster_id
             order by ct desc
         ) as ci
diff --git a/streampark-console/streampark-console-service/src/main/resources/mapper/core/FlinkClusterMapper.xml b/streampark-console/streampark-console-service/src/main/resources/mapper/core/FlinkClusterMapper.xml
index 769fe521c..e6f3f1cf8 100644
--- a/streampark-console/streampark-console-service/src/main/resources/mapper/core/FlinkClusterMapper.xml
+++ b/streampark-console/streampark-console-service/src/main/resources/mapper/core/FlinkClusterMapper.xml
@@ -25,7 +25,7 @@
         <result column="cluster_name" jdbcType="VARCHAR" property="clusterName"/>
         <result column="options" jdbcType="LONGVARCHAR" property="options"/>
         <result column="yarn_queue" jdbcType="VARCHAR" property="yarnQueue"/>
-        <result column="execution_mode" jdbcType="TINYINT" property="executionMode"/>
+        <result column="execution_mode" jdbcType="TINYINT" property="executionModeEnum"/>
         <result column="version_id" jdbcType="BIGINT" property="versionId"/>
         <result column="k8s_namespace" jdbcType="VARCHAR" property="k8sNamespace"/>
         <result column="service_account" jdbcType="VARCHAR" property="serviceAccount"/>
@@ -36,9 +36,9 @@
         <result column="k8s_rest_exposed_type" jdbcType="TINYINT" property="k8sRestExposedType"/>
         <result column="k8s_hadoop_integration" jdbcType="BOOLEAN" property="k8sHadoopIntegration"/>
         <result column="k8s_conf" jdbcType="VARCHAR" property="k8sConf"/>
-        <result column="resolve_order" jdbcType="INTEGER" property="resolveOrder"/>
+        <result column="resolve_order" jdbcType="INTEGER" property="resolveOrderEnum"/>
         <result column="exception" jdbcType="LONGVARCHAR" property="exception"/>
-        <result column="cluster_state" jdbcType="TINYINT" property="clusterState"/>
+        <result column="cluster_state" jdbcType="TINYINT" property="clusterStateEnum"/>
         <result column="create_time" jdbcType="DATE" property="createTime"/>
         <result column="start_time" jdbcType="DATE" property="startTime"/>
         <result column="end_time" jdbcType="DATE" property="endTime"/>
diff --git a/streampark-console/streampark-console-service/src/main/resources/mapper/core/ProjectMapper.xml b/streampark-console/streampark-console-service/src/main/resources/mapper/core/ProjectMapper.xml
index 46f733e58..773737927 100644
--- a/streampark-console/streampark-console-service/src/main/resources/mapper/core/ProjectMapper.xml
+++ b/streampark-console/streampark-console-service/src/main/resources/mapper/core/ProjectMapper.xml
@@ -22,7 +22,7 @@
         <id column="id" jdbcType="INTEGER" property="id"/>
         <result column="team_id" jdbcType="BIGINT" property="teamId"/>
         <result column="name" jdbcType="VARCHAR" property="name"/>
-        <result column="git_credential" jdbcType="INTEGER" property="gitCredential"/>
+        <result column="git_credential" jdbcType="INTEGER" property="gitCredentialEnum"/>
         <result column="url" jdbcType="VARCHAR" property="url"/>
         <result column="repository" jdbcType="INTEGER" property="repository"/>
         <result column="branches" jdbcType="VARCHAR" property="branches"/>
@@ -34,7 +34,7 @@
         <result column="type" jdbcType="INTEGER" property="type"/>
         <result column="description" jdbcType="VARCHAR" property="description"/>
         <result column="last_build" jdbcType="TIMESTAMP" property="lastBuild"/>
-        <result column="build_state" jdbcType="INTEGER" property="buildState"/>
+        <result column="build_state" jdbcType="INTEGER" property="buildStateEnum"/>
         <result column="create_time" jdbcType="TIMESTAMP" property="createTime"/>
         <result column="modify_time" jdbcType="TIMESTAMP" property="modifyTime"/>
     </resultMap>
@@ -85,8 +85,8 @@
             <if test="project.name != null and project.name != ''">
                 and t.name like '%${project.name}%'
             </if>
-            <if test="project.buildState != null">
-                and t.build_state = #{project.buildState}
+            <if test="project.buildStateEnum != null">
+                and t.build_state = #{project.buildStateEnum}
             </if>
         </where>
     </select>
diff --git a/streampark-console/streampark-console-service/src/main/resources/mapper/system/UserMapper.xml b/streampark-console/streampark-console-service/src/main/resources/mapper/system/UserMapper.xml
index aeec2eb4f..a196cb1ce 100644
--- a/streampark-console/streampark-console-service/src/main/resources/mapper/system/UserMapper.xml
+++ b/streampark-console/streampark-console-service/src/main/resources/mapper/system/UserMapper.xml
@@ -31,7 +31,7 @@
         <result column="sex" jdbcType="CHAR" property="sex"/>
         <result column="description" jdbcType="VARCHAR" property="description"/>
         <result column="last_team_id" jdbcType="BIGINT" property="lastTeamId"/>
-        <result column="login_type" jdbcType="INTEGER" property="loginType"/>
+        <result column="login_type" jdbcType="INTEGER" property="loginTypeEnum"/>
     </resultMap>
 
     <select id="findUserDetail" resultType="org.apache.streampark.console.system.entity.User" parameterType="org.apache.streampark.console.system.entity.User">
diff --git a/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/task/FlinkK8sObserverBroker.scala b/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/task/FlinkK8sObserverBroker.scala
index 6725bba4b..f9b54d467 100644
--- a/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/task/FlinkK8sObserverBroker.scala
+++ b/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/task/FlinkK8sObserverBroker.scala
@@ -18,13 +18,13 @@
 package org.apache.streampark.console.core.task
 
 import org.apache.streampark.common.conf.K8sFlinkConfig
-import org.apache.streampark.common.enums.{ClusterState, ExecutionMode}
+import org.apache.streampark.common.enums.{ClusterStateEnum, ExecutionModeEnum}
 import org.apache.streampark.common.zio.{OptionTraversableOps, PrettyStringOps}
 import org.apache.streampark.common.zio.ZIOContainerSubscription.{ConcurrentMapExtension, RefMapExtension}
 import org.apache.streampark.common.zio.ZIOExt.{IterableZStreamConverter, OptionZIOOps, UIOOps, ZStreamOptionEffectOps}
 import org.apache.streampark.console.core.bean.AlertTemplate
 import org.apache.streampark.console.core.entity.{Application, FlinkCluster}
-import org.apache.streampark.console.core.enums.{FlinkAppState, OptionState}
+import org.apache.streampark.console.core.enums.{FlinkAppStateEnum, OptionStateEnum}
 import org.apache.streampark.console.core.service.FlinkClusterService
 import org.apache.streampark.console.core.service.alert.AlertService
 import org.apache.streampark.console.core.service.application.ApplicationInfoService
@@ -61,18 +61,18 @@ class FlinkK8sObserverBroker @Autowired() (
 
   private val observer = FlinkK8sObserver
 
-  private val alertJobStateList: Array[FlinkAppState] = Array(
-    FlinkAppState.FAILED,
-    FlinkAppState.LOST,
-    FlinkAppState.RESTARTING,
-    FlinkAppState.FINISHED
+  private val alertJobStateList: Array[FlinkAppStateEnum] = Array(
+    FlinkAppStateEnum.FAILED,
+    FlinkAppStateEnum.LOST,
+    FlinkAppStateEnum.RESTARTING,
+    FlinkAppStateEnum.FINISHED
   )
 
   private val alertClusterStateList = Array(
-    ClusterState.FAILED,
-    ClusterState.UNKNOWN,
-    ClusterState.LOST,
-    ClusterState.KILLED
+    ClusterStateEnum.FAILED,
+    ClusterStateEnum.UNKNOWN,
+    ClusterStateEnum.LOST,
+    ClusterStateEnum.KILLED
   )
 
   private lazy val allDaemonEffects: Array[UIO[Unit]] = Array(
@@ -113,7 +113,7 @@ class FlinkK8sObserverBroker @Autowired() (
     val fromApplicationRecords: UIO[Unit] = it
       .safeFindApplication(
         new LambdaQueryWrapper[Application]
-          .in(Application.SFunc.EXECUTION_MODE, ExecutionMode.getKubernetesMode)
+          .in(Application.SFunc.EXECUTION_MODE, ExecutionModeEnum.getKubernetesMode)
       )(10)
       .map(apps => apps.map(app => applicationToTrackKey(app)).filterSome.toVector)
       .tap(keys => logInfo(s"Restore Flink K8s track-keys from Application records:\n${keys.prettyStr}"))
@@ -122,7 +122,7 @@ class FlinkK8sObserverBroker @Autowired() (
     val fromFlinkClusterRecords: UIO[Unit] = it
       .safeFindFlinkClusterRecord(
         new LambdaQueryWrapper[FlinkCluster]
-          .eq(FlinkCluster.SFunc.EXECUTION_MODE, ExecutionMode.KUBERNETES_NATIVE_SESSION.getMode)
+          .eq(FlinkCluster.SFunc.EXECUTION_MODE, ExecutionModeEnum.KUBERNETES_NATIVE_SESSION.getMode)
       )(10)
       .map(clusters => clusters.map(e => TrackKey.cluster(e.getId, e.getK8sNamespace, e.getClusterId)))
       .tap(keys => logInfo(s"Restore Flink K8s track-keys from FlinkCluster records:\n${keys.prettyStr}"))
@@ -138,12 +138,12 @@ class FlinkK8sObserverBroker @Autowired() (
       // Convert EvalJobState to FlinkAppState
       .map(snap => snap -> FlinkK8sDataTypeConverter.k8sEvalJobStateToFlinkAppState(snap.evalState))
       // Update the corresponding columns of Application record
-      .tap { case (snap: JobSnapshot, convertedState: FlinkAppState) =>
+      .tap { case (snap: JobSnapshot, convertedState: FlinkAppStateEnum) =>
         safeUpdateApplicationRecord(snap.appId) {
 
           var update = new LambdaUpdateWrapper[Application]
             .set(Application.SFunc.STATE, convertedState.getValue)
-            .set(Application.SFunc.OPTIONS, OptionState.NONE.getValue)
+            .set(Application.SFunc.OPTIONS, OptionStateEnum.NONE.getValue)
           // update JobStatus related columns
           snap.jobStatus.foreach { status =>
             update = update
@@ -154,7 +154,7 @@ class FlinkK8sObserverBroker @Autowired() (
               .set(Application.SFunc.TOTAL_TASK, status.tasks.map(_.total).getOrElse(0))
           }
           // Copy the logic from resources/mapper/core/ApplicationMapper.xml:persistMetrics
-          if (FlinkAppState.isEndState(convertedState.getValue)) {
+          if (FlinkAppStateEnum.isEndState(convertedState.getValue)) {
             update = update
               .set(Application.SFunc.TOTAL_TM, null)
               .set(Application.SFunc.TOTAL_SLOT, null)
diff --git a/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/utils/FlinkK8sDataTypeConverter.scala b/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/utils/FlinkK8sDataTypeConverter.scala
index 846ca61b2..5f3fdb5e6 100644
--- a/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/utils/FlinkK8sDataTypeConverter.scala
+++ b/streampark-console/streampark-console-service/src/main/scala/org/apache/streampark/console/core/utils/FlinkK8sDataTypeConverter.scala
@@ -18,9 +18,9 @@
 package org.apache.streampark.console.core.utils
 
 import org.apache.streampark.common.conf.{InternalConfigHolder, K8sFlinkConfig}
-import org.apache.streampark.common.enums.{ClusterState, ExecutionMode}
+import org.apache.streampark.common.enums.{ClusterStateEnum, ExecutionModeEnum}
 import org.apache.streampark.console.core.entity.{Application, FlinkCluster}
-import org.apache.streampark.console.core.enums.FlinkAppState
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum
 import org.apache.streampark.console.core.service.SettingService
 import org.apache.streampark.console.core.utils.FlinkK8sDataTypeConverter.genSessionJobCRName
 import org.apache.streampark.flink.kubernetes.model.FlinkMetricCV
@@ -72,20 +72,20 @@ object FlinkK8sDataTypeConverter {
     s"$clusterId-${UUID.randomUUID().toString.replace("-", "").take(8)}"
   }
 
-  /** Convert [[EvalJobState]] to [[FlinkAppState]]. */
-  def k8sEvalJobStateToFlinkAppState(jobState: EvalJobState): FlinkAppState = {
-    Try(FlinkAppState.valueOf(jobState.toString)).getOrElse(FlinkAppState.OTHER)
+  /** Convert [[EvalJobState]] to [[FlinkAppStateEnum]]. */
+  def k8sEvalJobStateToFlinkAppState(jobState: EvalJobState): FlinkAppStateEnum = {
+    Try(FlinkAppStateEnum.valueOf(jobState.toString)).getOrElse(FlinkAppStateEnum.OTHER)
   }
 
-  /** Convert [[DeployCRStatus]] to [[ClusterState]]. */
-  def k8sDeployStateToClusterState(crState: DeployCRStatus): ClusterState = {
+  /** Convert [[DeployCRStatus]] to [[ClusterStateEnum]]. */
+  def k8sDeployStateToClusterState(crState: DeployCRStatus): ClusterStateEnum = {
     crState.evalState match {
-      case EvalState.DEPLOYING => ClusterState.STARTING
-      case EvalState.READY     => ClusterState.RUNNING
-      case EvalState.SUSPENDED => ClusterState.CANCELED
-      case EvalState.FAILED    => ClusterState.FAILED
-      case EvalState.DELETED   => ClusterState.KILLED
-      case _                   => ClusterState.UNKNOWN
+      case EvalState.DEPLOYING => ClusterStateEnum.STARTING
+      case EvalState.READY     => ClusterStateEnum.RUNNING
+      case EvalState.SUSPENDED => ClusterStateEnum.CANCELED
+      case EvalState.FAILED    => ClusterStateEnum.FAILED
+      case EvalState.DELETED   => ClusterStateEnum.KILLED
+      case _                   => ClusterStateEnum.UNKNOWN
     }
   }
 
@@ -110,7 +110,7 @@ object FlinkK8sDataTypeConverter {
     val isLegal = {
       flinkCluster != null &&
       flinkCluster.getId != null &&
-      ExecutionMode.isKubernetesSessionMode(flinkCluster.getExecutionMode) &&
+      ExecutionModeEnum.isKubernetesSessionMode(flinkCluster.getExecutionMode) &&
       StringUtils.isNoneBlank(flinkCluster.getClusterId) &&
       StringUtils.isNoneBlank(flinkCluster.getK8sNamespace)
     }
@@ -120,12 +120,12 @@ object FlinkK8sDataTypeConverter {
 
   /** Convert [[Application]] to [[TrackKey]]. */
   def applicationToTrackKey(app: Application): Option[TrackKey] = {
-    import ExecutionMode._
+    import org.apache.streampark.common.enums.ExecutionModeEnum._
 
     val isLegal = {
       app != null &&
       app.getId != null &&
-      ExecutionMode.isKubernetesSessionMode(app.getExecutionMode) &&
+      ExecutionModeEnum.isKubernetesSessionMode(app.getExecutionMode) &&
       StringUtils.isNoneBlank(app.getClusterId) &&
       StringUtils.isNoneBlank(app.getK8sNamespace)
     }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/SpringUnitTestBase.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/SpringUnitTestBase.java
index 490df51c0..0ade0bf3e 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/SpringUnitTestBase.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/SpringUnitTestBase.java
@@ -19,7 +19,7 @@ package org.apache.streampark.console;
 
 import org.apache.streampark.common.conf.CommonConfig;
 import org.apache.streampark.common.conf.ConfigConst;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.YarnQueue;
@@ -90,17 +90,17 @@ public abstract class SpringUnitTestBase {
     cluster.setClusterName(name);
     cluster.setYarnQueue(yarnQueue);
     cluster.setVersionId(versionId);
-    cluster.setExecutionMode(ExecutionMode.YARN_SESSION.getMode());
+    cluster.setExecutionMode(ExecutionModeEnum.YARN_SESSION.getMode());
     return cluster;
   }
 
   protected Application mockYarnModeJobApp(
-      Long teamId, String name, String yarnQueue, ExecutionMode executionMode) {
+      Long teamId, String name, String yarnQueue, ExecutionModeEnum executionModeEnum) {
     Application application = new Application();
     application.setYarnQueue(yarnQueue);
     application.setTeamId(teamId);
     application.setJobName(name);
-    application.setExecutionMode(executionMode.getMode());
+    application.setExecutionMode(executionModeEnum.getMode());
     application.doSetHotParams();
     return application;
   }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/entity/ProjectTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/entity/ProjectTest.java
index f2d060d76..6c1f6807a 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/entity/ProjectTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/entity/ProjectTest.java
@@ -17,8 +17,8 @@
 
 package org.apache.streampark.console.core.entity;
 
-import org.apache.streampark.console.core.enums.GitAuthorizedError;
-import org.apache.streampark.console.core.enums.GitCredential;
+import org.apache.streampark.console.core.enums.GitAuthorizedErrorEnum;
+import org.apache.streampark.console.core.enums.GitCredentialEnum;
 
 import lombok.extern.slf4j.Slf4j;
 import org.junit.jupiter.api.BeforeEach;
@@ -35,7 +35,7 @@ class ProjectTest {
   @BeforeEach
   void before() {
     project.setUrl("https://github.com/apache/incubator-streampark.git");
-    project.setGitCredential(GitCredential.HTTPS.getValue());
+    project.setGitCredential(GitCredentialEnum.HTTPS.getValue());
   }
 
   @Disabled("This test case can't be runnable due to external service is not available.")
@@ -48,7 +48,7 @@ class ProjectTest {
   @Disabled("This test case can't be runnable due to external service is not available.")
   @Test
   void testGitCheckAuth() {
-    GitAuthorizedError error = project.gitCheck();
+    GitAuthorizedErrorEnum error = project.gitCheck();
     log.error("{}", error);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeEnumTest.java
similarity index 80%
rename from streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeTest.java
rename to streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeEnumTest.java
index c19755c79..148823de8 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/enums/AlertTypeEnumTest.java
@@ -24,16 +24,16 @@ import java.util.List;
 
 import static org.assertj.core.api.Assertions.assertThat;
 
-class AlertTypeTest {
+class AlertTypeEnumTest {
   @Test
   void testDecode() {
-    List<AlertType> notifyTypes = AlertType.decode(5);
-    assertThat(notifyTypes).containsExactly(AlertType.EMAIL, AlertType.WE_COM);
+    List<AlertTypeEnum> notifyTypes = AlertTypeEnum.decode(5);
+    assertThat(notifyTypes).containsExactly(AlertTypeEnum.EMAIL, AlertTypeEnum.WE_COM);
   }
 
   @Test
   void testEncode() {
-    int level = AlertType.encode(Arrays.asList(AlertType.DING_TALK, AlertType.EMAIL));
+    int level = AlertTypeEnum.encode(Arrays.asList(AlertTypeEnum.DING_TALK, AlertTypeEnum.EMAIL));
     assertThat(level).isEqualTo(3);
   }
 }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceITest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceITest.java
index 79ee5ca43..4b5be37bf 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceITest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceITest.java
@@ -17,15 +17,15 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.DeflaterUtils;
 import org.apache.streampark.console.SpringIntegrationTestBase;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.FlinkEnv;
 import org.apache.streampark.console.core.entity.FlinkSql;
-import org.apache.streampark.console.core.enums.FlinkAppState;
-import org.apache.streampark.console.core.enums.ReleaseState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
+import org.apache.streampark.console.core.enums.ReleaseStateEnum;
 import org.apache.streampark.console.core.service.application.ApplicationActionService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.console.core.service.impl.FlinkClusterServiceImpl;
@@ -102,7 +102,7 @@ class ApplicationManageServiceITest extends SpringIntegrationTestBase {
     FlinkCluster flinkCluster = new FlinkCluster();
     flinkCluster.setId(1L);
     flinkCluster.setAddress(cluster.getFlinkJobManagerUrl());
-    flinkCluster.setExecutionMode(ExecutionMode.REMOTE.getMode());
+    flinkCluster.setExecutionMode(ExecutionModeEnum.REMOTE.getMode());
     flinkCluster.setClusterName("docker-Cluster-1.17.1");
     flinkCluster.setVersionId(1L);
     flinkCluster.setUserId(100000L);
@@ -114,7 +114,7 @@ class ApplicationManageServiceITest extends SpringIntegrationTestBase {
     application.setFlinkClusterId(1L);
     application.setSqlId(100000L);
     application.setVersionId(1L);
-    application.setExecutionMode(ExecutionMode.REMOTE.getMode());
+    application.setExecutionMode(ExecutionModeEnum.REMOTE.getMode());
 
     // Avoid exceptional error.
     application.setFlinkSql(
@@ -132,7 +132,7 @@ class ApplicationManageServiceITest extends SpringIntegrationTestBase {
             () -> {
               while (true) {
                 Application app = applicationManageService.getById(100000L);
-                if (app != null && app.getReleaseState() == ReleaseState.DONE) {
+                if (app != null && app.getReleaseState() == ReleaseStateEnum.DONE) {
                   break;
                 }
               }
@@ -146,7 +146,7 @@ class ApplicationManageServiceITest extends SpringIntegrationTestBase {
             () -> {
               while (true) {
                 if (flinkAppHttpWatcher.tryQueryFlinkAppState(application.getId())
-                    == FlinkAppState.RUNNING) {
+                    == FlinkAppStateEnum.RUNNING) {
                   break;
                 }
               }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceTest.java
index 3f67ae7dd..eef17730e 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/ApplicationManageServiceTest.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.SpringUnitTestBase;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.YarnQueue;
@@ -111,7 +111,7 @@ class ApplicationManageServiceTest extends SpringUnitTestBase {
     YarnQueue yarnQueue = mockYarnQueue(targetTeamId, queueLabel);
     yarnQueueService.save(yarnQueue);
     Application application =
-        mockYarnModeJobApp(targetTeamId, "app1", queueLabel, ExecutionMode.YARN_APPLICATION);
+        mockYarnModeJobApp(targetTeamId, "app1", queueLabel, ExecutionModeEnum.YARN_APPLICATION);
     assertThat(applicationServiceImpl.validateQueueIfNeeded(application)).isTrue();
 
     // Test application without available queue
@@ -127,9 +127,9 @@ class ApplicationManageServiceTest extends SpringUnitTestBase {
 
     // Test update for both versions in yarn-app or per-job with same yarn queue
     Application app1 =
-        mockYarnModeJobApp(teamId2, appName, queueLabel1, ExecutionMode.YARN_APPLICATION);
+        mockYarnModeJobApp(teamId2, appName, queueLabel1, ExecutionModeEnum.YARN_APPLICATION);
     Application app2 =
-        mockYarnModeJobApp(teamId2, appName, queueLabel1, ExecutionMode.YARN_PER_JOB);
+        mockYarnModeJobApp(teamId2, appName, queueLabel1, ExecutionModeEnum.YARN_PER_JOB);
     assertThat(applicationServiceImpl.validateQueueIfNeeded(app1, app2)).isTrue();
 
     // Test available queue
@@ -141,7 +141,7 @@ class ApplicationManageServiceTest extends SpringUnitTestBase {
     assertThat(applicationServiceImpl.validateQueueIfNeeded(app1, app2)).isTrue();
 
     // Test non-existed queue
-    app1.setExecutionMode(ExecutionMode.KUBERNETES_NATIVE_APPLICATION.getMode());
+    app1.setExecutionMode(ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION.getMode());
     app2.setYarnQueue(nonExistedQueue);
     assertThat(applicationServiceImpl.validateQueueIfNeeded(app1, app2)).isFalse();
   }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/FlinkClusterServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/FlinkClusterServiceTest.java
index c1fef933c..b72ca2e97 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/FlinkClusterServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/FlinkClusterServiceTest.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.SpringUnitTestBase;
 import org.apache.streampark.console.core.entity.FlinkCluster;
 import org.apache.streampark.console.core.entity.YarnQueue;
@@ -95,7 +95,7 @@ class FlinkClusterServiceTest extends SpringUnitTestBase {
     assertThat(clusterServiceImpl.validateQueueIfNeeded(cluster1, cluster2)).isTrue();
 
     // Test non-existed queue
-    cluster1.setExecutionMode(ExecutionMode.KUBERNETES_NATIVE_APPLICATION.getMode());
+    cluster1.setExecutionMode(ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION.getMode());
     cluster2.setYarnQueue(nonExistedQueue);
     assertThat(clusterServiceImpl.validateQueueIfNeeded(cluster1, cluster2)).isFalse();
   }
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/SavePointServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/SavePointServiceTest.java
index 47decc93b..20d749a48 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/SavePointServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/SavePointServiceTest.java
@@ -17,17 +17,17 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ApplicationType;
-import org.apache.streampark.common.enums.DevelopmentMode;
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ApplicationTypeEnum;
+import org.apache.streampark.common.enums.DevelopmentModeEnum;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.common.util.DeflaterUtils;
 import org.apache.streampark.console.SpringUnitTestBase;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.ApplicationConfig;
 import org.apache.streampark.console.core.entity.Effective;
 import org.apache.streampark.console.core.entity.FlinkEnv;
-import org.apache.streampark.console.core.enums.ConfigFileType;
-import org.apache.streampark.console.core.enums.EffectiveType;
+import org.apache.streampark.console.core.enums.ConfigFileTypeEnum;
+import org.apache.streampark.console.core.enums.EffectiveTypeEnum;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.console.core.service.impl.SavePointServiceImpl;
 
@@ -94,17 +94,17 @@ class SavePointServiceTest extends SpringUnitTestBase {
     app.setId(appId);
 
     // Test for non-(StreamPark job Or FlinkSQL job)
-    app.setAppType(ApplicationType.APACHE_FLINK.getType());
+    app.setAppType(ApplicationTypeEnum.APACHE_FLINK.getType());
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app)).isNull();
-    app.setAppType(ApplicationType.STREAMPARK_FLINK.getType());
-    app.setJobType(DevelopmentMode.CUSTOM_CODE.getMode());
+    app.setAppType(ApplicationTypeEnum.STREAMPARK_FLINK.getType());
+    app.setJobType(DevelopmentModeEnum.CUSTOM_CODE.getMode());
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app)).isNull();
 
     // Test for (StreamPark job Or FlinkSQL job) without application config.
-    app.setAppType(ApplicationType.STREAMPARK_FLINK.getType());
+    app.setAppType(ApplicationTypeEnum.STREAMPARK_FLINK.getType());
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app)).isNull();
-    app.setAppType(ApplicationType.STREAMPARK_FLINK.getType());
-    app.setJobType(DevelopmentMode.CUSTOM_CODE.getMode());
+    app.setAppType(ApplicationTypeEnum.STREAMPARK_FLINK.getType());
+    app.setJobType(DevelopmentModeEnum.CUSTOM_CODE.getMode());
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app)).isNull();
 
     // Test for (StreamPark job Or FlinkSQL job) with application config just disabled checkpoint.
@@ -112,7 +112,7 @@ class SavePointServiceTest extends SpringUnitTestBase {
     appCfg.setId(appCfgId);
     appCfg.setAppId(appId);
     appCfg.setContent("state.savepoints.dir=hdfs:///test");
-    appCfg.setFormat(ConfigFileType.PROPERTIES.getValue());
+    appCfg.setFormat(ConfigFileTypeEnum.PROPERTIES.getValue());
     configService.save(appCfg);
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app)).isNull();
 
@@ -133,7 +133,7 @@ class SavePointServiceTest extends SpringUnitTestBase {
     Effective effective = new Effective();
     effective.setTargetId(appCfg.getId());
     effective.setAppId(appId);
-    effective.setTargetType(EffectiveType.CONFIG.getType());
+    effective.setTargetType(EffectiveTypeEnum.CONFIG.getType());
     effectiveService.save(effective);
     assertThat(savePointServiceImpl.getSavepointFromAppCfgIfStreamParkOrSQLJob(app))
         .isEqualTo("hdfs:///test");
@@ -149,7 +149,7 @@ class SavePointServiceTest extends SpringUnitTestBase {
     application.setId(appId);
     application.setTeamId(teamId);
     application.setVersionId(idOfFlinkEnv);
-    application.setExecutionMode(ExecutionMode.YARN_APPLICATION.getMode());
+    application.setExecutionMode(ExecutionModeEnum.YARN_APPLICATION.getMode());
     applicationManageService.save(application);
 
     FlinkEnv flinkEnv = new FlinkEnv();
@@ -169,7 +169,7 @@ class SavePointServiceTest extends SpringUnitTestBase {
     Long clusterId = 1L;
 
     // Test for it without cluster.
-    application.setExecutionMode(ExecutionMode.REMOTE.getMode());
+    application.setExecutionMode(ExecutionModeEnum.REMOTE.getMode());
     application.setFlinkClusterId(clusterId);
     assertThatThrownBy(() -> savePointServiceImpl.getSavepointFromDeployLayer(application))
         .isInstanceOf(NullPointerException.class);
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/UserServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/UserServiceTest.java
index cd742c58e..67619d4b9 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/UserServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/UserServiceTest.java
@@ -21,9 +21,9 @@ import org.apache.streampark.console.SpringUnitTestBase;
 import org.apache.streampark.console.base.domain.RestResponse;
 import org.apache.streampark.console.core.entity.Application;
 import org.apache.streampark.console.core.entity.Resource;
-import org.apache.streampark.console.core.enums.EngineType;
-import org.apache.streampark.console.core.enums.ResourceType;
-import org.apache.streampark.console.core.enums.UserType;
+import org.apache.streampark.console.core.enums.EngineTypeEnum;
+import org.apache.streampark.console.core.enums.ResourceTypeEnum;
+import org.apache.streampark.console.core.enums.UserTypeEnum;
 import org.apache.streampark.console.core.service.application.ApplicationInfoService;
 import org.apache.streampark.console.core.service.application.ApplicationManageService;
 import org.apache.streampark.console.system.entity.User;
@@ -52,7 +52,7 @@ class UserServiceTest extends SpringUnitTestBase {
     user.setUsername("test");
     user.setNickName("test");
     user.setPassword("test");
-    user.setUserType(UserType.USER);
+    user.setUserTypeEnum(UserTypeEnum.USER);
     user.setStatus(User.STATUS_VALID);
     userService.createUser(user);
     // lock user
@@ -74,8 +74,8 @@ class UserServiceTest extends SpringUnitTestBase {
 
     Resource resource = new Resource();
     resource.setResourceName("test");
-    resource.setResourceType(ResourceType.FLINK_APP);
-    resource.setEngineType(EngineType.FLINK);
+    resource.setResourceTypeEnum(ResourceTypeEnum.FLINK_APP);
+    resource.setEngineTypeEnum(EngineTypeEnum.FLINK);
     resource.setTeamId(1L);
     resource.setCreatorId(user.getUserId());
     resourceService.save(resource);
@@ -95,14 +95,14 @@ class UserServiceTest extends SpringUnitTestBase {
     user.setUsername("test");
     user.setNickName("test");
     user.setPassword("test");
-    user.setUserType(UserType.USER);
+    user.setUserTypeEnum(UserTypeEnum.USER);
     user.setStatus(User.STATUS_VALID);
     userService.save(user);
 
     Resource resource = new Resource();
     resource.setResourceName("test");
-    resource.setResourceType(ResourceType.FLINK_APP);
-    resource.setEngineType(EngineType.FLINK);
+    resource.setResourceTypeEnum(ResourceTypeEnum.FLINK_APP);
+    resource.setEngineTypeEnum(EngineTypeEnum.FLINK);
     resource.setTeamId(1L);
     resource.setCreatorId(user.getUserId());
     resourceService.save(resource);
@@ -116,7 +116,7 @@ class UserServiceTest extends SpringUnitTestBase {
     targetUser.setUsername("test0");
     targetUser.setNickName("test0");
     targetUser.setPassword("test0");
-    targetUser.setUserType(UserType.USER);
+    targetUser.setUserTypeEnum(UserTypeEnum.USER);
     targetUser.setStatus(User.STATUS_VALID);
     userService.save(targetUser);
 
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/YarnQueueServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/YarnQueueServiceTest.java
index 85c5f4432..00bec0e25 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/YarnQueueServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/YarnQueueServiceTest.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.console.core.service;
 
-import org.apache.streampark.common.enums.ExecutionMode;
+import org.apache.streampark.common.enums.ExecutionModeEnum;
 import org.apache.streampark.console.SpringUnitTestBase;
 import org.apache.streampark.console.base.domain.RestRequest;
 import org.apache.streampark.console.base.exception.ApiAlertException;
@@ -261,27 +261,27 @@ class YarnQueueServiceTest extends SpringUnitTestBase {
     yarnQueueServiceImpl.checkNotReferencedByApplications(targetTeamId, queueLabel, operation);
 
     // Test for existed applications that don't belong to the same team, not in yarn mode.
-    applicationManageService.save(mockYarnModeJobApp(2L, "app1", null, ExecutionMode.REMOTE));
+    applicationManageService.save(mockYarnModeJobApp(2L, "app1", null, ExecutionModeEnum.REMOTE));
     yarnQueueServiceImpl.checkNotReferencedByApplications(targetTeamId, queueLabel, operation);
 
     // Test for existed applications that don't belong to the same team, in yarn mode
     applicationManageService.save(
-        mockYarnModeJobApp(2L, "app2", null, ExecutionMode.YARN_APPLICATION));
+        mockYarnModeJobApp(2L, "app2", null, ExecutionModeEnum.YARN_APPLICATION));
     yarnQueueServiceImpl.checkNotReferencedByApplications(targetTeamId, queueLabel, operation);
 
     // Test for existed applications that belong to the same team, but not in yarn mode.
     applicationManageService.save(
-        mockYarnModeJobApp(targetTeamId, "app3", null, ExecutionMode.REMOTE));
+        mockYarnModeJobApp(targetTeamId, "app3", null, ExecutionModeEnum.REMOTE));
     yarnQueueServiceImpl.checkNotReferencedByApplications(targetTeamId, queueLabel, operation);
 
     // Test for existed applications that belong to the same team, but without yarn queue value.
     applicationManageService.save(
-        mockYarnModeJobApp(targetTeamId, "app4", null, ExecutionMode.YARN_PER_JOB));
+        mockYarnModeJobApp(targetTeamId, "app4", null, ExecutionModeEnum.YARN_PER_JOB));
     yarnQueueServiceImpl.checkNotReferencedByApplications(targetTeamId, queueLabel, operation);
 
     // Test for existed applications, some apps belong to the same team, but others don't belong to.
     applicationManageService.save(
-        mockYarnModeJobApp(targetTeamId, "app5", queueLabel, ExecutionMode.YARN_PER_JOB));
+        mockYarnModeJobApp(targetTeamId, "app5", queueLabel, ExecutionModeEnum.YARN_PER_JOB));
     assertThatThrownBy(
             () ->
                 yarnQueueServiceImpl.checkNotReferencedByApplications(
diff --git a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/alert/AlertServiceTest.java b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/alert/AlertServiceTest.java
index 54e592c46..c529e02bb 100644
--- a/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/alert/AlertServiceTest.java
+++ b/streampark-console/streampark-console-service/src/test/java/org/apache/streampark/console/core/service/alert/AlertServiceTest.java
@@ -27,7 +27,7 @@ import org.apache.streampark.console.core.bean.AlertTemplate;
 import org.apache.streampark.console.core.bean.AlertWeComParams;
 import org.apache.streampark.console.core.bean.EmailConfig;
 import org.apache.streampark.console.core.entity.Application;
-import org.apache.streampark.console.core.enums.FlinkAppState;
+import org.apache.streampark.console.core.enums.FlinkAppStateEnum;
 import org.apache.streampark.console.core.service.alert.impl.DingTalkAlertNotifyServiceImpl;
 import org.apache.streampark.console.core.service.alert.impl.LarkAlertNotifyServiceImpl;
 import org.apache.streampark.console.core.service.alert.impl.WeComAlertNotifyServiceImpl;
@@ -169,7 +169,7 @@ class AlertServiceTest {
     application.setCpFailureRateInterval(30);
     application.setCpMaxFailureInterval(5);
 
-    FlinkAppState appState = FlinkAppState.FAILED;
+    FlinkAppStateEnum appState = FlinkAppStateEnum.FAILED;
 
     try {
       AlertTemplate mail = getAlertBaseInfo(application);
diff --git a/streampark-console/streampark-console-webapp/src/views/base/login/LoginForm.vue b/streampark-console/streampark-console-webapp/src/views/base/login/LoginForm.vue
index a2cd069a3..1e37eb5f1 100644
--- a/streampark-console/streampark-console-webapp/src/views/base/login/LoginForm.vue
+++ b/streampark-console/streampark-console-webapp/src/views/base/login/LoginForm.vue
@@ -121,7 +121,7 @@
   const loading = ref(false);
   const userId = ref('');
   const modelVisible = ref(false);
-  const loginType = ref(LoginTypeEnum.PASSWORD);
+  const loginTypeEnum = ref(LoginTypeEnum.PASSWORD);
   const enableSSO = ref(false);
   const enableLDAP = ref(false);
 
@@ -133,7 +133,7 @@
   const loginText = computed(() => {
     const localText = t('sys.login.loginButton');
     const ldapText = t('sys.login.ldapTip');
-    if (loginType.value === LoginTypeEnum.PASSWORD) {
+    if (loginTypeEnum.value === LoginTypeEnum.PASSWORD) {
       return { buttonText: localText, linkText: t('sys.login.ldapTip') };
     }
     return { buttonText: ldapText, linkText: t('sys.login.passwordTip') };
@@ -158,7 +158,7 @@
       {
         password: loginFormValue.password,
         username: loginFormValue.account,
-        loginType: LoginTypeEnum[loginType.value],
+        loginTypeEnum: LoginTypeEnum[loginTypeEnum.value],
       },
       'none',
     );
@@ -222,11 +222,11 @@
   }
 
   function changeLoginType() {
-    if (loginType.value === LoginTypeEnum.PASSWORD) {
-      loginType.value = LoginTypeEnum.LDAP;
+    if (loginTypeEnum.value === LoginTypeEnum.PASSWORD) {
+      loginTypeEnum.value = LoginTypeEnum.LDAP;
       return;
     }
-    loginType.value = LoginTypeEnum.PASSWORD;
+    loginTypeEnum.value = LoginTypeEnum.PASSWORD;
   }
 
   onMounted(() => {
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/Add.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/Add.vue
index 39421b216..bb41ca4d2 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/Add.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/Add.vue
@@ -96,7 +96,7 @@
   /* Initialize the form */
   async function handleInitForm() {
     const defaultValue = {
-      resolveOrder: 0,
+      resolveOrderEnum: 0,
       k8sRestExposedType: 0,
     };
     options.forEach((item) => {
@@ -137,18 +137,18 @@
 
   function handleCluster(values: Recordable) {
     let flinkClusterId =
-      values.executionMode == ExecModeEnum.YARN_SESSION
+      values.executionModeEnum == ExecModeEnum.YARN_SESSION
         ? values.yarnSessionClusterId
         : values.flinkClusterId;
     const cluster =
       unref(flinkClusters).filter((c) => {
         if (flinkClusterId) {
-          return c.id == flinkClusterId && c.clusterState === ClusterStateEnum.RUNNING;
+          return c.id == flinkClusterId && c.clusterStateEnum === ClusterStateEnum.RUNNING;
         }
       })[0] || null;
     if (cluster) {
       Object.assign(values, { flinkClusterId: cluster.id });
-      if (values.executionMode == ExecModeEnum.KUBERNETES_SESSION) {
+      if (values.executionModeEnum == ExecModeEnum.KUBERNETES_SESSION) {
         Object.assign(values, { clusterId: cluster.clusterId });
       }
     }
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/Detail.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/Detail.vue
index 9b43c936c..7db24b4f5 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/Detail.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/Detail.vue
@@ -155,7 +155,7 @@
           ExecModeEnum.YARN_PER_JOB,
           ExecModeEnum.YARN_SESSION,
           ExecModeEnum.YARN_APPLICATION,
-        ].includes(res.executionMode)
+        ].includes(res.executionModeEnum)
       ) {
         handleYarn();
       }
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/EditFlink.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/EditFlink.vue
index 88691cc99..a68bd976e 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/EditFlink.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/EditFlink.vue
@@ -73,7 +73,7 @@
   });
 
   /* Form reset */
-  function handleReset(executionMode?: string) {
+  function handleReset(executionModeEnum?: string) {
     nextTick(async () => {
       let selectAlertId: string | undefined;
       if (app.alertId) {
@@ -88,8 +88,8 @@
         jar: app.jar,
         description: app.description,
         dynamicProperties: app.dynamicProperties,
-        resolveOrder: app.resolveOrder,
-        executionMode: app.executionMode,
+        resolveOrderEnum: app.resolveOrderEnum,
+        executionModeEnum: app.executionModeEnum,
         yarnQueue: app.yarnQueue,
         restartSize: app.restartSize,
         checkPointFailure: {
@@ -100,7 +100,7 @@
         versionId: app.versionId || null,
         k8sRestExposedType: app.k8sRestExposedType,
         clusterId: app.clusterId,
-        [app.executionMode == ExecModeEnum.YARN_SESSION
+        [app.executionModeEnum == ExecModeEnum.YARN_SESSION
           ? 'yarnSessionClusterId'
           : 'flinkClusterId']: app.flinkClusterId,
         flinkImage: app.flinkImage,
@@ -110,8 +110,8 @@
         module: app.module,
         ...resetParams,
       };
-      if (!executionMode) {
-        Object.assign(defaultParams, { executionMode: app.executionMode });
+      if (!executionModeEnum) {
+        Object.assign(defaultParams, { executionModeEnum: app.executionModeEnum });
       }
       setFieldsValue(defaultParams);
       app.args && programArgRef.value?.setContent(app.args);
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/EditStreamPark.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/EditStreamPark.vue
index 2225c38c6..bb7296ab9 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/EditStreamPark.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/EditStreamPark.vue
@@ -96,7 +96,7 @@
   const [registerReviewDrawer, { openDrawer: openReviewDrawer }] = useDrawer();
 
   /* Form reset */
-  function handleReset(executionMode?: string) {
+  function handleReset(executionModeEnum?: string) {
     let selectAlertId = '';
     if (app.alertId) {
       selectAlertId = unref(alerts).filter((t) => t.id == app.alertId)[0]?.id;
@@ -109,7 +109,7 @@
         args: app.args || '',
         description: app.description,
         dynamicProperties: app.dynamicProperties,
-        resolveOrder: app.resolveOrder,
+        resolveOrderEnum: app.resolveOrderEnum,
         versionId: app.versionId || null,
         teamResource: handleTeamResource(app.teamResource),
         k8sRestExposedType: app.k8sRestExposedType,
@@ -122,15 +122,15 @@
           cpFailureAction: app.cpFailureAction,
         },
         clusterId: app.clusterId,
-        [app.executionMode == ExecModeEnum.YARN_SESSION
+        [app.executionModeEnum == ExecModeEnum.YARN_SESSION
           ? 'yarnSessionClusterId'
           : 'flinkClusterId']: app.flinkClusterId,
         flinkImage: app.flinkImage,
         k8sNamespace: app.k8sNamespace,
         ...resetParams,
       };
-      if (!executionMode) {
-        Object.assign(defaultParams, { executionMode: app.executionMode });
+      if (!executionModeEnum) {
+        Object.assign(defaultParams, { executionModeEnum: app.executionModeEnum });
       }
       setFieldsValue(defaultParams);
       app.args && programArgRef.value?.setContent(app.args);
@@ -249,10 +249,10 @@
 
   /* Send submission interface */
   async function handleUpdateApp(params: Recordable) {
-    if (params.executionMode == ExecModeEnum.KUBERNETES_SESSION) {
+    if (params.executionModeEnum == ExecModeEnum.KUBERNETES_SESSION) {
       const cluster =
         unref(flinkClusters).filter((c) => {
-          return c.id == params.flinkClusterId && c.clusterState === ClusterStateEnum.RUNNING;
+          return c.id == params.flinkClusterId && c.clusterStateEnum === ClusterStateEnum.RUNNING;
         })[0] || null;
       if (cluster) {
         Object.assign(params, { clusterId: cluster.clusterId });
@@ -307,7 +307,7 @@
     setFieldsValue({
       jobType: res.jobType,
       appType: res.appType,
-      executionMode: res.executionMode,
+      executionModeEnum: res.executionModeEnum,
       flinkSql: res.flinkSql ? decodeByBase64(res.flinkSql) : '',
       dependency: '',
       module: res.module,
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/View.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/View.vue
index c75d24e9d..c2bb754fa 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/View.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/View.vue
@@ -89,7 +89,7 @@
             availableSlot: x.availableSlot,
           },
         ];
-        if (x['optionState'] === OptionStateEnum.NONE) {
+        if (x['optionStateEnum'] === OptionStateEnum.NONE) {
           if (optionApps.starting.get(x.id)) {
             if (timestamp - optionApps.starting.get(x.id) > 2000 * 2) {
               optionApps.starting.delete(x.id);
@@ -149,7 +149,7 @@
     // Task is running, restarting, in savePoint
     if (
       [AppStateEnum.RESTARTING, AppStateEnum.RUNNING].includes(app.state) ||
-      app['optionState'] === OptionStateEnum.SAVEPOINTING
+      app['optionStateEnum'] === OptionStateEnum.SAVEPOINTING
     ) {
       // yarn-per-job|yarn-session|yarn-application
       handleView(app, unref(yarn));
@@ -209,7 +209,7 @@
             :class="{
               'cursor-pointer':
                 [AppStateEnum.RESTARTING, AppStateEnum.RUNNING].includes(record.state) ||
-                record['optionState'] === OptionStateEnum.SAVEPOINTING,
+                record['optionStateEnum'] === OptionStateEnum.SAVEPOINTING,
             }"
             @click="handleJobView(record)"
           >
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/app/components/AppView/StartApplicationModal.vue b/streampark-console/streampark-console-webapp/src/views/flink/app/components/AppView/StartApplicationModal.vue
index 32bb55830..473ff6d27 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/app/components/AppView/StartApplicationModal.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/app/components/AppView/StartApplicationModal.vue
@@ -93,7 +93,7 @@
         required: true,
       },
       {
-        field: 'restoreMode',
+        field: 'restoreModeEnum',
         label: 'restore mode',
         component: 'Select',
         defaultValue: RestoreModeEnum.NO_CLAIM,
@@ -127,10 +127,10 @@
       const formValue = (await validate()) as Recordable;
       const savePointed = formValue.startSavePointed;
       const savePointPath = savePointed ? formValue['startSavePoint'] : null;
-      const restoreMode = savePointed ? formValue['restoreMode'] : null;
+      const restoreModeEnum = savePointed ? formValue['restoreModeEnum'] : null;
       const { data } = await fetchStart({
         id: receiveData.application.id,
-        restoreMode,
+        restoreModeEnum,
         savePointed,
         savePoint: savePointPath,
         allowNonRestored: formValue.allowNonRestoredState || false,
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/project/Add.vue b/streampark-console/streampark-console-webapp/src/views/flink/project/Add.vue
index 41c6215d9..babd8effb 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/project/Add.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/project/Add.vue
@@ -40,7 +40,7 @@
     try {
       const res = await createProject({
         name: values.name,
-        gitCredential: values.gitCredential,
+        gitCredentialEnum: values.gitCredentialEnum,
         url: values.url,
         repository: values.repository,
         type: values.type,
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/project/Edit.vue b/streampark-console/streampark-console-webapp/src/views/flink/project/Edit.vue
index 99748c59f..3c09e87ef 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/project/Edit.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/project/Edit.vue
@@ -37,7 +37,7 @@
   const route = useRoute();
   const router = useRouter();
 
-  const buildState = ref<Nullable<number | string> | undefined>(null);
+  const buildStateEnum = ref<Nullable<number | string> | undefined>(null);
   if (!route.query.id) {
     router.go(-1);
   }
@@ -48,9 +48,9 @@
       projectResource.branches !== values.branches ||
       projectResource.pom !== values.pom
     ) {
-      buildState.value = BuildStateEnum.NEED_REBUILD;
+      buildStateEnum.value = BuildStateEnum.NEED_REBUILD;
     } else {
-      buildState.value = projectResource.buildState;
+      buildStateEnum.value = projectResource.buildStateEnum;
     }
   }
   /* Update project */
@@ -60,7 +60,7 @@
       const { data } = await updateProject({
         id: route.query.id,
         name: values.name,
-        gitCredential: values.gitCredential,
+        gitCredentialEnum: values.gitCredentialEnum,
         url: values.url,
         repository: values.repository,
         type: values.type,
@@ -71,7 +71,7 @@
         pom: values.pom,
         buildArgs: values.buildArgs,
         description: values.description,
-        buildState: buildState.value,
+        buildStateEnum: buildStateEnum.value,
       });
       if (data.data) {
         router.go(-1);
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/project/View.vue b/streampark-console/streampark-console-webapp/src/views/flink/project/View.vue
index 6c2ebc2e9..337771bc4 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/project/View.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/project/View.vue
@@ -18,7 +18,7 @@
   <PageWrapper contentFullHeight contentBackground contentClass="px-20px">
     <a-card class="header" :bordered="false">
       <template #extra>
-        <a-radio-group v-model:value="queryParams.buildState">
+        <a-radio-group v-model:value="queryParams.buildStateEnum">
           <a-radio-button
             v-for="item in buttonList"
             @click="handleQuery(item.key)"
@@ -106,7 +106,7 @@
       const [registerLogModal, { openModal: openLogModal }] = useModal();
       const buttonList = reactive(statusList);
       const loading = ref(false);
-      const buildState = ref('');
+      const buildStateEnum = ref('');
       const searchValue = ref('');
       const pageInfo = reactive({
         currentPage: 1,
@@ -114,8 +114,8 @@
         total: 0,
       });
 
-      const queryParams = reactive<{ buildState: string; name?: string }>({
-        buildState: '',
+      const queryParams = reactive<{ buildStateEnum: string; name?: string }>({
+        buildStateEnum: '',
       });
 
       let projectDataSource = ref<Array<ProjectRecord>>([]);
@@ -146,7 +146,7 @@
 
       const handleQuery = function (val: string | undefined) {
         pageInfo.currentPage = 1;
-        queryParams.buildState = val!;
+        queryParams.buildStateEnum = val!;
         queryParams.name = searchValue.value;
         queryData();
       };
@@ -189,7 +189,7 @@
         t,
         searchValue,
         pageInfo,
-        buildState,
+        buildStateEnum,
         buttonList,
         handleQuery,
         queryParams,
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/project/components/ListItem.vue b/streampark-console/streampark-console-webapp/src/views/flink/project/components/ListItem.vue
index 0368c992f..e56e42a77 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/project/components/ListItem.vue
+++ b/streampark-console/streampark-console-webapp/src/views/flink/project/components/ListItem.vue
@@ -58,9 +58,9 @@
         <p>{{ item.lastBuild || '--' }}</p>
       </li>
       <li class="list-content_item build_state">
-        <span>{{ t('flink.project.form.buildState') }}</span>
+        <span>{{ t('flink.project.form.buildStateEnum') }}</span>
         <p>
-          <a-tag :color="buildState.color" :class="tagClass">{{ buildState.label }}</a-tag>
+          <a-tag :color="buildStateEnum.color" :class="tagClass">{{ buildStateEnum.label }}</a-tag>
         </p>
       </li>
     </ul>
@@ -139,12 +139,12 @@
   const props = defineProps({
     item: { type: Object as PropType<ProjectRecord>, required: true },
   });
-  const needBuild = computed(() => props.item.buildState == BuildStateEnum.NEED_REBUILD);
-  const isBuilding = computed(() => props.item.buildState == BuildStateEnum.BUILDING);
-  const buildState = computed(() => {
-    return buildStateMap[props.item.buildState] || buildStateMap[BuildStateEnum.FAILED];
+  const needBuild = computed(() => props.item.buildStateEnum == BuildStateEnum.NEED_REBUILD);
+  const isBuilding = computed(() => props.item.buildStateEnum == BuildStateEnum.BUILDING);
+  const buildStateEnum = computed(() => {
+    return buildStateMap[props.item.buildStateEnum] || buildStateMap[BuildStateEnum.FAILED];
   });
-  const tagClass = computed(() => buildState.value.className || '');
+  const tagClass = computed(() => buildStateEnum.value.className || '');
   const svgName = computed(() => {
     return (
       {
diff --git a/streampark-console/streampark-console-webapp/src/views/flink/project/components/RepositoryGroup.tsx b/streampark-console/streampark-console-webapp/src/views/flink/project/components/RepositoryGroup.tsx
index 5f317752f..c549a3bbb 100644
--- a/streampark-console/streampark-console-webapp/src/views/flink/project/components/RepositoryGroup.tsx
+++ b/streampark-console/streampark-console-webapp/src/views/flink/project/components/RepositoryGroup.tsx
@@ -18,7 +18,7 @@ import { defineComponent, h } from 'vue';
 import type { PropType } from 'vue';
 import { Form, Input, Select, Tag } from 'ant-design-vue';
 export interface RepositoryProps {
-  gitCredential: string | number;
+  gitCredentialEnum: string | number;
   url: string;
 }
 import { useI18n } from '/@/hooks/web/useI18n';
@@ -61,10 +61,10 @@ export default defineComponent({
         <div>
           <Input.Group compact class="!flex custom-input-group">
             <Select
-              name="gitCredential"
+              name="gitCredentialEnum"
               style="width: 80px"
               placeholder={t('flink.project.form.gitCredentialPlaceholder')}
-              value={props.value?.gitCredential}
+              value={props.value?.gitCredentialEnum}
               onChange={(e: any) => handleProtocolChange(e)}
               options={options}
             />
@@ -82,10 +82,10 @@ export default defineComponent({
               <Tag color="#2db7f5" class="tag-note">
                 {t('flink.app.noteInfo.note')}
               </Tag>
-              {props.value?.gitCredential === 1 && (
+              {props.value?.gitCredentialEnum === 1 && (
                 <span>{t('flink.project.operationTips.httpsCredential')}</span>
               )}
-              {props.value?.gitCredential === 2 && (
+              {props.value?.gitCredentialEnum === 2 && (
                 <span>{t('flink.project.operationTips.sshCredential')}</span>
               )}
             </span>
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertModal.vue b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertModal.vue
index 422f8df56..963bf557d 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertModal.vue
+++ b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertModal.vue
@@ -22,7 +22,7 @@
 <script setup lang="ts" name="AlertModal">
   import { ref, defineComponent, h } from 'vue';
   import { omit } from 'lodash-es';
-  import { alertFormSchema, alertTypes } from '../index.data';
+  import { alertFormSchema, alertTypeEnums } from '../index.data';
   import { BasicModal, useModalInner } from '/@/components/Modal';
   import { BasicForm, useForm } from '/@/components/Form';
   import { Form, Select, Input, Divider } from 'ant-design-vue';
@@ -38,7 +38,7 @@
   const { t } = useI18n();
   const emit = defineEmits(['reload', 'register']);
   const alertId = ref<string | null>(null);
-  const alertType = ref<string[]>([]);
+  const alertTypeEnum = ref<string[]>([]);
 
   const { Swal } = useMessage();
   const userStore = useUserStore();
@@ -99,10 +99,10 @@
   const [registerModal, { changeOkLoading, closeModal }] = useModalInner((data) => {
     resetFields();
     alertId.value = '';
-    alertType.value = [];
+    alertTypeEnum.value = [];
     if (data && Object.keys(data).length > 0) {
       alertId.value = data.alertId;
-      alertType.value = data.alertType;
+      alertTypeEnum.value = data.alertTypeEnum;
       setFieldsValue(omit(data, 'alertId'));
     }
   });
@@ -116,7 +116,7 @@
         id: alertId.value,
         alertName: formValue.alertName,
         userId: userStore.getUserInfo?.userId,
-        alertType: eval(formValue.alertType.join('+')),
+        alertTypeEnum: eval(formValue.alertTypeEnum.join('+')),
         emailParams: { contacts: formValue.alertEmail },
         dingTalkParams: {
           token: formValue.dingtalkToken,
@@ -209,10 +209,10 @@
           :placeholder="t('setting.alarm.faultAlertType')"
           allowClear
           mode="multiple"
-          @change="(value: string[]) => (alertType = value)"
+          @change="(value: string[]) => (alertTypeEnum = value)"
         >
           <SelectOption
-            v-for="(v, k) in alertTypes"
+            v-for="(v, k) in alertTypeEnums"
             :key="`alertType_${k}`"
             :disabled="v.disabled"
             :value="k"
@@ -224,7 +224,7 @@
       </template>
       <template #alertEmail="{ model, field }">
         <!-- Alert Email -->
-        <template v-if="(alertType || []).includes('1')">
+        <template v-if="(alertTypeEnum || []).includes('1')">
           <Divider>
             <SvgIcon name="mail" size="20" />
             {{ t('setting.alarm.email') }}
@@ -254,7 +254,7 @@
         </template>
       </template>
 
-      <template #alertDingURL="{ model, field }" v-if="(alertType || []).includes('2')">
+      <template #alertDingURL="{ model, field }" v-if="(alertTypeEnum || []).includes('2')">
         <Divider>
           <SvgIcon name="dingtalk" size="20" />
           {{ t('setting.alarm.dingTalk') }}
@@ -280,14 +280,14 @@
       </template>
 
       <!-- WeChat -->
-      <template #weToken="{ model, field, schema }" v-if="(alertType || []).includes('4')">
+      <template #weToken="{ model, field, schema }" v-if="(alertTypeEnum || []).includes('4')">
         <Divider> <SvgIcon name="wecom" size="20" /> {{ t('setting.alarm.weChat') }} </Divider>
         <FormItem :label="schema.label" :name="field" :rules="schema.rules">
           <InputTextArea v-model:value="model[field]" v-bind="schema.componentProps" />
         </FormItem>
       </template>
 
-      <template #alertSms="{ model, field, schema }" v-if="(alertType || []).includes('8')">
+      <template #alertSms="{ model, field, schema }" v-if="(alertTypeEnum || []).includes('8')">
         <Divider> <SvgIcon name="message" size="20" /> {{ t('setting.alarm.sms') }} </Divider>
         <FormItem :label="schema.label" :name="field" :rules="schema.rules">
           <Input v-model:value="model[field]" v-bind="schema.componentProps" />
@@ -295,7 +295,7 @@
       </template>
 
       <!-- lark -->
-      <template #larkToken="{ model, field, schema }" v-if="(alertType || []).includes('16')">
+      <template #larkToken="{ model, field, schema }" v-if="(alertTypeEnum || []).includes('16')">
         <Divider> <SvgIcon name="lark" size="20" /> {{ t('setting.alarm.lark') }} </Divider>
         <FormItem :label="schema.label" :name="field" :rules="schema.rules">
           <Input
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertTypeInfo.vue b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertTypeInfo.vue
index c76f9cc38..fb80346bb 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertTypeInfo.vue
+++ b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/components/AlertTypeInfo.vue
@@ -19,12 +19,12 @@
   import { Descriptions, Tag } from 'ant-design-vue';
   import { computed, toRefs } from 'vue';
   import { SvgIcon } from '/@/components/Icon';
-  import { alertTypes } from '../index.data';
+  import { alertTypeEnums } from '../index.data';
   import { useI18n } from '/@/hooks/web/useI18n';
 
   const { t } = useI18n();
   const props = defineProps({
-    alertType: {
+    alertTypeEnum: {
       type: String,
       validator: (v: string) => ['1', '2', '4', '8', '16'].includes(v),
       required: true,
@@ -34,7 +34,7 @@
       required: true,
     },
   });
-  const { alertType, alertSource } = toRefs(props);
+  const { alertTypeEnum, alertSource } = toRefs(props);
   const DescriptionsItem = Descriptions.Item;
 
   const emailInfo = computed(() => {
@@ -57,19 +57,19 @@
 <template>
   <div class="mt-10px border-dot text-16px cursor-pointer flex items-center">
     <div class="flex items-center">
-      <SvgIcon :name="alertTypes[alertType].icon" :size="20" class="!align-middle" />
+      <SvgIcon :name="alertTypeEnums[alertTypeEnum].icon" :size="20" class="!align-middle" />
       <span class="pl-10px">
-        {{ alertTypes[alertType].name }}
+        {{ alertTypeEnums[alertTypeEnum].name }}
       </span>
     </div>
   </div>
   <Descriptions size="small" :column="1" class="pl-15px mt-10px">
-    <template v-if="alertType === '1'">
+    <template v-if="alertTypeEnum === '1'">
       <DescriptionsItem :label="t('setting.alarm.alertEmail')">
         <span class="text-blue-500">{{ emailInfo.contacts || '' }}</span>
       </DescriptionsItem>
     </template>
-    <template v-else-if="alertType === '2'">
+    <template v-else-if="alertTypeEnum === '2'">
       <DescriptionsItem :label="t('setting.alarm.dingTalkUser')">
         {{ dingTalk.contacts || '' }}
       </DescriptionsItem>
@@ -79,12 +79,12 @@
         </Tag>
       </DescriptionsItem>
     </template>
-    <template v-else-if="alertType === '4'">
+    <template v-else-if="alertTypeEnum === '4'">
       <DescriptionsItem :label="t('setting.alarm.weChattoken')">
         {{ desensitization(weChat.token || '') }}
       </DescriptionsItem>
     </template>
-    <template v-else-if="alertType === '16'">
+    <template v-else-if="alertTypeEnum === '16'">
       <DescriptionsItem :label="t('setting.alarm.larkIsAtAll')">
         <Tag :color="lark.isAtAll ? 'green' : 'red'" class="!leading-20px">
           {{ lark.isAtAll }}
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.data.ts b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.data.ts
index 0becf9887..dcebfffb7 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.data.ts
+++ b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.data.ts
@@ -175,7 +175,7 @@ export const alertFormSchema: Array<FormSchema> = [
   },
 ];
 
-export const alertTypes = {
+export const alertTypeEnums = {
   '1': { name: t('setting.alarm.email'), value: 1, disabled: false, icon: 'mail' },
   '2': { name: t('setting.alarm.dingTalk'), value: 2, disabled: false, icon: 'dingtalk' },
   '4': { name: t('setting.alarm.weChat'), value: 4, disabled: false, icon: 'wecom' },
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.vue b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.vue
index 74f995316..2660393cc 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.vue
+++ b/streampark-console/streampark-console-webapp/src/views/setting/Alarm/index.vue
@@ -50,7 +50,7 @@
   /* Get alert configuration */
   async function getAlertSetting() {
     const res = await fetchAlertSetting();
-    res.map((a) => (a.alertTypeTags = computeAlertType(a.alertType)));
+    res.map((a) => (a.alertTypeTags = computeAlertType(a.alertTypeEnum)));
     alerts.value = res;
   }
   const alertTypeMap = {
@@ -117,7 +117,7 @@
     openAlertModal(true, {
       alertId: item.id,
       alertName: item.alertName,
-      alertType: item.alertTypeTags,
+      alertTypeEnum: item.alertTypeTags,
       alertEmail: emailParams.contacts,
       alertDingURL: dingTalkParams.alertDingURL,
       dingtalkToken: dingTalkParams.token,
@@ -261,27 +261,27 @@
             </template>
 
             <AlertTypeInfo
-              :alertType="String(AlertTypeEnum.MAIL)"
+              :alertTypeEnum="String(AlertTypeEnum.MAIL)"
               :alertSource="item"
               v-if="item.alertTypeTags.includes(String(AlertTypeEnum.MAIL))"
             />
             <AlertTypeInfo
-              :alertType="String(AlertTypeEnum.WECOM)"
+              :alertTypeEnum="String(AlertTypeEnum.WECOM)"
               :alertSource="item"
               v-if="item.alertTypeTags.includes(String(AlertTypeEnum.WECOM))"
             />
             <AlertTypeInfo
-              :alertType="String(AlertTypeEnum.DINGTALK)"
+              :alertTypeEnum="String(AlertTypeEnum.DINGTALK)"
               :alertSource="item"
               v-if="item.alertTypeTags.includes(String(AlertTypeEnum.DINGTALK))"
             />
             <AlertTypeInfo
-              :alertType="String(AlertTypeEnum.MESSAGE)"
+              :alertTypeEnum="String(AlertTypeEnum.MESSAGE)"
               :alertSource="item"
               v-if="item.alertTypeTags.includes(String(AlertTypeEnum.MESSAGE))"
             />
             <AlertTypeInfo
-              :alertType="String(AlertTypeEnum.LARK)"
+              :alertTypeEnum="String(AlertTypeEnum.LARK)"
               :alertSource="item"
               v-if="item.alertTypeTags.includes(String(AlertTypeEnum.LARK))"
             />
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/EditCluster.vue b/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/EditCluster.vue
index 9c2590226..817351228 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/EditCluster.vue
+++ b/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/EditCluster.vue
@@ -107,11 +107,11 @@
       setFieldsValue({
         clusterName: cluster.clusterName,
         clusterId: cluster.clusterId,
-        executionMode: cluster.executionMode,
+        executionModeEnum: cluster.executionModeEnum,
         address: cluster.address,
         description: cluster.description,
         dynamicProperties: cluster.dynamicProperties,
-        resolveOrder: cluster.resolveOrder,
+        resolveOrderEnum: cluster.resolveOrderEnum,
         yarnQueue: cluster.yarnQueue,
         alertId: selectAlertId,
         versionId: cluster.versionId || null,
diff --git a/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/index.vue b/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/index.vue
index 4b330ca73..452e54a95 100644
--- a/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/index.vue
+++ b/streampark-console/streampark-console-webapp/src/views/setting/FlinkCluster/index.vue
@@ -59,7 +59,7 @@
   const clusters = ref<FlinkCluster[]>([]);
   const loading = ref(false);
   function handleIsStart(item) {
-    return item.clusterState === ClusterStateEnum.RUNNING;
+    return item.clusterStateEnum === ClusterStateEnum.RUNNING;
   }
 
   /* Go to edit cluster */
@@ -158,7 +158,7 @@
           </ListItemMeta>
           <div class="list-content" style="width: 10%">
             <div class="list-content-item">
-              <span>{{ t('setting.flinkCluster.form.executionMode') }}</span>
+              <span>{{ t('setting.flinkCluster.form.executionModeEnum') }}</span>
               <p style="margin-top: 10px">
                 {{ item.executionModeEnum.toLowerCase() }}
               </p>
@@ -168,8 +168,8 @@
             class="list-content"
             style="width: 40%"
             v-if="
-              item.executionMode === ExecModeEnum.REMOTE ||
-              item.executionMode === ExecModeEnum.YARN_SESSION
+              item.executionModeEnum === ExecModeEnum.REMOTE ||
+              item.executionModeEnum === ExecModeEnum.YARN_SESSION
             "
           >
             <div class="list-content-item">
@@ -185,14 +185,14 @@
             class="list-content"
             style="width: 10%"
             v-if="
-              item.executionMode === ExecModeEnum.REMOTE ||
-              item.executionMode === ExecModeEnum.YARN_SESSION
+              item.executionModeEnum === ExecModeEnum.REMOTE ||
+              item.executionModeEnum === ExecModeEnum.YARN_SESSION
             "
           >
             <div class="list-content-item">
               <span>{{ t('setting.flinkCluster.form.runState') }}</span>
               <p style="margin-top: 10px">
-                <State :data="{ clusterState: item.clusterState }" />
+                <State :data="{ clusterStateEnum: item.clusterStateEnum }" />
               </p>
             </div>
           </div>
@@ -212,7 +212,7 @@
             <template v-if="handleIsStart(item)">
               <Tooltip :title="t('setting.flinkCluster.stop')">
                 <a-button
-                  :disabled="item.executionMode === ExecModeEnum.REMOTE"
+                  :disabled="item.executionModeEnum === ExecModeEnum.REMOTE"
                   v-auth="'cluster:create'"
                   @click="handleShutdownCluster(item)"
                   shape="circle"
@@ -227,7 +227,7 @@
             <template v-else>
               <Tooltip :title="t('setting.flinkCluster.start')">
                 <a-button
-                  :disabled="item.executionMode === ExecModeEnum.REMOTE"
+                  :disabled="item.executionModeEnum === ExecModeEnum.REMOTE"
                   v-auth="'cluster:create'"
                   @click="handleDeployCluster(item)"
                   shape="circle"
diff --git a/streampark-console/streampark-console-webapp/src/views/system/user/User.vue b/streampark-console/streampark-console-webapp/src/views/system/user/User.vue
index b04cac08a..756c9b0b1 100644
--- a/streampark-console/streampark-console-webapp/src/views/system/user/User.vue
+++ b/streampark-console/streampark-console-webapp/src/views/system/user/User.vue
@@ -107,7 +107,7 @@
             tooltip: t('system.user.table.reset'),
             ifShow: () =>
               (record.username !== 'admin' || userName.value === 'admin') &&
-              record.loginType == LoginTypeEnum[LoginTypeEnum.PASSWORD],
+              record.loginTypeEnum == LoginTypeEnum[LoginTypeEnum.PASSWORD],
             popConfirm: {
               title: t('system.user.table.resetTip'),
               confirm: handleReset.bind(null, record),
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
index f06db3d9d..d9bf7f633 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/CancelRequest.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.{FlinkVersion, K8sFlinkConfig}
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 
 import javax.annotation.Nullable
 
@@ -27,7 +27,7 @@ import java.util.{Map => JavaMap}
 case class CancelRequest(
     id: Long,
     flinkVersion: FlinkVersion,
-    executionMode: ExecutionMode,
+    executionMode: ExecutionModeEnum,
     @Nullable properties: JavaMap[String, Any],
     clusterId: String,
     jobId: String,
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
index db7bbcd3d..e2ba9fc6c 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/DeployRequest.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.{FlinkVersion, Workspace}
-import org.apache.streampark.common.enums.{ExecutionMode, FlinkK8sRestExposedType}
+import org.apache.streampark.common.enums.{ExecutionModeEnum, FlinkK8sRestExposedTypeEnum}
 import org.apache.streampark.flink.util.FlinkUtils
 
 import org.apache.commons.io.FileUtils
@@ -31,7 +31,7 @@ import java.util.{Map => JavaMap}
 
 case class DeployRequest(
     flinkVersion: FlinkVersion,
-    executionMode: ExecutionMode,
+    executionMode: ExecutionModeEnum,
     properties: JavaMap[String, Any],
     clusterId: String,
     @Nullable k8sDeployParam: KubernetesDeployParam) {
@@ -69,4 +69,5 @@ case class KubernetesDeployParam(
     kubeConf: String = "~/.kube/config",
     serviceAccount: String = KubernetesConfigOptions.KUBERNETES_SERVICE_ACCOUNT.defaultValue(),
     flinkImage: String = KubernetesConfigOptions.CONTAINER_IMAGE.defaultValue(),
-    @Nullable flinkRestExposedType: FlinkK8sRestExposedType = FlinkK8sRestExposedType.CLUSTER_IP)
+    @Nullable flinkRestExposedType: FlinkK8sRestExposedTypeEnum =
+      FlinkK8sRestExposedTypeEnum.CLUSTER_IP)
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/KubernetesSubmitParam.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/KubernetesSubmitParam.scala
index e3351200f..8671d88bd 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/KubernetesSubmitParam.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/KubernetesSubmitParam.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.client.bean
 
-import org.apache.streampark.common.enums.FlinkK8sRestExposedType
+import org.apache.streampark.common.enums.FlinkK8sRestExposedTypeEnum
 import org.apache.streampark.flink.kubernetes.v2.model.IngressDef
 
 import javax.annotation.Nullable
@@ -51,7 +51,7 @@ case class KubernetesSubmitParam(
     taskManagerEphemeralStorage: Option[String] = None,
     taskManagerPodTemplate: Option[String] = None,
     logConfiguration: JMap[String, String] = new util.HashMap[String, String](),
-    flinkRestExposedType: Option[FlinkK8sRestExposedType] = None,
+    flinkRestExposedType: Option[FlinkK8sRestExposedTypeEnum] = None,
     ingressDefinition: Option[IngressDef] = None
 )
 
@@ -72,7 +72,7 @@ object KubernetesSubmitParam {
       kubernetesName: String,
       kubernetesNamespace: String,
       baseImage: String,
-      @Nullable flinkRestExposedType: FlinkK8sRestExposedType,
+      @Nullable flinkRestExposedType: FlinkK8sRestExposedTypeEnum,
       @Nullable ingressDefinition: IngressDef): KubernetesSubmitParam =
     KubernetesSubmitParam(
       clusterId = clusterId,
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
index 34bc21c61..3d2c5b542 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SavepointRequestTrait.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.{FlinkVersion, K8sFlinkConfig}
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 
 import javax.annotation.Nullable
 
@@ -28,7 +28,7 @@ trait SavepointRequestTrait {
 
   val flinkVersion: FlinkVersion
 
-  val executionMode: ExecutionMode
+  val executionMode: ExecutionModeEnum
 
   val clusterId: String
 
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
index a6b75bbdf..86c777a68 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/ShutDownRequest.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.FlinkVersion
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 
 import javax.annotation.Nullable
 
@@ -26,7 +26,7 @@ import java.util.{Map => JavaMap}
 
 case class ShutDownRequest(
     flinkVersion: FlinkVersion,
-    executionMode: ExecutionMode,
+    executionMode: ExecutionModeEnum,
     @Nullable properties: JavaMap[String, Any],
     clusterId: String,
     @Nullable kubernetesDeployParam: KubernetesDeployParam)
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
index 7650b84b3..ef440a54c 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/SubmitRequest.scala
@@ -38,17 +38,17 @@ import scala.util.Try
 
 case class SubmitRequest(
     flinkVersion: FlinkVersion,
-    executionMode: ExecutionMode,
+    executionMode: ExecutionModeEnum,
     properties: JavaMap[String, Any],
     flinkYaml: String,
-    developmentMode: DevelopmentMode,
+    developmentMode: DevelopmentModeEnum,
     id: Long,
     jobId: String,
     appName: String,
     appConf: String,
-    applicationType: ApplicationType,
+    applicationType: ApplicationTypeEnum,
     savePoint: String,
-    restoreMode: RestoreMode,
+    restoreMode: RestoreModeEnum,
     args: String,
     @Nullable buildResult: BuildResult,
     @Nullable k8sSubmitParam: KubernetesSubmitParam,
@@ -59,8 +59,8 @@ case class SubmitRequest(
   lazy val appOption: Map[String, String] = getParameterMap(KEY_FLINK_OPTION_PREFIX)
 
   lazy val appMain: String = this.developmentMode match {
-    case DevelopmentMode.FLINK_SQL => ConfigConst.STREAMPARK_FLINKSQL_CLIENT_CLASS
-    case DevelopmentMode.PYFLINK => ConfigConst.PYTHON_DRIVER_CLASS_NAME
+    case DevelopmentModeEnum.FLINK_SQL => ConfigConst.STREAMPARK_FLINKSQL_CLIENT_CLASS
+    case DevelopmentModeEnum.PYFLINK => ConfigConst.PYTHON_DRIVER_CLASS_NAME
     case _ => appProperties(KEY_FLINK_APPLICATION_MAIN_CLASS)
   }
 
@@ -82,7 +82,7 @@ case class SubmitRequest(
 
   lazy val userJarFile: File = {
     executionMode match {
-      case ExecutionMode.KUBERNETES_NATIVE_APPLICATION => null
+      case ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION => null
       case _ =>
         checkBuildResult()
         new File(buildResult.asInstanceOf[ShadedBuildResponse].shadedJarPath)
@@ -169,7 +169,7 @@ case class SubmitRequest(
   @throws[Exception]
   def checkBuildResult(): Unit = {
     executionMode match {
-      case ExecutionMode.KUBERNETES_NATIVE_SESSION =>
+      case ExecutionModeEnum.KUBERNETES_NATIVE_SESSION =>
         if (buildResult == null) {
           throw new Exception(
             s"[flink-submit] current job: ${this.effectiveAppName} was not yet built, buildResult is empty" +
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
index 85c788da7..4c4acb9f7 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-api/src/main/scala/org/apache/streampark/flink/client/bean/TriggerSavepointRequest.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.bean
 
 import org.apache.streampark.common.conf.{FlinkVersion, K8sFlinkConfig}
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 
 import javax.annotation.Nullable
 
@@ -28,7 +28,7 @@ import java.util.{Map => JavaMap}
 case class TriggerSavepointRequest(
     id: Long,
     flinkVersion: FlinkVersion,
-    executionMode: ExecutionMode,
+    executionMode: ExecutionModeEnum,
     @Nullable properties: JavaMap[String, Any],
     clusterId: String,
     jobId: String,
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/FlinkClientEndpoint.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/FlinkClientEndpoint.scala
index c36c2d1bf..58a0da06b 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/FlinkClientEndpoint.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/FlinkClientEndpoint.scala
@@ -18,15 +18,15 @@
 package org.apache.streampark.flink.client
 
 import org.apache.streampark.common.conf.K8sFlinkConfig
-import org.apache.streampark.common.enums.ExecutionMode
-import org.apache.streampark.common.enums.ExecutionMode._
+import org.apache.streampark.common.enums.ExecutionModeEnum
+import org.apache.streampark.common.enums.ExecutionModeEnum._
 import org.apache.streampark.flink.client.`trait`.FlinkClientTrait
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.client.impl._
 
 object FlinkClientEndpoint {
 
-  private[this] val clients: Map[ExecutionMode, FlinkClientTrait] = Map(
+  private[this] val clients: Map[ExecutionModeEnum, FlinkClientTrait] = Map(
     LOCAL -> LocalClient,
     REMOTE -> RemoteClient,
     YARN_APPLICATION -> YarnApplicationClient,
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
index 209a21045..d08be791c 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeApplicationClient.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.client.impl
 
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 import org.apache.streampark.common.util.Utils
 import org.apache.streampark.flink.client.`trait`.KubernetesNativeClientTrait
 import org.apache.streampark.flink.client.bean._
@@ -94,7 +94,7 @@ object KubernetesNativeApplicationClient extends KubernetesNativeClientTrait {
       flinkConfig: Configuration): CancelResponse = {
     flinkConfig.safeSet(
       DeploymentOptions.TARGET,
-      ExecutionMode.KUBERNETES_NATIVE_APPLICATION.getName)
+      ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION.getName)
     super.doCancel(cancelRequest, flinkConfig)
   }
 
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
index a63e91298..23ebb93d2 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/KubernetesNativeSessionClient.scala
@@ -17,14 +17,14 @@
 
 package org.apache.streampark.flink.client.impl
 
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 import org.apache.streampark.common.util.{Logger, Utils}
 import org.apache.streampark.flink.client.`trait`.KubernetesNativeClientTrait
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.client.tool.FlinkSessionSubmitHelper
 import org.apache.streampark.flink.core.FlinkKubernetesClient
 import org.apache.streampark.flink.kubernetes.KubernetesRetriever
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum
 import org.apache.streampark.flink.kubernetes.model.ClusterKey
 
 import io.fabric8.kubernetes.api.model.{Config => _}
@@ -62,7 +62,7 @@ object KubernetesNativeSessionClient extends KubernetesNativeClientTrait with Lo
     Try {
       // get jm rest url of flink session cluster
       val clusterKey = ClusterKey(
-        FlinkK8sExecuteMode.SESSION,
+        FlinkK8sExecuteModeEnum.SESSION,
         submitRequest.k8sSubmitParam.kubernetesNamespace,
         submitRequest.k8sSubmitParam.clusterId)
       val jmRestUrl = KubernetesRetriever
@@ -122,7 +122,9 @@ object KubernetesNativeSessionClient extends KubernetesNativeClientTrait with Lo
   override def doCancel(
       cancelRequest: CancelRequest,
       flinkConfig: Configuration): CancelResponse = {
-    flinkConfig.safeSet(DeploymentOptions.TARGET, ExecutionMode.KUBERNETES_NATIVE_SESSION.getName)
+    flinkConfig.safeSet(
+      DeploymentOptions.TARGET,
+      ExecutionModeEnum.KUBERNETES_NATIVE_SESSION.getName)
     super.doCancel(cancelRequest, flinkConfig)
   }
 
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
index b585a1df8..52441fd82 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/impl/YarnApplicationClient.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.client.impl
 
 import org.apache.streampark.common.conf.{ConfigConst, Workspace}
-import org.apache.streampark.common.enums.DevelopmentMode
+import org.apache.streampark.common.enums.DevelopmentModeEnum
 import org.apache.streampark.common.fs.FsOperator
 import org.apache.streampark.common.util.{FileUtils, HdfsUtils, Utils}
 import org.apache.streampark.flink.client.`trait`.YarnClientTrait
@@ -69,7 +69,7 @@ object YarnApplicationClient extends YarnClientTrait {
         submitRequest.hdfsWorkspace.appPlugins
       )
       submitRequest.developmentMode match {
-        case DevelopmentMode.FLINK_SQL =>
+        case DevelopmentModeEnum.FLINK_SQL =>
           array += s"${workspace.APP_SHIMS}/flink-${submitRequest.flinkVersion.majorVersion}"
           val jobLib = s"${workspace.APP_WORKSPACE}/${submitRequest.id}/lib"
           if (HdfsUtils.exists(jobLib)) {
@@ -95,7 +95,7 @@ object YarnApplicationClient extends YarnClientTrait {
       // yarn application Type
       .safeSet(YarnConfigOptions.APPLICATION_TYPE, submitRequest.applicationType.getName)
 
-    if (submitRequest.developmentMode == DevelopmentMode.PYFLINK) {
+    if (submitRequest.developmentMode == DevelopmentModeEnum.PYFLINK) {
       val pyVenv: String = workspace.APP_PYTHON_VENV
       if (!FsOperator.hdfs.exists(pyVenv)) {
         throw new RuntimeException(s"$pyVenv File does not exist")
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
index 839520eba..ee706f9d1 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/FlinkClientTrait.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.flink.client.`trait`
 
 import org.apache.streampark.common.conf.{ConfigConst, Workspace}
 import org.apache.streampark.common.conf.ConfigConst._
-import org.apache.streampark.common.enums.{ApplicationType, DevelopmentMode, ExecutionMode, RestoreMode}
+import org.apache.streampark.common.enums.{ApplicationTypeEnum, DevelopmentModeEnum, ExecutionModeEnum, RestoreModeEnum}
 import org.apache.streampark.common.fs.FsOperator
 import org.apache.streampark.common.util.{DeflaterUtils, FileUtils, Logger, SystemPropertyUtils}
 import org.apache.streampark.flink.client.bean._
@@ -93,7 +93,7 @@ trait FlinkClientTrait extends Logger {
     val (commandLine, flinkConfig) = getCommandLineAndFlinkConfig(submitRequest)
 
     submitRequest.developmentMode match {
-      case DevelopmentMode.PYFLINK =>
+      case DevelopmentModeEnum.PYFLINK =>
         val flinkOptPath: String = System.getenv(ConfigConstants.ENV_FLINK_OPT_DIR)
         if (StringUtils.isBlank(flinkOptPath)) {
           logWarn(s"Get environment variable ${ConfigConstants.ENV_FLINK_OPT_DIR} fail")
@@ -140,9 +140,9 @@ trait FlinkClientTrait extends Logger {
         submitRequest.allowNonRestoredState)
       if (
         submitRequest.flinkVersion.checkVersion(
-          RestoreMode.SINCE_FLINK_VERSION) && submitRequest.restoreMode != null
+          RestoreModeEnum.SINCE_FLINK_VERSION) && submitRequest.restoreMode != null
       ) {
-        flinkConfig.setString(RestoreMode.RESTORE_MODE, submitRequest.restoreMode.getName);
+        flinkConfig.setString(RestoreModeEnum.RESTORE_MODE, submitRequest.restoreMode.getName);
       }
     }
 
@@ -236,7 +236,7 @@ trait FlinkClientTrait extends Logger {
   private[client] def getJobGraph(
       submitRequest: SubmitRequest,
       flinkConfig: Configuration): (PackagedProgram, JobGraph) = {
-    if (submitRequest.developmentMode == DevelopmentMode.PYFLINK) {
+    if (submitRequest.developmentMode == DevelopmentModeEnum.PYFLINK) {
       val pythonVenv: String = Workspace.local.APP_PYTHON_VENV
       if (!FsOperator.lfs.exists(pythonVenv)) {
         throw new RuntimeException(s"$pythonVenv File does not exist")
@@ -487,14 +487,14 @@ trait FlinkClientTrait extends Logger {
       }
     }
 
-    if (submitRequest.applicationType == ApplicationType.STREAMPARK_FLINK) {
+    if (submitRequest.applicationType == ApplicationTypeEnum.STREAMPARK_FLINK) {
 
       programArgs += PARAM_KEY_FLINK_CONF += submitRequest.flinkYaml
       programArgs += PARAM_KEY_APP_NAME += DeflaterUtils.zipString(submitRequest.effectiveAppName)
       programArgs += PARAM_KEY_FLINK_PARALLELISM += getParallelism(submitRequest).toString
 
       submitRequest.developmentMode match {
-        case DevelopmentMode.FLINK_SQL =>
+        case DevelopmentModeEnum.FLINK_SQL =>
           programArgs += PARAM_KEY_FLINK_SQL += submitRequest.flinkSQL
           if (submitRequest.appConf != null) {
             programArgs += PARAM_KEY_APP_CONF += submitRequest.appConf
@@ -514,8 +514,8 @@ trait FlinkClientTrait extends Logger {
     }
 
     if (
-      submitRequest.developmentMode == DevelopmentMode.PYFLINK
-      && submitRequest.executionMode != ExecutionMode.YARN_APPLICATION
+      submitRequest.developmentMode == DevelopmentModeEnum.PYFLINK
+      && submitRequest.executionMode != ExecutionModeEnum.YARN_APPLICATION
     ) {
       // python file
       programArgs.add("-py")
@@ -601,7 +601,7 @@ trait FlinkClientTrait extends Logger {
             .key(CheckpointingOptions.SAVEPOINT_DIRECTORY.key())
             .stringType()
             .defaultValue {
-              if (request.executionMode == ExecutionMode.YARN_APPLICATION) {
+              if (request.executionMode == ExecutionModeEnum.YARN_APPLICATION) {
                 Workspace.remote.APP_SAVEPOINTS
               } else null
             }
diff --git a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/KubernetesNativeClientTrait.scala b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/KubernetesNativeClientTrait.scala
index 310d42a14..1ce9a8304 100644
--- a/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/KubernetesNativeClientTrait.scala
+++ b/streampark-flink/streampark-flink-client/streampark-flink-client-core/src/main/scala/org/apache/streampark/flink/client/trait/KubernetesNativeClientTrait.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.client.`trait`
 
-import org.apache.streampark.common.enums.{ExecutionMode, FlinkK8sRestExposedType}
+import org.apache.streampark.common.enums.{ExecutionModeEnum, FlinkK8sRestExposedTypeEnum}
 import org.apache.streampark.flink.client.bean._
 import org.apache.streampark.flink.kubernetes.PodTemplateTool
 import org.apache.streampark.flink.packer.pipeline.DockerImageBuildResponse
@@ -48,7 +48,7 @@ trait KubernetesNativeClientTrait extends FlinkClientTrait {
         covertToServiceExposedType(submitRequest.k8sSubmitParam.flinkRestExposedType.get))
 
     if (submitRequest.buildResult != null) {
-      if (submitRequest.executionMode == ExecutionMode.KUBERNETES_NATIVE_APPLICATION) {
+      if (submitRequest.executionMode == ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION) {
         val buildResult = submitRequest.buildResult.asInstanceOf[DockerImageBuildResponse]
         buildResult.podTemplatePaths.foreach(
           p => {
@@ -170,11 +170,12 @@ trait KubernetesNativeClientTrait extends FlinkClientTrait {
         KubernetesConfigOptions.CLUSTER_ID)}, " +
       s"namespace=${conf.get(KubernetesConfigOptions.NAMESPACE)}"
 
-  private def covertToServiceExposedType(exposedType: FlinkK8sRestExposedType): ServiceExposedType =
+  private def covertToServiceExposedType(
+      exposedType: FlinkK8sRestExposedTypeEnum): ServiceExposedType =
     exposedType match {
-      case FlinkK8sRestExposedType.CLUSTER_IP => ServiceExposedType.ClusterIP
-      case FlinkK8sRestExposedType.LOAD_BALANCER => ServiceExposedType.LoadBalancer
-      case FlinkK8sRestExposedType.NODE_PORT => ServiceExposedType.NodePort
+      case FlinkK8sRestExposedTypeEnum.CLUSTER_IP => ServiceExposedType.ClusterIP
+      case FlinkK8sRestExposedTypeEnum.LOAD_BALANCER => ServiceExposedType.LoadBalancer
+      case FlinkK8sRestExposedTypeEnum.NODE_PORT => ServiceExposedType.NodePort
       case _ => ServiceExposedType.LoadBalancer
     }
 
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/DefaultFlinkK8sWatcher.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/DefaultFlinkK8sWatcher.scala
index 21975d248..6e5a28f45 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/DefaultFlinkK8sWatcher.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/DefaultFlinkK8sWatcher.scala
@@ -18,8 +18,8 @@
 package org.apache.streampark.flink.kubernetes
 
 import org.apache.streampark.common.conf.K8sFlinkConfig
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode.{APPLICATION, SESSION}
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum.{APPLICATION, SESSION}
 import org.apache.streampark.flink.kubernetes.event.{BuildInEvent, FlinkJobStateEvent, FlinkJobStatusChangeEvent}
 import org.apache.streampark.flink.kubernetes.model._
 import org.apache.streampark.flink.kubernetes.watcher.{FlinkCheckpointWatcher, FlinkJobStatusWatcher, FlinkK8sEventWatcher, FlinkMetricWatcher, FlinkWatcher}
@@ -97,8 +97,8 @@ class DefaultFlinkK8sWatcher(conf: FlinkTrackConfig = FlinkTrackConfig.defaultCo
   override def checkIsInRemoteCluster(trackId: TrackId): Boolean = {
     if (!trackId.isLegal) false;
     else {
-      val nonLost = (state: FlinkJobState.Value) =>
-        state != FlinkJobState.LOST || state != FlinkJobState.SILENT
+      val nonLost = (state: FlinkJobStateEnum.Value) =>
+        state != FlinkJobStateEnum.LOST || state != FlinkJobStateEnum.SILENT
       trackId.executeMode match {
         case SESSION =>
           jobStatusWatcher.touchSessionJob(trackId).exists(e => nonLost(e.jobState))
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
index 7418e8c50..edd679ad7 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/KubernetesRetriever.scala
@@ -20,7 +20,7 @@ package org.apache.streampark.flink.kubernetes
 import org.apache.streampark.common.conf.ConfigConst
 import org.apache.streampark.common.util.ImplicitsUtils._
 import org.apache.streampark.common.util.Logger
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum
 import org.apache.streampark.flink.kubernetes.ingress.IngressController
 import org.apache.streampark.flink.kubernetes.model.ClusterKey
 
@@ -65,7 +65,7 @@ object KubernetesRetriever extends Logger {
   def newFinkClusterClient(
       clusterId: String,
       @Nullable namespace: String,
-      executeMode: FlinkK8sExecuteMode.Value): Option[ClusterClient[String]] = {
+      executeMode: FlinkK8sExecuteModeEnum.Value): Option[ClusterClient[String]] = {
     // build flink config
     val flinkConfig = new Configuration()
     flinkConfig.setString(DeploymentOptions.TARGET, executeMode.toString)
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobState.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobStateEnum.scala
similarity index 90%
rename from streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobState.scala
rename to streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobStateEnum.scala
index feefe1ad9..cc789a46d 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobState.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkJobStateEnum.scala
@@ -19,7 +19,7 @@ package org.apache.streampark.flink.kubernetes.enums
 
 /** flink job status on kubernetes */
 @Deprecated
-object FlinkJobState extends Enumeration {
+object FlinkJobStateEnum extends Enumeration {
 
   // flink job has been submit by the streampark.
   val STARTING,
@@ -44,11 +44,11 @@ object FlinkJobState extends Enumeration {
   // ending flink states, the tracking monitor will stop tracking these states of flink job.
   val endingStates = Seq(FAILED, CANCELED, FINISHED, POS_TERMINATED, TERMINATED, LOST)
 
-  def of(value: String): FlinkJobState.Value = {
+  def of(value: String): FlinkJobStateEnum.Value = {
     this.values.find(_.toString == value).getOrElse(OTHER)
   }
 
   // whether flink job state is ending state
-  def isEndState(state: FlinkJobState.Value): Boolean = endingStates.contains(state)
+  def isEndState(state: FlinkJobStateEnum.Value): Boolean = endingStates.contains(state)
 
 }
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteMode.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteModeEnum.scala
similarity index 60%
rename from streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteMode.scala
rename to streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteModeEnum.scala
index c1e1f945c..8d52a1434 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteMode.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/enums/FlinkK8sExecuteModeEnum.scala
@@ -17,26 +17,26 @@
 
 package org.apache.streampark.flink.kubernetes.enums
 
-import org.apache.streampark.common.enums.ExecutionMode
+import org.apache.streampark.common.enums.ExecutionModeEnum
 
 /** execution mode of flink on kubernetes */
-object FlinkK8sExecuteMode extends Enumeration {
+object FlinkK8sExecuteModeEnum extends Enumeration {
 
-  val SESSION: FlinkK8sExecuteMode.Value = Value("kubernetes-session")
-  val APPLICATION: FlinkK8sExecuteMode.Value = Value("kubernetes-application")
+  val SESSION: FlinkK8sExecuteModeEnum.Value = Value("kubernetes-session")
+  val APPLICATION: FlinkK8sExecuteModeEnum.Value = Value("kubernetes-application")
 
-  def of(mode: ExecutionMode): Value = {
+  def of(mode: ExecutionModeEnum): Value = {
     mode match {
-      case ExecutionMode.KUBERNETES_NATIVE_SESSION => SESSION
-      case ExecutionMode.KUBERNETES_NATIVE_APPLICATION => APPLICATION
+      case ExecutionModeEnum.KUBERNETES_NATIVE_SESSION => SESSION
+      case ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION => APPLICATION
       case _ => throw new IllegalStateException(s"Illegal K8sExecuteMode, ${mode.name}")
     }
   }
 
-  def toExecutionMode(mode: FlinkK8sExecuteMode.Value): ExecutionMode = {
+  def toExecutionMode(mode: FlinkK8sExecuteModeEnum.Value): ExecutionModeEnum = {
     mode match {
-      case SESSION => ExecutionMode.KUBERNETES_NATIVE_SESSION
-      case APPLICATION => ExecutionMode.KUBERNETES_NATIVE_APPLICATION
+      case SESSION => ExecutionModeEnum.KUBERNETES_NATIVE_SESSION
+      case APPLICATION => ExecutionModeEnum.KUBERNETES_NATIVE_APPLICATION
     }
   }
 
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/event/FlinkJobStateEvent.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/event/FlinkJobStateEvent.scala
index 8d46008c4..53485dd21 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/event/FlinkJobStateEvent.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/event/FlinkJobStateEvent.scala
@@ -17,11 +17,11 @@
 
 package org.apache.streampark.flink.kubernetes.event
 
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum
 import org.apache.streampark.flink.kubernetes.model.TrackId
 
 /**
  * Notification of expecting changes to flink job state cache held internally by K8sFlinkMonitor.
  */
-case class FlinkJobStateEvent(trackId: TrackId, jobState: FlinkJobState.Value, pollTime: Long)
+case class FlinkJobStateEvent(trackId: TrackId, jobState: FlinkJobStateEnum.Value, pollTime: Long)
   extends BuildInEvent
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/ClusterKey.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/ClusterKey.scala
index 3a2c72aa8..237dfd5d9 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/ClusterKey.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/ClusterKey.scala
@@ -18,11 +18,11 @@
 package org.apache.streampark.flink.kubernetes.model
 
 import org.apache.streampark.common.conf.K8sFlinkConfig
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum
 
 /** flink cluster identifier on kubernetes */
 case class ClusterKey(
-    executeMode: FlinkK8sExecuteMode.Value,
+    executeMode: FlinkK8sExecuteModeEnum.Value,
     namespace: String = K8sFlinkConfig.DEFAULT_KUBERNETES_NAMESPACE,
     clusterId: String)
 
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/JobStatusCV.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/JobStatusCV.scala
index 7bd5bd52a..df790a11e 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/JobStatusCV.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/JobStatusCV.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.kubernetes.model
 
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum
 
 /**
  * @param jobState
@@ -34,7 +34,7 @@ import org.apache.streampark.flink.kubernetes.enums.FlinkJobState
  *   tracking polling result receive timestamp
  */
 case class JobStatusCV(
-    jobState: FlinkJobState.Value,
+    jobState: FlinkJobStateEnum.Value,
     jobId: String,
     jobName: String = "",
     jobStartTime: Long = -1,
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/TrackId.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/TrackId.scala
index 410836e6a..34e6dd30d 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/TrackId.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/model/TrackId.scala
@@ -19,13 +19,13 @@ package org.apache.streampark.flink.kubernetes.model
 
 import org.apache.streampark.common.conf.K8sFlinkConfig
 import org.apache.streampark.common.util.Utils
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum
 
 import scala.util.Try
 
 /** tracking identifier for flink on kubernetes */
 case class TrackId(
-    executeMode: FlinkK8sExecuteMode.Value,
+    executeMode: FlinkK8sExecuteModeEnum.Value,
     namespace: String = K8sFlinkConfig.DEFAULT_KUBERNETES_NAMESPACE,
     clusterId: String,
     appId: Long,
@@ -34,9 +34,9 @@ case class TrackId(
 
   def isLegal: Boolean = {
     executeMode match {
-      case FlinkK8sExecuteMode.APPLICATION =>
+      case FlinkK8sExecuteModeEnum.APPLICATION =>
         Try(namespace.nonEmpty).getOrElse(false) && Try(clusterId.nonEmpty).getOrElse(false)
-      case FlinkK8sExecuteMode.SESSION =>
+      case FlinkK8sExecuteModeEnum.SESSION =>
         Try(namespace.nonEmpty).getOrElse(false) && Try(clusterId.nonEmpty).getOrElse(false) && Try(
           jobId.nonEmpty).getOrElse(false)
       case _ => false
@@ -74,7 +74,7 @@ object TrackId {
       appId: Long,
       jobId: String,
       groupId: String): TrackId = {
-    this(FlinkK8sExecuteMode.SESSION, namespace, clusterId, appId, jobId, groupId)
+    this(FlinkK8sExecuteModeEnum.SESSION, namespace, clusterId, appId, jobId, groupId)
   }
 
   def onApplication(
@@ -83,6 +83,6 @@ object TrackId {
       appId: Long,
       jobId: String = null,
       groupId: String): TrackId = {
-    this(FlinkK8sExecuteMode.APPLICATION, namespace, clusterId, appId, jobId, groupId)
+    this(FlinkK8sExecuteModeEnum.APPLICATION, namespace, clusterId, appId, jobId, groupId)
   }
 }
diff --git a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
index e6b0d7923..7dbfcfd0a 100644
--- a/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
+++ b/streampark-flink/streampark-flink-kubernetes/src/main/scala/org/apache/streampark/flink/kubernetes/watcher/FlinkJobStatusWatcher.scala
@@ -20,8 +20,8 @@ package org.apache.streampark.flink.kubernetes.watcher
 import org.apache.streampark.common.conf.Workspace
 import org.apache.streampark.common.util.Logger
 import org.apache.streampark.flink.kubernetes.{ChangeEventBus, FlinkK8sWatchController, JobStatusWatcherConfig, KubernetesRetriever}
-import org.apache.streampark.flink.kubernetes.enums.FlinkJobState
-import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteMode.{APPLICATION, SESSION}
+import org.apache.streampark.flink.kubernetes.enums.FlinkJobStateEnum
+import org.apache.streampark.flink.kubernetes.enums.FlinkK8sExecuteModeEnum.{APPLICATION, SESSION}
 import org.apache.streampark.flink.kubernetes.event.FlinkJobStatusChangeEvent
 import org.apache.streampark.flink.kubernetes.helper.KubernetesDeploymentHelper
 import org.apache.streampark.flink.kubernetes.model._
@@ -124,7 +124,7 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
                 watchController.trackIds.update(trackId)
                 eventBus.postSync(FlinkJobStatusChangeEvent(trackId, jobState))
               }
-              if (FlinkJobState.isEndState(jobState.jobState)) {
+              if (FlinkJobStateEnum.isEndState(jobState.jobState)) {
                 // remove trackId from cache of job that needs to be untracked
                 watchController.unWatching(trackId)
                 if (trackId.executeMode == APPLICATION) {
@@ -167,11 +167,11 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
 
     val rsMap = touchSessionAllJob(clusterId, namespace, appId, trackId.groupId).toMap
     val id = TrackId.onSession(namespace, clusterId, appId, jobId, trackId.groupId)
-    val jobState = rsMap.get(id).filter(_.jobState != FlinkJobState.SILENT).getOrElse {
+    val jobState = rsMap.get(id).filter(_.jobState != FlinkJobStateEnum.SILENT).getOrElse {
       val preCache = watchController.jobStatuses.get(id)
       val state = inferSilentOrLostFromPreCache(preCache)
       val nonFirstSilent =
-        state == FlinkJobState.SILENT && preCache != null && preCache.jobState == FlinkJobState.SILENT
+        state == FlinkJobStateEnum.SILENT && preCache != null && preCache.jobState == FlinkJobStateEnum.SILENT
       if (nonFirstSilent) {
         JobStatusCV(
           jobState = state,
@@ -286,7 +286,7 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
     logger.info(
       s"Query the local cache result:${watchController.canceling.has(trackId).toString},trackId ${trackId.toString}.")
     val jobState = {
-      if (watchController.canceling.has(trackId)) FlinkJobState.CANCELED
+      if (watchController.canceling.has(trackId)) FlinkJobStateEnum.CANCELED
       else {
         // whether deployment exists on kubernetes cluster
         val isDeployExists =
@@ -299,7 +299,7 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
         if (isDeployExists) {
           if (!deployStateOfTheError) {
             logger.info("Task Enter the initialization process.")
-            FlinkJobState.K8S_INITIALIZING
+            FlinkJobStateEnum.K8S_INITIALIZING
           } else if (isConnection) {
             logger.info("Enter the task failure deletion process.")
             KubernetesDeploymentHelper.watchPodTerminatedLog(
@@ -307,13 +307,13 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
               trackId.clusterId,
               trackId.jobId)
             KubernetesDeploymentHelper.deleteTaskDeployment(trackId.namespace, trackId.clusterId)
-            FlinkJobState.FAILED
+            FlinkJobStateEnum.FAILED
           } else {
             inferSilentOrLostFromPreCache(latest)
           }
         } else if (isConnection) {
           logger.info("The deployment is deleted and enters the task failure process.")
-          FlinkJobState.of(FlinkHistoryArchives.getJobStateFromArchiveFile(trackId.jobId))
+          FlinkJobStateEnum.of(FlinkHistoryArchives.getJobStateFromArchiveFile(trackId.jobId))
         } else {
           inferSilentOrLostFromPreCache(latest)
         }
@@ -328,7 +328,7 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
       pollAckTime = System.currentTimeMillis)
 
     if (
-      jobState == FlinkJobState.SILENT && latest != null && latest.jobState == FlinkJobState.SILENT
+      jobState == FlinkJobStateEnum.SILENT && latest != null && latest.jobState == FlinkJobStateEnum.SILENT
     ) {
       Some(jobStatusCV.copy(pollEmitTime = latest.pollEmitTime, pollAckTime = latest.pollAckTime))
     } else {
@@ -337,21 +337,21 @@ class FlinkJobStatusWatcher(conf: JobStatusWatcherConfig = JobStatusWatcherConfi
   }
 
   private[this] def inferSilentOrLostFromPreCache(preCache: JobStatusCV) = preCache match {
-    case preCache if preCache == null => FlinkJobState.SILENT
+    case preCache if preCache == null => FlinkJobStateEnum.SILENT
     case preCache
-        if preCache.jobState == FlinkJobState.SILENT &&
+        if preCache.jobState == FlinkJobStateEnum.SILENT &&
           System
             .currentTimeMillis() - preCache.pollAckTime >= conf.silentStateJobKeepTrackingSec * 1000 =>
-      FlinkJobState.LOST
-    case _ => FlinkJobState.SILENT
+      FlinkJobStateEnum.LOST
+    case _ => FlinkJobStateEnum.SILENT
   }
 
 }
 
 object FlinkJobStatusWatcher {
 
-  private val effectEndStates: Seq[FlinkJobState.Value] =
-    FlinkJobState.endingStates.filter(_ != FlinkJobState.LOST)
+  private val effectEndStates: Seq[FlinkJobStateEnum.Value] =
+    FlinkJobStateEnum.endingStates.filter(_ != FlinkJobStateEnum.LOST)
 
   /**
    * infer flink job state before persistence. so drama, so sad.
@@ -362,18 +362,18 @@ object FlinkJobStatusWatcher {
    *   previous flink job state from persistent storage
    */
   def inferFlinkJobStateFromPersist(
-      current: FlinkJobState.Value,
-      previous: FlinkJobState.Value): FlinkJobState.Value = {
+      current: FlinkJobStateEnum.Value,
+      previous: FlinkJobStateEnum.Value): FlinkJobStateEnum.Value = {
     current match {
-      case FlinkJobState.LOST =>
-        if (effectEndStates.contains(current)) previous else FlinkJobState.TERMINATED
-      case FlinkJobState.POS_TERMINATED | FlinkJobState.TERMINATED =>
+      case FlinkJobStateEnum.LOST =>
+        if (effectEndStates.contains(current)) previous else FlinkJobStateEnum.TERMINATED
+      case FlinkJobStateEnum.POS_TERMINATED | FlinkJobStateEnum.TERMINATED =>
         previous match {
-          case FlinkJobState.CANCELLING => FlinkJobState.CANCELED
-          case FlinkJobState.FAILING => FlinkJobState.FAILED
+          case FlinkJobStateEnum.CANCELLING => FlinkJobStateEnum.CANCELED
+          case FlinkJobStateEnum.FAILING => FlinkJobStateEnum.FAILED
           case _ =>
-            if (current == FlinkJobState.POS_TERMINATED) FlinkJobState.FINISHED
-            else FlinkJobState.TERMINATED
+            if (current == FlinkJobStateEnum.POS_TERMINATED) FlinkJobStateEnum.FINISHED
+            else FlinkJobStateEnum.TERMINATED
         }
       case _ => current
     }
@@ -394,7 +394,7 @@ private[kubernetes] case class JobDetail(
     tasks: JobTask) {
   def toJobStatusCV(pollEmitTime: Long, pollAckTime: Long): JobStatusCV = {
     JobStatusCV(
-      jobState = FlinkJobState.of(state),
+      jobState = FlinkJobStateEnum.of(state),
       jobId = jid,
       jobName = name,
       jobStartTime = startTime,
diff --git a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatus.java b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatusEnum.java
similarity index 90%
rename from streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatus.java
rename to streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatusEnum.java
index 5368dbe2c..b002e7907 100644
--- a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatus.java
+++ b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStatusEnum.java
@@ -34,7 +34,7 @@ import com.fasterxml.jackson.annotation.JsonValue;
  * </pre>
  */
 @SuppressWarnings("AlibabaEnumConstantsMustHaveComment")
-public enum PipelineStatus {
+public enum PipelineStatusEnum {
   unknown(0),
   pending(1),
   running(2),
@@ -43,13 +43,13 @@ public enum PipelineStatus {
 
   private final Integer code;
 
-  PipelineStatus(Integer code) {
+  PipelineStatusEnum(Integer code) {
     this.code = code;
   }
 
   @JsonCreator
-  public static PipelineStatus of(Integer code) {
-    for (PipelineStatus status : PipelineStatus.values()) {
+  public static PipelineStatusEnum of(Integer code) {
+    for (PipelineStatusEnum status : PipelineStatusEnum.values()) {
       if (status.getCode().equals(code)) {
         return status;
       }
diff --git a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatus.java b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatusEnum.java
similarity index 90%
rename from streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatus.java
rename to streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatusEnum.java
index 029bc4044..ac4c35ce3 100644
--- a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatus.java
+++ b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineStepStatusEnum.java
@@ -37,7 +37,7 @@ import com.fasterxml.jackson.annotation.JsonValue;
  * </pre>
  */
 @SuppressWarnings("AlibabaEnumConstantsMustHaveComment")
-public enum PipelineStepStatus {
+public enum PipelineStepStatusEnum {
   unknown(0),
   waiting(1),
   running(2),
@@ -47,13 +47,13 @@ public enum PipelineStepStatus {
 
   private final Integer code;
 
-  PipelineStepStatus(Integer code) {
+  PipelineStepStatusEnum(Integer code) {
     this.code = code;
   }
 
   @JsonCreator
-  public static PipelineStepStatus of(Integer code) {
-    for (PipelineStepStatus status : PipelineStepStatus.values()) {
+  public static PipelineStepStatusEnum of(Integer code) {
+    for (PipelineStepStatusEnum status : PipelineStepStatusEnum.values()) {
       if (status.getCode().equals(code)) {
         return status;
       }
diff --git a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineType.java b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineTypeEnum.java
similarity index 93%
rename from streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineType.java
rename to streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineTypeEnum.java
index 0c0e8e4b1..bf460db4a 100644
--- a/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineType.java
+++ b/streampark-flink/streampark-flink-packer/src/main/java/org/apache/streampark/flink/packer/pipeline/PipelineTypeEnum.java
@@ -24,7 +24,7 @@ import com.google.common.collect.ImmutableMap;
 import java.util.Map;
 
 /** Building pipeline type. */
-public enum PipelineType {
+public enum PipelineTypeEnum {
 
   /** unknown type */
   UNKNOWN(0, "", ImmutableMap.of(), null),
@@ -89,7 +89,7 @@ public enum PipelineType {
   /** type of result */
   private final Class<? extends BuildResult> resultType;
 
-  PipelineType(
+  PipelineTypeEnum(
       Integer code,
       String desc,
       Map<Integer, String> steps,
@@ -101,8 +101,8 @@ public enum PipelineType {
   }
 
   @JsonCreator
-  public static PipelineType of(Integer code) {
-    for (PipelineType type : PipelineType.values()) {
+  public static PipelineTypeEnum of(Integer code) {
+    for (PipelineTypeEnum type : PipelineTypeEnum.values()) {
       if (type.getCode().equals(code)) {
         return type;
       }
@@ -110,8 +110,8 @@ public enum PipelineType {
     return UNKNOWN;
   }
 
-  public static PipelineType ofName(String name) {
-    return PipelineType.valueOf(name);
+  public static PipelineTypeEnum ofName(String name) {
+    return PipelineTypeEnum.valueOf(name);
   }
 
   @JsonValue
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
index d1113722e..bd690667f 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildPipeline.scala
@@ -31,7 +31,7 @@ import scala.util.{Failure, Success, Try}
 trait BuildPipelineProcess {
 
   /** the type of pipeline */
-  def pipeType: PipelineType
+  def pipeType: PipelineTypeEnum
 
   /**
    * the actual build process. the effective steps progress should be implemented in multiple
@@ -48,13 +48,13 @@ trait BuildPipelineProcess {
 trait BuildPipelineExpose {
 
   /** get current state of the pipeline instance */
-  def getPipeStatus: PipelineStatus
+  def getPipeStatus: PipelineStatusEnum
 
   /** get error of pipeline instance */
   def getError: PipeError
 
   /** get all of the steps status StepSeq -> (PipeStepStatus -> status update timestamp) */
-  def getStepsStatus: Map[Int, (PipelineStepStatus, Long)]
+  def getStepsStatus: Map[Int, (PipelineStepStatusEnum, Long)]
 
   /** get current build step index */
   def getCurStep: Int
@@ -71,16 +71,16 @@ trait BuildPipelineExpose {
 /** Building pipeline trait. */
 trait BuildPipeline extends BuildPipelineProcess with BuildPipelineExpose with Logger {
 
-  protected var pipeStatus: PipelineStatus = PipelineStatus.pending
+  protected var pipeStatus: PipelineStatusEnum = PipelineStatusEnum.pending
 
   protected var error: PipeError = PipeError.empty()
 
   protected var curStep: Int = 0
 
-  protected val stepsStatus: mutable.Map[Int, (PipelineStepStatus, Long)] =
+  protected val stepsStatus: mutable.Map[Int, (PipelineStepStatusEnum, Long)] =
     mutable.Map(
       pipeType.getSteps.asScala
-        .map(e => e._1.toInt -> (PipelineStepStatus.waiting -> System.currentTimeMillis))
+        .map(e => e._1.toInt -> (PipelineStepStatusEnum.waiting -> System.currentTimeMillis))
         .toSeq: _*)
 
   /** use to identify the log record that belongs to which pipeline instance */
@@ -96,19 +96,19 @@ trait BuildPipeline extends BuildPipelineProcess with BuildPipelineExpose with L
   protected def execStep[R](seq: Int)(process: => R): Option[R] = {
     Try {
       curStep = seq
-      stepsStatus(seq) = PipelineStepStatus.running -> System.currentTimeMillis
+      stepsStatus(seq) = PipelineStepStatusEnum.running -> System.currentTimeMillis
       logInfo(s"building pipeline step[$seq/$allSteps] running => ${pipeType.getSteps.get(seq)}")
       watcher.onStepStateChange(snapshot)
       process
     } match {
       case Success(result) =>
-        stepsStatus(seq) = PipelineStepStatus.success -> System.currentTimeMillis
+        stepsStatus(seq) = PipelineStepStatusEnum.success -> System.currentTimeMillis
         logInfo(s"building pipeline step[$seq/$allSteps] success")
         watcher.onStepStateChange(snapshot)
         Some(result)
       case Failure(cause) =>
-        stepsStatus(seq) = PipelineStepStatus.failure -> System.currentTimeMillis
-        pipeStatus = PipelineStatus.failure
+        stepsStatus(seq) = PipelineStepStatusEnum.failure -> System.currentTimeMillis
+        pipeStatus = PipelineStatusEnum.failure
         error = PipeError.of(cause.getMessage, cause)
         logInfo(s"building pipeline step[$seq/$allSteps] failure => ${pipeType.getSteps.get(seq)}")
         watcher.onStepStateChange(snapshot)
@@ -118,14 +118,14 @@ trait BuildPipeline extends BuildPipelineProcess with BuildPipelineExpose with L
 
   protected def skipStep(step: Int): Unit = {
     curStep = step
-    stepsStatus(step) = PipelineStepStatus.skipped -> System.currentTimeMillis
+    stepsStatus(step) = PipelineStepStatusEnum.skipped -> System.currentTimeMillis
     logInfo(s"building pipeline step[$step/$allSteps] skipped => ${pipeType.getSteps.get(step)}")
     watcher.onStepStateChange(snapshot)
   }
 
   /** Launch the building pipeline. */
   override def launch(): BuildResult = {
-    pipeStatus = PipelineStatus.running
+    pipeStatus = PipelineStatusEnum.running
     Try {
       watcher.onStart(snapshot)
       logInfo(s"building pipeline is launching, params=${offerBuildParam.toString}")
@@ -136,12 +136,12 @@ trait BuildPipeline extends BuildPipelineProcess with BuildPipelineExpose with L
         .get(20, TimeUnit.MINUTES)
     } match {
       case Success(result) =>
-        pipeStatus = PipelineStatus.success
+        pipeStatus = PipelineStatusEnum.success
         logInfo(s"building pipeline has finished successfully.")
         watcher.onFinish(snapshot, result)
         result
       case Failure(cause) =>
-        pipeStatus = PipelineStatus.failure
+        pipeStatus = PipelineStatusEnum.failure
         error = PipeError.of(cause.getMessage, cause)
         // log and print error trace stack
         logError(s"building pipeline has failed.", cause)
@@ -151,11 +151,11 @@ trait BuildPipeline extends BuildPipelineProcess with BuildPipelineExpose with L
     }
   }
 
-  override def getPipeStatus: PipelineStatus = pipeStatus
+  override def getPipeStatus: PipelineStatusEnum = pipeStatus
 
   override def getError: PipeError = error.copy()
 
-  override def getStepsStatus: Map[Int, (PipelineStepStatus, Long)] = stepsStatus.toMap
+  override def getStepsStatus: Map[Int, (PipelineStepStatusEnum, Long)] = stepsStatus.toMap
 
   override def getCurStep: Int = curStep
 
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildRequest.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildRequest.scala
index 41e6e4f46..1dbfeab32 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildRequest.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/BuildRequest.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.packer.pipeline
 
 import org.apache.streampark.common.conf.{FlinkVersion, Workspace}
-import org.apache.streampark.common.enums.{DevelopmentMode, ExecutionMode}
+import org.apache.streampark.common.enums.{DevelopmentModeEnum, ExecutionModeEnum}
 import org.apache.streampark.flink.kubernetes.model.K8sPodTemplates
 import org.apache.streampark.flink.packer.docker.DockerConf
 import org.apache.streampark.flink.packer.maven.DependencyInfo
@@ -38,9 +38,9 @@ sealed trait FlinkBuildParam extends BuildParam {
 
   def workspace: String
 
-  def executionMode: ExecutionMode
+  def executionMode: ExecutionModeEnum
 
-  def developmentMode: DevelopmentMode
+  def developmentMode: DevelopmentModeEnum
 
   def flinkVersion: FlinkVersion
 
@@ -51,7 +51,7 @@ sealed trait FlinkBuildParam extends BuildParam {
   lazy val providedLibs: DependencyInfo = {
     val providedLibs =
       ArrayBuffer(localWorkspace.APP_JARS, localWorkspace.APP_PLUGINS, customFlinkUserJar)
-    if (developmentMode == DevelopmentMode.FLINK_SQL) {
+    if (developmentMode == DevelopmentModeEnum.FLINK_SQL) {
       providedLibs += s"${localWorkspace.APP_SHIMS}/flink-${flinkVersion.majorVersion}"
     }
     dependencyInfo.merge(providedLibs.toSet)
@@ -76,8 +76,8 @@ case class FlinkK8sSessionBuildRequest(
     workspace: String,
     mainClass: String,
     customFlinkUserJar: String,
-    executionMode: ExecutionMode,
-    developmentMode: DevelopmentMode,
+    executionMode: ExecutionModeEnum,
+    developmentMode: DevelopmentModeEnum,
     flinkVersion: FlinkVersion,
     dependencyInfo: DependencyInfo,
     clusterId: String,
@@ -89,8 +89,8 @@ case class FlinkK8sApplicationBuildRequest(
     workspace: String,
     mainClass: String,
     customFlinkUserJar: String,
-    executionMode: ExecutionMode,
-    developmentMode: DevelopmentMode,
+    executionMode: ExecutionModeEnum,
+    developmentMode: DevelopmentModeEnum,
     flinkVersion: FlinkVersion,
     dependencyInfo: DependencyInfo,
     clusterId: String,
@@ -108,8 +108,8 @@ case class FlinkRemotePerJobBuildRequest(
     mainClass: String,
     customFlinkUserJar: String,
     skipBuild: Boolean,
-    executionMode: ExecutionMode,
-    developmentMode: DevelopmentMode,
+    executionMode: ExecutionModeEnum,
+    developmentMode: DevelopmentModeEnum,
     flinkVersion: FlinkVersion,
     dependencyInfo: DependencyInfo)
   extends FlinkBuildParam
@@ -119,6 +119,6 @@ case class FlinkYarnApplicationBuildRequest(
     mainClass: String,
     localWorkspace: String,
     yarnProvidedPath: String,
-    developmentMode: DevelopmentMode,
+    developmentMode: DevelopmentModeEnum,
     dependencyInfo: DependencyInfo)
   extends BuildParam
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipeSnapshot.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipeSnapshot.scala
index 036745da9..dc159a4f5 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipeSnapshot.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/PipeSnapshot.scala
@@ -36,24 +36,24 @@ import scala.collection.JavaConverters._
 
 case class PipeSnapshot(
     appName: String,
-    pipeType: PipelineType,
-    pipeStatus: PipelineStatus,
+    pipeType: PipelineTypeEnum,
+    pipeStatus: PipelineStatusEnum,
     curStep: Int,
     allSteps: Int,
-    stepStatus: Map[Int, (PipelineStepStatus, Long)],
+    stepStatus: Map[Int, (PipelineStepStatusEnum, Long)],
     error: PipeError,
     emitTime: Long) {
 
   def percent(): Double = Utils.calPercent(curStep, allSteps)
 
-  def stepStatusAsJava: JavaMap[Integer, (PipelineStepStatus, JavaLong)] = {
+  def stepStatusAsJava: JavaMap[Integer, (PipelineStepStatusEnum, JavaLong)] = {
     stepStatus.toSeq
       .map(e => Integer.valueOf(e._1) -> (e._2._1 -> JavaLong.valueOf(e._2._2)))
       .toMap
       .asJava
   }
 
-  def pureStepStatusAsJava: JavaMap[Integer, PipelineStepStatus] = {
+  def pureStepStatusAsJava: JavaMap[Integer, PipelineStepStatusEnum] = {
     stepStatus.toSeq.map(e => Integer.valueOf(e._1) -> e._2._1).toMap.asJava
   }
 
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipeline.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipeline.scala
index 5e0449aee..f94af2347 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipeline.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipeline.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.packer.pipeline.impl
 
-import org.apache.streampark.common.enums.DevelopmentMode
+import org.apache.streampark.common.enums.DevelopmentModeEnum
 import org.apache.streampark.common.fs.LfsOperator
 import org.apache.streampark.common.util.ThreadUtils
 import org.apache.streampark.flink.kubernetes.PodTemplateTool
@@ -42,7 +42,7 @@ import scala.language.postfixOps
 class FlinkK8sApplicationBuildPipeline(request: FlinkK8sApplicationBuildRequest)
   extends BuildPipeline {
 
-  override def pipeType: PipelineType = PipelineType.FLINK_NATIVE_K8S_APPLICATION
+  override def pipeType: PipelineTypeEnum = PipelineTypeEnum.FLINK_NATIVE_K8S_APPLICATION
 
   private var dockerProcessWatcher: DockerProgressWatcher = new SilentDockerProgressWatcher
 
@@ -91,8 +91,8 @@ class FlinkK8sApplicationBuildPipeline(request: FlinkK8sApplicationBuildRequest)
       execStep(3) {
         val shadedJarOutputPath = request.getShadedJarPath(buildWorkspace)
         val extJarLibs = request.developmentMode match {
-          case DevelopmentMode.FLINK_SQL => request.dependencyInfo.extJarLibs
-          case DevelopmentMode.CUSTOM_CODE => Set[String]()
+          case DevelopmentModeEnum.FLINK_SQL => request.dependencyInfo.extJarLibs
+          case DevelopmentModeEnum.CUSTOM_CODE => Set[String]()
         }
         val shadedJar =
           MavenTool.buildFatJar(request.mainClass, request.providedLibs, shadedJarOutputPath)
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipelineV2.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipelineV2.scala
index ba9b9c6bc..ea490bbb6 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipelineV2.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sApplicationBuildPipelineV2.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.packer.pipeline.impl
 
-import org.apache.streampark.common.enums.DevelopmentMode
+import org.apache.streampark.common.enums.DevelopmentModeEnum
 import org.apache.streampark.common.fs.LfsOperator
 import org.apache.streampark.flink.packer.maven.MavenTool
 import org.apache.streampark.flink.packer.pipeline._
@@ -30,7 +30,7 @@ import scala.language.postfixOps
 class FlinkK8sApplicationBuildPipelineV2(request: FlinkK8sApplicationBuildRequest)
   extends BuildPipeline {
 
-  override def pipeType: PipelineType = PipelineType.FLINK_K8S_APPLICATION_V2
+  override def pipeType: PipelineTypeEnum = PipelineTypeEnum.FLINK_K8S_APPLICATION_V2
 
   @throws[Throwable]
   override protected def buildProcess(): K8sAppModeBuildResponse = {
@@ -51,8 +51,8 @@ class FlinkK8sApplicationBuildPipelineV2(request: FlinkK8sApplicationBuildReques
       execStep(2) {
         val shadedJarOutputPath = request.getShadedJarPath(buildWorkspace)
         val extJarLibs = request.developmentMode match {
-          case DevelopmentMode.FLINK_SQL => request.dependencyInfo.extJarLibs
-          case DevelopmentMode.CUSTOM_CODE => Set.empty[String]
+          case DevelopmentModeEnum.FLINK_SQL => request.dependencyInfo.extJarLibs
+          case DevelopmentModeEnum.CUSTOM_CODE => Set.empty[String]
           case _ => Set.empty[String]
         }
         val shadedJar =
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sSessionBuildPipeline.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sSessionBuildPipeline.scala
index 048876fda..9b86e9b5f 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sSessionBuildPipeline.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkK8sSessionBuildPipeline.scala
@@ -24,7 +24,7 @@ import org.apache.streampark.flink.packer.pipeline._
 /** Building pipeline for flink kubernetes-native session mode */
 class FlinkK8sSessionBuildPipeline(request: FlinkK8sSessionBuildRequest) extends BuildPipeline {
 
-  override def pipeType: PipelineType = PipelineType.FLINK_NATIVE_K8S_SESSION
+  override def pipeType: PipelineTypeEnum = PipelineTypeEnum.FLINK_NATIVE_K8S_SESSION
 
   override def offerBuildParam: FlinkK8sSessionBuildRequest = request
 
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
index e3250b7c0..199e0b3db 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkRemoteBuildPipeline.scala
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.flink.packer.pipeline.impl
 
-import org.apache.streampark.common.enums.DevelopmentMode
+import org.apache.streampark.common.enums.DevelopmentModeEnum
 import org.apache.streampark.common.fs.{FsOperator, LfsOperator}
 import org.apache.streampark.flink.packer.maven.MavenTool
 import org.apache.streampark.flink.packer.pipeline._
@@ -30,7 +30,7 @@ import scala.collection.convert.ImplicitConversions._
 /** Building pipeline for flink standalone session mode */
 class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends BuildPipeline {
 
-  override def pipeType: PipelineType = PipelineType.FLINK_STANDALONE
+  override def pipeType: PipelineTypeEnum = PipelineTypeEnum.FLINK_STANDALONE
 
   override def offerBuildParam: FlinkRemotePerJobBuildRequest = request
 
@@ -50,7 +50,7 @@ class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends B
       val shadedJar =
         execStep(2) {
           request.developmentMode match {
-            case DevelopmentMode.FLINK_SQL =>
+            case DevelopmentModeEnum.FLINK_SQL =>
               val output = MavenTool.buildFatJar(
                 request.mainClass,
                 request.providedLibs,
@@ -64,7 +64,7 @@ class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends B
       val mavenJars =
         execStep(3) {
           request.developmentMode match {
-            case DevelopmentMode.PYFLINK =>
+            case DevelopmentModeEnum.PYFLINK =>
               val mavenArts = MavenTool.resolveArtifacts(request.dependencyInfo.mavenArts.asJava)
               mavenArts.map(_.getAbsolutePath) ++ request.dependencyInfo.extJarLibs
             case _ => List[String]()
@@ -73,7 +73,7 @@ class FlinkRemoteBuildPipeline(request: FlinkRemotePerJobBuildRequest) extends B
 
       execStep(4) {
         request.developmentMode match {
-          case DevelopmentMode.PYFLINK =>
+          case DevelopmentModeEnum.PYFLINK =>
             mavenJars.foreach(
               jar => {
                 val lfs: FsOperator = FsOperator.lfs
diff --git a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
index 2902c339c..36023cb8e 100644
--- a/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
+++ b/streampark-flink/streampark-flink-packer/src/main/scala/org/apache/streampark/flink/packer/pipeline/impl/FlinkYarnApplicationBuildPipeline.scala
@@ -18,7 +18,7 @@
 package org.apache.streampark.flink.packer.pipeline.impl
 
 import org.apache.streampark.common.conf.Workspace
-import org.apache.streampark.common.enums.DevelopmentMode
+import org.apache.streampark.common.enums.DevelopmentModeEnum
 import org.apache.streampark.common.fs.{FsOperator, HdfsOperator, LfsOperator}
 import org.apache.streampark.common.util.ImplicitsUtils._
 import org.apache.streampark.flink.packer.maven.MavenTool
@@ -35,7 +35,7 @@ class FlinkYarnApplicationBuildPipeline(request: FlinkYarnApplicationBuildReques
   extends BuildPipeline {
 
   /** the type of pipeline */
-  override def pipeType: PipelineType = PipelineType.FLINK_YARN_APPLICATION
+  override def pipeType: PipelineTypeEnum = PipelineTypeEnum.FLINK_YARN_APPLICATION
 
   override def offerBuildParam: FlinkYarnApplicationBuildRequest = request
 
@@ -47,7 +47,7 @@ class FlinkYarnApplicationBuildPipeline(request: FlinkYarnApplicationBuildReques
   override protected def buildProcess(): SimpleBuildResponse = {
     execStep(1) {
       request.developmentMode match {
-        case DevelopmentMode.FLINK_SQL | DevelopmentMode.PYFLINK =>
+        case DevelopmentModeEnum.FLINK_SQL | DevelopmentModeEnum.PYFLINK =>
           LfsOperator.mkCleanDirs(request.localWorkspace)
           HdfsOperator.mkCleanDirs(request.yarnProvidedPath)
         case _ =>
@@ -58,7 +58,7 @@ class FlinkYarnApplicationBuildPipeline(request: FlinkYarnApplicationBuildReques
     val mavenJars =
       execStep(2) {
         request.developmentMode match {
-          case DevelopmentMode.FLINK_SQL | DevelopmentMode.PYFLINK =>
+          case DevelopmentModeEnum.FLINK_SQL | DevelopmentModeEnum.PYFLINK =>
             val mavenArts = MavenTool.resolveArtifacts(request.dependencyInfo.mavenArts)
             mavenArts.map(_.getAbsolutePath) ++ request.dependencyInfo.extJarLibs
           case _ => List[String]()
diff --git a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkSqlValidator.scala b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkSqlValidator.scala
index 26c31f928..87d7ec91d 100644
--- a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkSqlValidator.scala
+++ b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkSqlValidator.scala
@@ -16,7 +16,7 @@
  */
 package org.apache.streampark.flink.core
 
-import org.apache.streampark.common.enums.FlinkSqlValidationFailedType
+import org.apache.streampark.common.enums.FlinkSqlValidationFailedTypeEnum
 import org.apache.streampark.common.util.{ExceptionUtils, Logger, Utils}
 import org.apache.streampark.flink.core.SqlCommand._
 
@@ -103,7 +103,7 @@ object FlinkSqlValidator extends Logger {
                 val errorLine = call.lineStart + line.toInt - 1
                 return FlinkSqlValidationResult(
                   success = false,
-                  failedType = FlinkSqlValidationFailedType.SYNTAX_ERROR,
+                  failedType = FlinkSqlValidationFailedTypeEnum.SYNTAX_ERROR,
                   lineStart = call.lineStart,
                   lineEnd = call.lineEnd,
                   errorLine = errorLine,
@@ -114,7 +114,7 @@ object FlinkSqlValidator extends Logger {
               } else {
                 return FlinkSqlValidationResult(
                   success = false,
-                  failedType = FlinkSqlValidationFailedType.SYNTAX_ERROR,
+                  failedType = FlinkSqlValidationFailedTypeEnum.SYNTAX_ERROR,
                   lineStart = call.lineStart,
                   lineEnd = call.lineEnd,
                   sql = call.originSql,
@@ -131,7 +131,7 @@ object FlinkSqlValidator extends Logger {
     } else {
       FlinkSqlValidationResult(
         success = false,
-        failedType = FlinkSqlValidationFailedType.SYNTAX_ERROR,
+        failedType = FlinkSqlValidationFailedTypeEnum.SYNTAX_ERROR,
         lineStart = sqlCommands.head.lineStart,
         lineEnd = sqlCommands.last.lineEnd,
         exception = "No 'INSERT' statement to trigger the execution of the Flink job."
diff --git a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
index bf2e2c68b..4ecba14a1 100644
--- a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
+++ b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkStreamingInitializer.scala
@@ -17,8 +17,8 @@
 package org.apache.streampark.flink.core
 
 import org.apache.streampark.common.conf.ConfigConst._
-import org.apache.streampark.common.enums.ApiType
-import org.apache.streampark.common.enums.ApiType.ApiType
+import org.apache.streampark.common.enums.ApiTypeEnum
+import org.apache.streampark.common.enums.ApiTypeEnum.ApiType
 import org.apache.streampark.common.util._
 import org.apache.streampark.flink.core.conf.FlinkConfiguration
 
@@ -36,13 +36,13 @@ private[flink] object FlinkStreamingInitializer {
 
   def initialize(args: Array[String], config: (StreamExecutionEnvironment, ParameterTool) => Unit)
       : (ParameterTool, StreamExecutionEnvironment) = {
-    val flinkInitializer = new FlinkStreamingInitializer(args, ApiType.scala)
+    val flinkInitializer = new FlinkStreamingInitializer(args, ApiTypeEnum.SCALA)
     flinkInitializer.streamEnvConfFunc = config
     (flinkInitializer.configuration.parameter, flinkInitializer.streamEnv)
   }
 
   def initialize(args: StreamEnvConfig): (ParameterTool, StreamExecutionEnvironment) = {
-    val flinkInitializer = new FlinkStreamingInitializer(args.args, ApiType.java)
+    val flinkInitializer = new FlinkStreamingInitializer(args.args, ApiTypeEnum.JAVA)
     flinkInitializer.javaStreamEnvConfFunc = args.conf
     (flinkInitializer.configuration.parameter, flinkInitializer.streamEnv)
   }
@@ -66,9 +66,9 @@ private[flink] class FlinkStreamingInitializer(args: Array[String], apiType: Api
       JavaStreamEnv.getExecutionEnvironment(configuration.envConfig))
 
     apiType match {
-      case ApiType.java if javaStreamEnvConfFunc != null =>
+      case ApiTypeEnum.JAVA if javaStreamEnvConfFunc != null =>
         javaStreamEnvConfFunc.configuration(env.getJavaEnv, configuration.parameter)
-      case ApiType.scala if streamEnvConfFunc != null =>
+      case ApiTypeEnum.SCALA if streamEnvConfFunc != null =>
         streamEnvConfFunc(env, configuration.parameter)
       case _ =>
     }
diff --git a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
index a401b437e..d45d42336 100644
--- a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
+++ b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/FlinkTableInitializer.scala
@@ -17,8 +17,8 @@
 package org.apache.streampark.flink.core
 
 import org.apache.streampark.common.conf.ConfigConst._
-import org.apache.streampark.common.enums.{ApiType, PlannerType}
-import org.apache.streampark.common.enums.ApiType.ApiType
+import org.apache.streampark.common.enums.{ApiTypeEnum, PlannerTypeEnum}
+import org.apache.streampark.common.enums.ApiTypeEnum.ApiType
 import org.apache.streampark.common.util.{DeflaterUtils, PropertiesUtils}
 import org.apache.streampark.flink.core.EnhancerImplicit._
 import org.apache.streampark.flink.core.conf.FlinkConfiguration
@@ -40,13 +40,13 @@ private[flink] object FlinkTableInitializer {
   def initialize(
       args: Array[String],
       config: (TableConfig, ParameterTool) => Unit): (ParameterTool, TableEnvironment) = {
-    val flinkInitializer = new FlinkTableInitializer(args, ApiType.scala)
+    val flinkInitializer = new FlinkTableInitializer(args, ApiTypeEnum.SCALA)
     flinkInitializer.tableConfFunc = config
     (flinkInitializer.configuration.parameter, flinkInitializer.tableEnv)
   }
 
   def initialize(args: TableEnvConfig): (ParameterTool, TableEnvironment) = {
-    val flinkInitializer = new FlinkTableInitializer(args.args, ApiType.java)
+    val flinkInitializer = new FlinkTableInitializer(args.args, ApiTypeEnum.JAVA)
     flinkInitializer.javaTableEnvConfFunc = args.conf
     (flinkInitializer.configuration.parameter, flinkInitializer.tableEnv)
   }
@@ -57,7 +57,7 @@ private[flink] object FlinkTableInitializer {
       configTable: (TableConfig, ParameterTool) => Unit)
       : (ParameterTool, StreamExecutionEnvironment, StreamTableEnvironment) = {
 
-    val flinkInitializer = new FlinkTableInitializer(args, ApiType.scala)
+    val flinkInitializer = new FlinkTableInitializer(args, ApiTypeEnum.SCALA)
     flinkInitializer.streamEnvConfFunc = configStream
     flinkInitializer.tableConfFunc = configTable
     (
@@ -68,7 +68,7 @@ private[flink] object FlinkTableInitializer {
 
   def initialize(args: StreamTableEnvConfig)
       : (ParameterTool, StreamExecutionEnvironment, StreamTableEnvironment) = {
-    val flinkInitializer = new FlinkTableInitializer(args.args, ApiType.java)
+    val flinkInitializer = new FlinkTableInitializer(args.args, ApiTypeEnum.JAVA)
     flinkInitializer.javaStreamEnvConfFunc = args.streamConfig
     flinkInitializer.javaTableEnvConfFunc = args.tableConfig
     (
@@ -86,9 +86,9 @@ private[flink] class FlinkTableInitializer(args: Array[String], apiType: ApiType
 
     val builder = EnvironmentSettings.newInstance()
 
-    Try(PlannerType.withName(parameter.get(KEY_FLINK_TABLE_PLANNER)))
-      .getOrElse(PlannerType.blink) match {
-      case PlannerType.blink =>
+    Try(PlannerTypeEnum.withName(parameter.get(KEY_FLINK_TABLE_PLANNER)))
+      .getOrElse(PlannerTypeEnum.BLINK) match {
+      case PlannerTypeEnum.BLINK =>
         val useBlinkPlanner =
           Try(builder.getClass.getDeclaredMethod("useBlinkPlanner")).getOrElse(null)
         if (useBlinkPlanner == null) {
@@ -98,7 +98,7 @@ private[flink] class FlinkTableInitializer(args: Array[String], apiType: ApiType
           useBlinkPlanner.invoke(builder)
           logInfo("blinkPlanner will be use.")
         }
-      case PlannerType.old =>
+      case PlannerTypeEnum.OLD =>
         val useOldPlanner = Try(builder.getClass.getDeclaredMethod("useOldPlanner")).getOrElse(null)
         if (useOldPlanner == null) {
           logWarn("useOldPlanner deprecated")
@@ -107,7 +107,7 @@ private[flink] class FlinkTableInitializer(args: Array[String], apiType: ApiType
           useOldPlanner.invoke(builder)
           logInfo("useOldPlanner will be use.")
         }
-      case PlannerType.any =>
+      case PlannerTypeEnum.ANY =>
         val useAnyPlanner = Try(builder.getClass.getDeclaredMethod("useAnyPlanner")).getOrElse(null)
         if (useAnyPlanner == null) {
           logWarn("useAnyPlanner deprecated")
@@ -142,9 +142,9 @@ private[flink] class FlinkTableInitializer(args: Array[String], apiType: ApiType
     envSettings.inBatchMode()
     val tableEnv = TableEnvironment.create(envSettings.build()).setAppName
     apiType match {
-      case ApiType.java if javaTableEnvConfFunc != null =>
+      case ApiTypeEnum.JAVA if javaTableEnvConfFunc != null =>
         javaTableEnvConfFunc.configuration(tableEnv.getConfig, parameter)
-      case ApiType.scala if tableConfFunc != null =>
+      case ApiTypeEnum.SCALA if tableConfFunc != null =>
         tableConfFunc(tableEnv.getConfig, parameter)
       case _ =>
     }
@@ -164,9 +164,9 @@ private[flink] class FlinkTableInitializer(args: Array[String], apiType: ApiType
     }
     val streamTableEnv = StreamTableEnvironment.create(streamEnv, setting).setAppName
     apiType match {
-      case ApiType.java if javaTableEnvConfFunc != null =>
+      case ApiTypeEnum.JAVA if javaTableEnvConfFunc != null =>
         javaTableEnvConfFunc.configuration(streamTableEnv.getConfig, parameter)
-      case ApiType.scala if tableConfFunc != null =>
+      case ApiTypeEnum.SCALA if tableConfFunc != null =>
         tableConfFunc(streamTableEnv.getConfig, parameter)
       case _ =>
     }
diff --git a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/SqlCommandParser.scala b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/SqlCommandParser.scala
index 0d21aab5c..4cd86aac8 100644
--- a/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/SqlCommandParser.scala
+++ b/streampark-flink/streampark-flink-shims/streampark-flink-shims-base/src/main/scala/org/apache/streampark/flink/core/SqlCommandParser.scala
@@ -17,7 +17,7 @@
 package org.apache.streampark.flink.core
 
 import org.apache.streampark.common.conf.ConfigConst.PARAM_PREFIX
-import org.apache.streampark.common.enums.FlinkSqlValidationFailedType
+import org.apache.streampark.common.enums.FlinkSqlValidationFailedTypeEnum
 import org.apache.streampark.common.util.Logger
 
 import enumeratum.EnumEntry
@@ -46,7 +46,7 @@ object SqlCommandParser extends Logger {
           validationCallback(
             FlinkSqlValidationResult(
               success = false,
-              failedType = FlinkSqlValidationFailedType.VERIFY_FAILED,
+              failedType = FlinkSqlValidationFailedTypeEnum.VERIFY_FAILED,
               exception = sqlEmptyError))
           null
         } else {
@@ -62,7 +62,7 @@ object SqlCommandParser extends Logger {
                 validationCallback(
                   FlinkSqlValidationResult(
                     success = false,
-                    failedType = FlinkSqlValidationFailedType.UNSUPPORTED_SQL,
+                    failedType = FlinkSqlValidationFailedTypeEnum.UNSUPPORTED_SQL,
                     lineStart = segment.start,
                     lineEnd = segment.end,
                     exception = s"unsupported sql",
@@ -80,7 +80,7 @@ object SqlCommandParser extends Logger {
               validationCallback(
                 FlinkSqlValidationResult(
                   success = false,
-                  failedType = FlinkSqlValidationFailedType.VERIFY_FAILED,
+                  failedType = FlinkSqlValidationFailedTypeEnum.VERIFY_FAILED,
                   exception = "flink sql syntax error, no executable sql"))
               null
             } else {
@@ -407,7 +407,7 @@ case class SqlCommandCall(
 
 case class FlinkSqlValidationResult(
     success: JavaBool = true,
-    failedType: FlinkSqlValidationFailedType = null,
+    failedType: FlinkSqlValidationFailedTypeEnum = null,
     lineStart: Int = 0,
     lineEnd: Int = 0,
     errorLine: Int = 0,
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatus.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatusEnum.java
similarity index 90%
rename from streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatus.java
rename to streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatusEnum.java
index 21c9c0e4f..201af4f19 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatus.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/OperationStatusEnum.java
@@ -23,7 +23,7 @@ import java.util.HashSet;
 import java.util.Set;
 
 /** Status to describe the {@code Operation}. */
-public enum OperationStatus {
+public enum OperationStatusEnum {
   /** The operation is newly created. */
   INITIALIZED(false),
 
@@ -50,12 +50,12 @@ public enum OperationStatus {
 
   private final boolean isTerminalStatus;
 
-  OperationStatus(boolean isTerminalStatus) {
+  OperationStatusEnum(boolean isTerminalStatus) {
     this.isTerminalStatus = isTerminalStatus;
   }
 
   public static boolean isValidStatusTransition(
-      OperationStatus fromStatus, OperationStatus toStatus) {
+      OperationStatusEnum fromStatus, OperationStatusEnum toStatus) {
     return toOperationStatusSet(fromStatus).contains(toStatus);
   }
 
@@ -63,7 +63,7 @@ public enum OperationStatus {
     return isTerminalStatus;
   }
 
-  private static Set<OperationStatus> toOperationStatusSet(OperationStatus fromStatus) {
+  private static Set<OperationStatusEnum> toOperationStatusSet(OperationStatusEnum fromStatus) {
     switch (fromStatus) {
       case INITIALIZED:
         return new HashSet<>(Arrays.asList(PENDING, CANCELED, CLOSED, TIMEOUT, ERROR));
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientation.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientationEnum.java
similarity index 96%
rename from streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientation.java
rename to streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientationEnum.java
index b6439a504..14d27c229 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientation.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/FetchOrientationEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.gateway.results;
 
 /** Orientation to fetch results. */
-public enum FetchOrientation {
+public enum FetchOrientationEnum {
 
   /** Fetch the next results. */
   FETCH_NEXT,
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/OperationInfo.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/OperationInfo.java
index 7699f210b..b50fb4525 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/OperationInfo.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/OperationInfo.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.gateway.results;
 
-import org.apache.streampark.gateway.OperationStatus;
+import org.apache.streampark.gateway.OperationStatusEnum;
 
 import javax.annotation.Nullable;
 
@@ -27,15 +27,15 @@ import java.util.Objects;
 /** Information of the {@code Operation}. */
 public class OperationInfo implements Serializable {
 
-  private final OperationStatus status;
+  private final OperationStatusEnum status;
   @Nullable private final Exception exception;
 
-  public OperationInfo(OperationStatus status, @Nullable Exception exception) {
+  public OperationInfo(OperationStatusEnum status, @Nullable Exception exception) {
     this.status = status;
     this.exception = exception;
   }
 
-  public OperationStatus getStatus() {
+  public OperationStatusEnum getStatus() {
     return status;
   }
 
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKind.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKindEnum.java
similarity index 97%
rename from streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKind.java
rename to streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKindEnum.java
index f2fd7e82e..05362ebed 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKind.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultKindEnum.java
@@ -18,7 +18,7 @@
 package org.apache.streampark.gateway.results;
 
 /** ResultKind defines the types of the result. */
-public enum ResultKind {
+public enum ResultKindEnum {
   /**
    * The statement (e.g. DDL, USE) executes successfully, and the result only contains a simple
    * "OK".
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultQueryCondition.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultQueryCondition.java
index cc740a950..5dafa0348 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultQueryCondition.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultQueryCondition.java
@@ -22,24 +22,24 @@ import java.util.Objects;
 
 /** Condition of result query. */
 public class ResultQueryCondition implements Serializable {
-  public FetchOrientation orientation;
+  public FetchOrientationEnum orientation;
 
   public long token;
   public int maxRows;
 
   public ResultQueryCondition() {}
 
-  public ResultQueryCondition(FetchOrientation orientation, long token, int maxRows) {
+  public ResultQueryCondition(FetchOrientationEnum orientation, long token, int maxRows) {
     this.orientation = orientation;
     this.token = token;
     this.maxRows = maxRows;
   }
 
-  public FetchOrientation getOrientation() {
+  public FetchOrientationEnum getOrientation() {
     return orientation;
   }
 
-  public void setOrientation(FetchOrientation orientation) {
+  public void setOrientation(FetchOrientationEnum orientation) {
     this.orientation = orientation;
   }
 
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultSet.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultSet.java
index fafc0a800..894c7bc39 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultSet.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/ResultSet.java
@@ -66,7 +66,7 @@ public class ResultSet implements Serializable {
   @Nullable private final JobID jobID;
 
   /** Gets the result kind of the result. */
-  private final ResultKind resultKind;
+  private final ResultKindEnum resultKindEnum;
 
   public ResultSet(
       ResultType resultType,
@@ -75,14 +75,14 @@ public class ResultSet implements Serializable {
       List<RowData> data,
       boolean isQueryResult,
       @Nullable JobID jobID,
-      ResultKind resultKind) {
+      ResultKindEnum resultKindEnum) {
     this.resultType = resultType;
     this.nextToken = nextToken;
     this.columns = columns;
     this.data = data;
     this.isQueryResult = isQueryResult;
     this.jobID = jobID;
-    this.resultKind = resultKind;
+    this.resultKindEnum = resultKindEnum;
   }
 
   public ResultType getResultType() {
@@ -111,8 +111,8 @@ public class ResultSet implements Serializable {
     return jobID;
   }
 
-  public ResultKind getResultKind() {
-    return resultKind;
+  public ResultKindEnum getResultKind() {
+    return resultKindEnum;
   }
 
   @Override
@@ -130,12 +130,12 @@ public class ResultSet implements Serializable {
         && Objects.equals(columns, resultSet.columns)
         && Objects.equals(data, resultSet.data)
         && Objects.equals(jobID, resultSet.jobID)
-        && resultKind == resultSet.resultKind;
+        && resultKindEnum == resultSet.resultKindEnum;
   }
 
   @Override
   public int hashCode() {
-    return Objects.hash(resultType, nextToken, columns, data, isQueryResult, jobID, resultKind);
+    return Objects.hash(resultType, nextToken, columns, data, isQueryResult, jobID, resultKindEnum);
   }
 
   @Override
@@ -154,7 +154,7 @@ public class ResultSet implements Serializable {
         + ", jobID="
         + jobID
         + ", resultKind="
-        + resultKind
+        + resultKindEnum
         + '}';
   }
 
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableInfo.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableInfo.java
index b096edb57..8f3f46765 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableInfo.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableInfo.java
@@ -23,19 +23,19 @@ import java.util.Objects;
 /** Information of the table or view. */
 public class TableInfo implements Serializable {
   private final ObjectIdentifier identifier;
-  private final TableKind tableKind;
+  private final TableKindEnum tableKindEnum;
 
-  public TableInfo(ObjectIdentifier identifier, TableKind tableKind) {
+  public TableInfo(ObjectIdentifier identifier, TableKindEnum tableKindEnum) {
     this.identifier = identifier;
-    this.tableKind = tableKind;
+    this.tableKindEnum = tableKindEnum;
   }
 
   public ObjectIdentifier getIdentifier() {
     return identifier;
   }
 
-  public TableKind getTableKind() {
-    return tableKind;
+  public TableKindEnum getTableKind() {
+    return tableKindEnum;
   }
 
   @Override
@@ -47,16 +47,17 @@ public class TableInfo implements Serializable {
       return false;
     }
     TableInfo tableInfo = (TableInfo) o;
-    return Objects.equals(identifier, tableInfo.identifier) && tableKind == tableInfo.tableKind;
+    return Objects.equals(identifier, tableInfo.identifier)
+        && tableKindEnum == tableInfo.tableKindEnum;
   }
 
   @Override
   public int hashCode() {
-    return Objects.hash(identifier, tableKind);
+    return Objects.hash(identifier, tableKindEnum);
   }
 
   @Override
   public String toString() {
-    return "TableInfo{" + "identifier=" + identifier + ", tableKind=" + tableKind + '}';
+    return "TableInfo{" + "identifier=" + identifier + ", tableKind=" + tableKindEnum + '}';
   }
 }
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKind.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKindEnum.java
similarity index 96%
rename from streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKind.java
rename to streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKindEnum.java
index cd76a613f..af214466d 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKind.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/results/TableKindEnum.java
@@ -17,7 +17,7 @@
 
 package org.apache.streampark.gateway.results;
 
-public enum TableKind {
+public enum TableKindEnum {
   TABLE,
   VIEW
 }
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/service/SqlGatewayService.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/service/SqlGatewayService.java
index b355ac9dc..8519a1eb4 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/service/SqlGatewayService.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-base/src/main/java/org/apache/streampark/gateway/service/SqlGatewayService.java
@@ -19,7 +19,7 @@ package org.apache.streampark.gateway.service;
 
 import org.apache.streampark.gateway.ExecutionConfiguration;
 import org.apache.streampark.gateway.OperationHandle;
-import org.apache.streampark.gateway.OperationStatus;
+import org.apache.streampark.gateway.OperationStatusEnum;
 import org.apache.streampark.gateway.exception.SqlGatewayException;
 import org.apache.streampark.gateway.results.Column;
 import org.apache.streampark.gateway.results.GatewayInfo;
@@ -113,7 +113,7 @@ public interface SqlGatewayService {
    * Get the result schema for the specified Operation.
    *
    * <p>Note: The result schema is available when the Operation is in the {@link
-   * OperationStatus#FINISHED}.
+   * OperationStatusEnum#FINISHED}.
    *
    * @param sessionHandle handle to identify the session.
    * @param operationHandle handle to identify the operation.
diff --git a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-flink-v1/src/main/java/org/apache/streampark/gateway/flink/FlinkSqlGatewayImpl.java b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-flink-v1/src/main/java/org/apache/streampark/gateway/flink/FlinkSqlGatewayImpl.java
index 263cb73ab..059953aa8 100644
--- a/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-flink-v1/src/main/java/org/apache/streampark/gateway/flink/FlinkSqlGatewayImpl.java
+++ b/streampark-flink/streampark-flink-sql-gateway/streampark-flink-sql-gateway-flink-v1/src/main/java/org/apache/streampark/gateway/flink/FlinkSqlGatewayImpl.java
@@ -19,7 +19,7 @@ package org.apache.streampark.gateway.flink;
 
 import org.apache.streampark.gateway.ExecutionConfiguration;
 import org.apache.streampark.gateway.OperationHandle;
-import org.apache.streampark.gateway.OperationStatus;
+import org.apache.streampark.gateway.OperationStatusEnum;
 import org.apache.streampark.gateway.exception.SqlGatewayException;
 import org.apache.streampark.gateway.flink.client.dto.ExecuteStatementRequestBody;
 import org.apache.streampark.gateway.flink.client.dto.FetchResultsResponseBody;
@@ -34,7 +34,7 @@ import org.apache.streampark.gateway.flink.client.rest.v1.DefaultApi;
 import org.apache.streampark.gateway.results.Column;
 import org.apache.streampark.gateway.results.GatewayInfo;
 import org.apache.streampark.gateway.results.OperationInfo;
-import org.apache.streampark.gateway.results.ResultKind;
+import org.apache.streampark.gateway.results.ResultKindEnum;
 import org.apache.streampark.gateway.results.ResultQueryCondition;
 import org.apache.streampark.gateway.results.ResultSet;
 import org.apache.streampark.gateway.results.RowData;
@@ -146,7 +146,7 @@ public class FlinkSqlGatewayImpl implements SqlGatewayService {
           defaultApi.getOperationStatus(
               UUID.fromString(sessionHandle.getIdentifier()),
               UUID.fromString(operationHandle.getIdentifier()));
-      return new OperationInfo(OperationStatus.valueOf(operationStatus.getStatus()), null);
+      return new OperationInfo(OperationStatusEnum.valueOf(operationStatus.getStatus()), null);
     } catch (ApiException e) {
       throw new SqlGatewayException("Flink native SqlGateWay closeOperation failed!", e);
     }
@@ -219,10 +219,10 @@ public class FlinkSqlGatewayImpl implements SqlGatewayService {
 
       resultsData.forEach(row -> data.add(new RowData(row.getKind().getValue(), row.getFields())));
 
-      ResultKind resultKind =
+      ResultKindEnum resultKindEnum =
           columns.size() == 1 && columns.get(0).getName().equals("result")
-              ? ResultKind.SUCCESS
-              : ResultKind.SUCCESS_WITH_CONTENT;
+              ? ResultKindEnum.SUCCESS
+              : ResultKindEnum.SUCCESS_WITH_CONTENT;
 
       return new ResultSet(
           ResultSet.ResultType.valueOf(resultTypeStr),
@@ -231,7 +231,7 @@ public class FlinkSqlGatewayImpl implements SqlGatewayService {
           data,
           true,
           null,
-          resultKind);
+          resultKindEnum);
     } catch (ApiException e) {
       throw new SqlGatewayException("Flink native SqlGateWay fetchResults failed!", e);
     }