You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@dolphinscheduler.apache.org by GitBox <gi...@apache.org> on 2022/09/30 08:56:51 UTC

[GitHub] [dolphinscheduler] qq774724635 opened a new issue, #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing

qq774724635 opened a new issue, #12235:
URL: https://github.com/apache/dolphinscheduler/issues/12235

   ### Search before asking
   
   - [X] I had searched in the [issues](https://github.com/apache/dolphinscheduler/issues?q=is%3Aissue) and found no similar issues.
   
   
   ### What happened
   
   ### Deployment model:standalone
   
   ### Environment:
   `export JAVA_HOME=/usr/lib/jvm/java-1.8.0-openjdk-1.8.0.342.b07-1.el7_9.x86_64/jre
   
   # Database related configuration, set database type, username and password
   # export DATABASE=${DATABASE:-postgresql}
   # export SPRING_PROFILES_ACTIVE=${DATABASE}
   # export SPRING_DATASOURCE_URL
   # export SPRING_DATASOURCE_USERNAME
   # export SPRING_DATASOURCE_PASSWORD
   export DATABASE=mysql
   export SPRING_PROFILES_ACTIVE=mysql
   export SPRING_DATASOURCE_URL="jdbc:mysql://192.168.1.181:3306/dolphin?useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=Asia/Shanghai"
   export SPRING_DATASOURCE_USERNAME=bigData
   export SPRING_DATASOURCE_PASSWORD=P@ssw0rd654321
   export SPRING_JACKSON_TIME_ZONE=${SPRING_JACKSON_TIME_ZONE:-Asia/Shanghai}
   # DolphinScheduler server related configuration
   export SPRING_CACHE_TYPE=${SPRING_CACHE_TYPE:-none}
   export SPRING_JACKSON_TIME_ZONE=${SPRING_JACKSON_TIME_ZONE:-UTC}
   export MASTER_FETCH_COMMAND_NUM=${MASTER_FETCH_COMMAND_NUM:-10}
   
   # Registry center configuration, determines the type and link of the registry center
   export REGISTRY_TYPE=${REGISTRY_TYPE:-zookeeper}
   export REGISTRY_ZOOKEEPER_CONNECT_STRING=${REGISTRY_ZOOKEEPER_CONNECT_STRING:-localhost:2181}
   
   # Tasks related configurations, need to change the configuration if you use the related tasks.
   export HADOOP_HOME=/opt/hadoop-2.7.3
   export HADOOP_CONF_DIR=/opt/hadoop-2.7.3/etc/hadoop
   export SPARK_HOME1=/opt/spark-2.4.0
   export SPARK_HOME2=/opt/spark-2.4.0
   export PYTHON_HOME=/usr
   export HIVE_HOME=/opt/hive-2.3.7
   #export FLINK_HOME=${FLINK_HOME:-/opt/soft/flink}
   export DATAX_HOME=/opt/datax
   
   export PATH=$HADOOP_HOME/bin:$SPARK_HOME1/bin:$SPARK_HOME2/bin:$PYTHON_HOME/bin:$JAVA_HOME/bin:$HIVE_HOME/bin:$FLINK_HOME/bin:$DATAX_HOME/bin:$PATH`
   
   common.properties:
   `# user data local directory path, please make sure the directory exists and have read write permissions
   data.basedir.path=/tmp/dolphinscheduler
   
   # resource storage type: HDFS, S3, NONE
   resource.storage.type=HDFS
   
   resource.upload.path=/dolphinscheduler
   
   # whether to startup kerberos
   hadoop.security.authentication.startup.state=false
   
   # java.security.krb5.conf path
   java.security.krb5.conf.path=/opt/krb5.conf
   
   # login user from keytab username
   login.user.keytab.username=hdfs-mycluster@ESZ.COM
   
   # login user from keytab path
   login.user.keytab.path=/opt/hdfs.headless.keytab
   
   # kerberos expire time, the unit is hour
   kerberos.expire.time=2
   # resource view suffixs
   #resource.view.suffixs=txt,log,sh,bat,conf,cfg,py,java,sql,xml,hql,properties,json,yml,yaml,ini,js
   # if resource.storage.type=HDFS, the user must have the permission to create directories under the HDFS root path
   hdfs.root.user=hdfs
   fs.defaultFS=file:///
   aws.access.key.id=minioadmin
   aws.secret.access.key=minioadmin
   aws.region=us-east-1
   aws.endpoint=http://localhost:9000
   # resourcemanager port, the default value is 8088 if not specified
   resource.manager.httpaddress.port=7777
   # if resourcemanager HA is enabled, please set the HA IPs; if resourcemanager is single, keep this value empty
   #yarn.resourcemanager.ha.rm.ids=192.168.xx.xx,192.168.xx.xx
   # if resourcemanager HA is enabled or not use resourcemanager, please keep the default value; If resourcemanager is single, you only need to replace ds1 to actual resourcemanager hostname
   yarn.application.status.address=http://192.168.1.181:%s/ws/v1/cluster/apps/%s
   # job history status url when application number threshold is reached(default 10000, maybe it was set to 1000)
   yarn.job.history.status.address=http://192.168.1.181:19888/ws/v1/history/mapreduce/jobs/%s
   
   # datasource encryption enable
   datasource.encryption.enable=false
   
   # datasource encryption salt
   datasource.encryption.salt=!@#$%^&*
   
   # data quality option
   data-quality.jar.name=dolphinscheduler-data-quality-dev-SNAPSHOT.jar
   
   #data-quality.error.output.path=/tmp/data-quality-error-data
   
   # Network IP gets priority, default inner outer
   
   # Whether hive SQL is executed in the same session
   support.hive.oneSession=false
   
   # use sudo or not, if set true, executing user is tenant user and deploy user needs sudo permissions; if set false, executing user is the deploy user and doesn't need sudo permissions
   sudo.enable=true
   
   # network interface preferred like eth0, default: empty
   #dolphin.scheduler.network.interface.preferred=
   
   # network IP gets priority, default: inner outer
   #dolphin.scheduler.network.priority.strategy=default
   
   # system env path
   #dolphinscheduler.env.path=dolphinscheduler_env.sh
   
   # development state
   development.state=false
   
   # rpc port
   alert.rpc.port=50052
   
   # Url endpoint for zeppelin RESTful API
   zeppelin.rest.url=http://localhost:8080`
   
   ### scene:
   After deploying spark, hadoop, hive, etc. and configuring spark to connect to hive metadata according to the information provided on the official website, the following error is always reported when performing data quality checks
   
   
   ### error:
   `[LOG-PATH]: /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/logs/20220930/7052280036032_2-376-859.log, [HOST]:  Host{address='192.168.1.181:1234', ip='192.168.1.181', port=1234}
   [INFO] 2022-09-30 16:43:59.405 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[83] - [WorkflowInstance-376][TaskInstance-859] - data quality task params {"localParams":[],"resourceList":[],"ruleId":6,"ruleInputParameter":{"check_type":"0","failure_strategy":"0","operator":"0","src_connector_type":0,"src_datasource_id":2,"src_field":"ID","src_table":"DIM_BE_MAINTAIN_FRT","threshold":"0"},"sparkParameters":{"deployMode":"client","driverCores":1,"driverMemory":"512M","executorCores":2,"executorMemory":"2G","numExecutors":2,"others":"--conf spark.yarn.maxAppAttempts=1 --driver-class-path /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/standalone-server/mysql-connector-java-8.0.29.jar"}}
   [INFO] 2022-09-30 16:43:59.413 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[181] - [WorkflowInstance-376][TaskInstance-859] - data quality task command: ${SPARK_HOME2}/bin/spark-submit --master yarn --deploy-mode client --driver-cores 1 --driver-memory 512M --num-executors 2 --executor-cores 2 --executor-memory 2G --queue default --conf spark.yarn.maxAppAttempts=1 --driver-class-path /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/standalone-server/mysql-connector-java-8.0.29.jar /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/dolphinscheduler-data-quality-dev-SNAPSHOT.jar "{\"name\":\"$t(uniqueness_check)\",\"env\":{\"type\":\"batch\",\"config\":null},\"readers\":[{\"type\":\"JDBC\",\"config\":{\"database\":\"warehouse_new\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"output_table\":\"warehouse_new_DIM_BE_MAINTAI
 N_FRT\",\"table\":\"DIM_BE_MAINTAIN_FRT\",\"url\":\"jdbc:mysql://192.168.1.181:3306/warehouse_new?allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\"} }],\"transformers\":[{\"type\":\"sql\",\"config\":{\"index\":1,\"output_table\":\"duplicate_items\",\"sql\":\"SELECT ID FROM warehouse_new_DIM_BE_MAINTAIN_FRT group by ID having count(*) > 1\"} },{\"type\":\"sql\",\"config\":{\"index\":2,\"output_table\":\"duplicate_count\",\"sql\":\"SELECT COUNT(*) AS duplicates FROM duplicate_items\"} }],\"writers\":[{\"type\":\"JDBC\",\"config\":{\"database\":\"dolphin\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"table\":\"t_ds_dq_execute_result\",\"url\":\"jdbc:mysql://192.168.1.181:3306/dolphin?useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=Asia/Shanghai&allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\",\"sql\":\"select 0 as
  rule_type,'$t(uniqueness_check)' as rule_name,0 as process_definition_id,376 as process_instance_id,859 as task_instance_id,duplicate_count.duplicates AS statistics_value,${comparison_name} AS comparison_value, AS comparison_type,0 as check_type,0 as threshold,0 as operator,0 as failure_strategy,'file:////user/admin/data_quality_error_data/0_376_test' as error_output_path,'2022-09-30 16:43:59' as create_time,'2022-09-30 16:43:59' as update_time from duplicate_count full join ${comparison_table}\"} },{\"type\":\"JDBC\",\"config\":{\"database\":\"dolphin\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"table\":\"t_ds_dq_task_statistics_value\",\"url\":\"jdbc:mysql://192.168.1.181:3306/dolphin?useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=Asia/Shanghai&allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\",\"sql\":\"select 0 as process_definition_id,859 as task_instance
 _id,6 as rule_id,'AOXATF9DIPSNL1TLP/YTDGK8TIRC3L+OUXFM8LSTGRK=' as unique_code,'duplicate_count.duplicates'AS statistics_name,duplicate_count.duplicates AS statistics_value,'2022-09-30 16:43:59' as data_time,'2022-09-30 16:43:59' as create_time,'2022-09-30 16:43:59' as update_time from duplicate_count\"} },{\"type\":\"hdfs_file\",\"config\":{\"path\":\"file:////user/admin/data_quality_error_data/0_376_test\",\"input_table\":\"duplicate_items\"} }]}"
   [INFO] 2022-09-30 16:43:59.414 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[85] - [WorkflowInstance-376][TaskInstance-859] - tenantCode user:admin, task dir:376_859
   [INFO] 2022-09-30 16:43:59.414 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[90] - [WorkflowInstance-376][TaskInstance-859] - create command file:/tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859/376_859.command
   [INFO] 2022-09-30 16:43:59.414 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[116] - [WorkflowInstance-376][TaskInstance-859] - command : #!/bin/sh
   BASEDIR=$(cd `dirname $0`; pwd)
   cd $BASEDIR
   source /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/conf/dolphinscheduler_env.sh
   ${SPARK_HOME2}/bin/spark-submit --master yarn --deploy-mode client --driver-cores 1 --driver-memory 512M --num-executors 2 --executor-cores 2 --executor-memory 2G --queue default --conf spark.yarn.maxAppAttempts=1 --driver-class-path /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/standalone-server/mysql-connector-java-8.0.29.jar /opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/dolphinscheduler-data-quality-dev-SNAPSHOT.jar "{\"name\":\"$t(uniqueness_check)\",\"env\":{\"type\":\"batch\",\"config\":null},\"readers\":[{\"type\":\"JDBC\",\"config\":{\"database\":\"warehouse_new\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"output_table\":\"warehouse_new_DIM_BE_MAINTAIN_FRT\",\"table\":\"DIM_BE_MAINTAIN_FRT\",\"url\":\"jdbc:mysql://192.168.1.181:3306/warehouse_new?allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\"} }],\"transformers\":[{\"type\":\"sql\",\"co
 nfig\":{\"index\":1,\"output_table\":\"duplicate_items\",\"sql\":\"SELECT ID FROM warehouse_new_DIM_BE_MAINTAIN_FRT group by ID having count(*) > 1\"} },{\"type\":\"sql\",\"config\":{\"index\":2,\"output_table\":\"duplicate_count\",\"sql\":\"SELECT COUNT(*) AS duplicates FROM duplicate_items\"} }],\"writers\":[{\"type\":\"JDBC\",\"config\":{\"database\":\"dolphin\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"table\":\"t_ds_dq_execute_result\",\"url\":\"jdbc:mysql://192.168.1.181:3306/dolphin?useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=Asia/Shanghai&allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\",\"sql\":\"select 0 as rule_type,'$t(uniqueness_check)' as rule_name,0 as process_definition_id,376 as process_instance_id,859 as task_instance_id,duplicate_count.duplicates AS statistics_value,${comparison_name} AS comparison_value, AS comparison_type,0 as check_ty
 pe,0 as threshold,0 as operator,0 as failure_strategy,'file:////user/admin/data_quality_error_data/0_376_test' as error_output_path,'2022-09-30 16:43:59' as create_time,'2022-09-30 16:43:59' as update_time from duplicate_count full join ${comparison_table}\"} },{\"type\":\"JDBC\",\"config\":{\"database\":\"dolphin\",\"password\":\"**************\",\"driver\":\"com.mysql.cj.jdbc.Driver\",\"user\":\"bigData\",\"table\":\"t_ds_dq_task_statistics_value\",\"url\":\"jdbc:mysql://192.168.1.181:3306/dolphin?useUnicode=true&characterEncoding=UTF-8&useSSL=false&serverTimezone=Asia/Shanghai&allowLoadLocalInfile=false&autoDeserialize=false&allowLocalInfile=false&allowUrlInLocalInfile=false\",\"sql\":\"select 0 as process_definition_id,859 as task_instance_id,6 as rule_id,'AOXATF9DIPSNL1TLP/YTDGK8TIRC3L+OUXFM8LSTGRK=' as unique_code,'duplicate_count.duplicates'AS statistics_name,duplicate_count.duplicates AS statistics_value,'2022-09-30 16:43:59' as data_time,'2022-09-30 16:43:59' as create_time
 ,'2022-09-30 16:43:59' as update_time from duplicate_count\"} },{\"type\":\"hdfs_file\",\"config\":{\"path\":\"file:////user/admin/data_quality_error_data/0_376_test\",\"input_table\":\"duplicate_items\"} }]}"
   [INFO] 2022-09-30 16:43:59.421 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[290] - [WorkflowInstance-376][TaskInstance-859] - task run command: sudo -u admin sh /tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859/376_859.command
   [INFO] 2022-09-30 16:43:59.422 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[181] - [WorkflowInstance-376][TaskInstance-859] - process start, process id is: 1788
   [INFO] 2022-09-30 16:44:01.422 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:00 WARN  NativeCodeLoader:62 - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
   [INFO] 2022-09-30 16:44:02.425 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:01 INFO  SparkContext:54 - Running Spark version 2.4.0
   	2022-09-30 16:44:01 INFO  SparkContext:54 - Submitted application: (uniqueness_check)
   	2022-09-30 16:44:01 INFO  SecurityManager:54 - Changing view acls to: admin
   	2022-09-30 16:44:01 INFO  SecurityManager:54 - Changing modify acls to: admin
   	2022-09-30 16:44:01 INFO  SecurityManager:54 - Changing view acls groups to: 
   	2022-09-30 16:44:01 INFO  SecurityManager:54 - Changing modify acls groups to: 
   	2022-09-30 16:44:01 INFO  SecurityManager:54 - SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(admin); groups with view permissions: Set(); users  with modify permissions: Set(admin); groups with modify permissions: Set()
   	2022-09-30 16:44:02 INFO  Utils:54 - Successfully started service 'sparkDriver' on port 35638.
   	2022-09-30 16:44:02 INFO  SparkEnv:54 - Registering MapOutputTracker
   	2022-09-30 16:44:02 INFO  SparkEnv:54 - Registering BlockManagerMaster
   	2022-09-30 16:44:02 INFO  BlockManagerMasterEndpoint:54 - Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
   	2022-09-30 16:44:02 INFO  BlockManagerMasterEndpoint:54 - BlockManagerMasterEndpoint up
   	2022-09-30 16:44:02 INFO  DiskBlockManager:54 - Created local directory at /tmp/blockmgr-0905802e-97fd-4924-9877-24fae1ba490c
   	2022-09-30 16:44:02 INFO  MemoryStore:54 - MemoryStore started with capacity 93.3 MB
   	2022-09-30 16:44:02 INFO  SparkEnv:54 - Registering OutputCommitCoordinator
   	2022-09-30 16:44:02 INFO  log:192 - Logging initialized @2662ms
   [INFO] 2022-09-30 16:44:03.426 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:02 INFO  Server:351 - jetty-9.3.z-SNAPSHOT, build timestamp: unknown, git hash: unknown
   	2022-09-30 16:44:02 INFO  Server:419 - Started @2777ms
   	2022-09-30 16:44:02 INFO  AbstractConnector:278 - Started ServerConnector@52fc5eb1{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   	2022-09-30 16:44:02 INFO  Utils:54 - Successfully started service 'SparkUI' on port 4040.
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@322803db{/jobs,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@1bbae752{/jobs/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@460b6d54{/jobs/job,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@76075d65{/jobs/job/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@3a4ba480{/stages,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@27b71f50{/stages/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@383790cf{/stages/stage,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@ccd1bc3{/stages/stage/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@878537d{/stages/pool,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@4455f57d{/stages/pool/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@29fc1a2b{/storage,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@4d0b0fd4{/storage/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@7a24eb3{/storage/rdd,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@6c37bd27{/storage/rdd/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@25d3cfc8{/environment,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@30331109{/environment/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@2571066a{/executors,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@74fe5966{/executors/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@4fe875be{/executors/threadDump,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@677b8e13{/executors/threadDump/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@4a9486c0{/static,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@562c877a{/,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@67001148{/api,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@64030b91{/jobs/job/kill,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@2032e725{/stages/stage/kill,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:02 INFO  SparkUI:54 - Bound SparkUI to 0.0.0.0, and started at http://enoch:4040
   	2022-09-30 16:44:02 INFO  SparkContext:54 - Added JAR file:/opt/apache-dolphinscheduler-3.0.0-bin/standalone-server/libs/dolphinscheduler-data-quality-dev-SNAPSHOT.jar at spark://enoch:35638/jars/dolphinscheduler-data-quality-dev-SNAPSHOT.jar with timestamp 1664527442612
   	2022-09-30 16:44:03 INFO  RMProxy:98 - Connecting to ResourceManager at /0.0.0.0:8032
   [INFO] 2022-09-30 16:44:04.427 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:03 INFO  Client:54 - Requesting a new application from cluster with 1 NodeManagers
   	2022-09-30 16:44:03 INFO  Client:54 - Verifying our application has not requested more than the maximum memory capability of the cluster (8192 MB per container)
   	2022-09-30 16:44:03 INFO  Client:54 - Will allocate AM container, with 896 MB memory including 384 MB overhead
   	2022-09-30 16:44:03 INFO  Client:54 - Setting up container launch context for our AM
   	2022-09-30 16:44:03 INFO  Client:54 - Setting up the launch environment for our AM container
   	2022-09-30 16:44:03 INFO  Client:54 - Preparing resources for our AM container
   	2022-09-30 16:44:03 WARN  Client:66 - Neither spark.yarn.jars nor spark.yarn.archive is set, falling back to uploading libraries under SPARK_HOME.
   [INFO] 2022-09-30 16:44:09.428 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:09 INFO  Client:54 - Uploading resource file:/tmp/spark-ed1d373e-5e95-4d3b-b97b-0fa14e0266c9/__spark_libs__1561782813166038027.zip -> hdfs://192.168.1.181:54310/user/admin/.sparkStaging/application_1664519295996_0007/__spark_libs__1561782813166038027.zip
   [INFO] 2022-09-30 16:44:11.429 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:10 INFO  Client:54 - Uploading resource file:/tmp/spark-ed1d373e-5e95-4d3b-b97b-0fa14e0266c9/__spark_conf__5327828120016222911.zip -> hdfs://192.168.1.181:54310/user/admin/.sparkStaging/application_1664519295996_0007/__spark_conf__.zip
   	2022-09-30 16:44:10 INFO  SecurityManager:54 - Changing view acls to: admin
   	2022-09-30 16:44:10 INFO  SecurityManager:54 - Changing modify acls to: admin
   	2022-09-30 16:44:10 INFO  SecurityManager:54 - Changing view acls groups to: 
   	2022-09-30 16:44:10 INFO  SecurityManager:54 - Changing modify acls groups to: 
   	2022-09-30 16:44:10 INFO  SecurityManager:54 - SecurityManager: authentication disabled; ui acls disabled; users  with view permissions: Set(admin); groups with view permissions: Set(); users  with modify permissions: Set(admin); groups with modify permissions: Set()
   [INFO] 2022-09-30 16:44:12.430 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:11 INFO  Client:54 - Submitting application application_1664519295996_0007 to ResourceManager
   	2022-09-30 16:44:11 INFO  YarnClientImpl:273 - Submitted application application_1664519295996_0007
   	2022-09-30 16:44:11 INFO  SchedulerExtensionServices:54 - Starting Yarn extension services with app application_1664519295996_0007 and attemptId None
   [INFO] 2022-09-30 16:44:13.431 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:12 INFO  Client:54 - Application report for application_1664519295996_0007 (state: ACCEPTED)
   	2022-09-30 16:44:12 INFO  Client:54 - 
   		 client token: N/A
   		 diagnostics: N/A
   		 ApplicationMaster host: N/A
   		 ApplicationMaster RPC port: -1
   		 queue: default
   		 start time: 1664527451736
   		 final status: UNDEFINED
   		 tracking URL: http://enoch:7777/proxy/application_1664519295996_0007/
   		 user: admin
   [INFO] 2022-09-30 16:44:14.432 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:13 INFO  Client:54 - Application report for application_1664519295996_0007 (state: ACCEPTED)
   [INFO] 2022-09-30 16:44:15.433 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:14 INFO  Client:54 - Application report for application_1664519295996_0007 (state: ACCEPTED)
   [INFO] 2022-09-30 16:44:16.434 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:15 INFO  Client:54 - Application report for application_1664519295996_0007 (state: ACCEPTED)
   [INFO] 2022-09-30 16:44:17.435 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:16 INFO  Client:54 - Application report for application_1664519295996_0007 (state: ACCEPTED)
   	2022-09-30 16:44:17 INFO  YarnClientSchedulerBackend:54 - Add WebUI Filter. org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter, Map(PROXY_HOSTS -> enoch, PROXY_URI_BASES -> http://enoch:7777/proxy/application_1664519295996_0007), /proxy/application_1664519295996_0007
   	2022-09-30 16:44:17 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /jobs, /jobs/json, /jobs/job, /jobs/job/json, /stages, /stages/json, /stages/stage, /stages/stage/json, /stages/pool, /stages/pool/json, /storage, /storage/json, /storage/rdd, /storage/rdd/json, /environment, /environment/json, /executors, /executors/json, /executors/threadDump, /executors/threadDump/json, /static, /, /api, /jobs/job/kill, /stages/stage/kill.
   [INFO] 2022-09-30 16:44:18.436 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:17 INFO  YarnSchedulerBackend$YarnSchedulerEndpoint:54 - ApplicationMaster registered as NettyRpcEndpointRef(spark-client://YarnAM)
   	2022-09-30 16:44:17 INFO  Client:54 - Application report for application_1664519295996_0007 (state: RUNNING)
   	2022-09-30 16:44:17 INFO  Client:54 - 
   		 client token: N/A
   		 diagnostics: N/A
   		 ApplicationMaster host: 192.168.1.181
   		 ApplicationMaster RPC port: -1
   		 queue: default
   		 start time: 1664527451736
   		 final status: UNDEFINED
   		 tracking URL: http://enoch:7777/proxy/application_1664519295996_0007/
   		 user: admin
   	2022-09-30 16:44:17 INFO  YarnClientSchedulerBackend:54 - Application application_1664519295996_0007 has started running.
   	2022-09-30 16:44:17 INFO  Utils:54 - Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 44485.
   	2022-09-30 16:44:17 INFO  NettyBlockTransferService:54 - Server created on enoch:44485
   	2022-09-30 16:44:17 INFO  BlockManager:54 - Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
   	2022-09-30 16:44:17 INFO  BlockManagerMaster:54 - Registering BlockManager BlockManagerId(driver, enoch, 44485, None)
   	2022-09-30 16:44:17 INFO  BlockManagerMasterEndpoint:54 - Registering block manager enoch:44485 with 93.3 MB RAM, BlockManagerId(driver, enoch, 44485, None)
   	2022-09-30 16:44:17 INFO  BlockManagerMaster:54 - Registered BlockManager BlockManagerId(driver, enoch, 44485, None)
   	2022-09-30 16:44:17 INFO  BlockManager:54 - Initialized BlockManager: BlockManagerId(driver, enoch, 44485, None)
   	2022-09-30 16:44:17 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /metrics/json.
   	2022-09-30 16:44:17 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@27e44e9c{/metrics/json,null,AVAILABLE,@Spark}
   [INFO] 2022-09-30 16:44:22.437 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:21 INFO  YarnSchedulerBackend$YarnDriverEndpoint:54 - Registered executor NettyRpcEndpointRef(spark-client://Executor) (192.168.1.181:58620) with ID 1
   	2022-09-30 16:44:21 INFO  BlockManagerMasterEndpoint:54 - Registering block manager enoch:34122 with 912.3 MB RAM, BlockManagerId(1, enoch, 34122, None)
   [INFO] 2022-09-30 16:44:23.438 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:23 INFO  YarnSchedulerBackend$YarnDriverEndpoint:54 - Registered executor NettyRpcEndpointRef(spark-client://Executor) (192.168.1.181:58628) with ID 2
   	2022-09-30 16:44:23 INFO  YarnClientSchedulerBackend:54 - SchedulerBackend is ready for scheduling beginning after reached minRegisteredResourcesRatio: 0.8
   [INFO] 2022-09-30 16:44:24.439 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:23 INFO  SharedState:54 - loading hive config file: file:/opt/spark-2.4.0/conf/hive-site.xml
   	2022-09-30 16:44:23 INFO  BlockManagerMasterEndpoint:54 - Registering block manager enoch:40145 with 912.3 MB RAM, BlockManagerId(2, enoch, 40145, None)
   	2022-09-30 16:44:23 INFO  SharedState:54 - Setting hive.metastore.warehouse.dir ('null') to the value of spark.sql.warehouse.dir ('file:/tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859/spark-warehouse').
   	2022-09-30 16:44:23 INFO  SharedState:54 - Warehouse path is 'file:/tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859/spark-warehouse'.
   	2022-09-30 16:44:23 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /SQL.
   	2022-09-30 16:44:23 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@1068176{/SQL,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:23 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /SQL/json.
   	2022-09-30 16:44:23 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@71789580{/SQL/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:23 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /SQL/execution.
   	2022-09-30 16:44:23 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@9d3d54e{/SQL/execution,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:23 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /SQL/execution/json.
   	2022-09-30 16:44:23 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@333398f{/SQL/execution/json,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:23 INFO  JettyUtils:54 - Adding filter org.apache.hadoop.yarn.server.webproxy.amfilter.AmIpFilter to /static/sql.
   	2022-09-30 16:44:23 INFO  ContextHandler:781 - Started o.s.j.s.ServletContextHandler@77896335{/static/sql,null,AVAILABLE,@Spark}
   	2022-09-30 16:44:24 INFO  StateStoreCoordinatorRef:54 - Registered StateStoreCoordinator endpoint
   [INFO] 2022-09-30 16:44:26.440 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:25 INFO  HiveUtils:54 - Initializing HiveMetastoreConnection version 1.2.1 using Spark classes.
   	2022-09-30 16:44:26 INFO  metastore:376 - Trying to connect to metastore with URI thrift://192.168.1.181:9083
   	2022-09-30 16:44:26 INFO  metastore:472 - Connected to metastore.
   	2022-09-30 16:44:26 INFO  SessionState:641 - Created local directory: /tmp/admin
   	2022-09-30 16:44:26 INFO  SessionState:641 - Created local directory: /tmp/50f5c351-bbbf-48fc-90fb-375a4ac8bbf2_resources
   [INFO] 2022-09-30 16:44:27.228 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[375] - [WorkflowInstance-376][TaskInstance-859] - find app id: application_1664519295996_0007
   [INFO] 2022-09-30 16:44:27.228 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[205] - [WorkflowInstance-376][TaskInstance-859] - process has exited, execute path:/tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859, processId:1788 ,exitStatusCode:1 ,processWaitForStatus:true ,processExitValue:1
   [INFO] 2022-09-30 16:44:27.440 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[63] - [WorkflowInstance-0][TaskInstance-0] -  -> 2022-09-30 16:44:26 INFO  SessionState:641 - Created HDFS directory: /tmp/hive/admin/50f5c351-bbbf-48fc-90fb-375a4ac8bbf2
   	2022-09-30 16:44:26 INFO  SessionState:641 - Created local directory: /tmp/admin/50f5c351-bbbf-48fc-90fb-375a4ac8bbf2
   	2022-09-30 16:44:26 INFO  SessionState:641 - Created HDFS directory: /tmp/hive/admin/50f5c351-bbbf-48fc-90fb-375a4ac8bbf2/_tmp_space.db
   	2022-09-30 16:44:26 INFO  HiveClientImpl:54 - Warehouse location for Hive client (version 1.2.2) is file:/tmp/dolphinscheduler/exec/process/6883044913472/7052280036032_2/376/859/spark-warehouse
   	Exception in thread "main" org.apache.spark.sql.catalyst.parser.ParseException: 
   	mismatched input 'from' expecting <EOF>(line 1, pos 441)
   	
   	== SQL ==
   	select 0 as rule_type,'(uniqueness_check)' as rule_name,0 as process_definition_id,376 as process_instance_id,859 as task_instance_id,duplicate_count.duplicates AS statistics_value, AS comparison_value, AS comparison_type,0 as check_type,0 as threshold,0 as operator,0 as failure_strategy,'file:////user/admin/data_quality_error_data/0_376_test' as error_output_path,'2022-09-30 16:43:59' as create_time,'2022-09-30 16:43:59' as update_time from duplicate_count full join 
   	---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------^^^
   	
   		at org.apache.spark.sql.catalyst.parser.ParseException.withCommand(ParseDriver.scala:241)
   		at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parse(ParseDriver.scala:117)
   		at org.apache.spark.sql.execution.SparkSqlParser.parse(SparkSqlParser.scala:48)
   		at org.apache.spark.sql.catalyst.parser.AbstractSqlParser.parsePlan(ParseDriver.scala:69)
   		at org.apache.spark.sql.SparkSession.sql(SparkSession.scala:642)
   		at org.apache.dolphinscheduler.data.quality.flow.batch.writer.JdbcWriter.write(JdbcWriter.java:74)
   		at org.apache.dolphinscheduler.data.quality.execution.SparkBatchExecution.executeWriter(SparkBatchExecution.java:130)
   		at org.apache.dolphinscheduler.data.quality.execution.SparkBatchExecution.execute(SparkBatchExecution.java:58)
   		at org.apache.dolphinscheduler.data.quality.context.DataQualityContext.execute(DataQualityContext.java:62)
   		at org.apache.dolphinscheduler.data.quality.DataQualityApplication.main(DataQualityApplication.java:70)
   		at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
   		at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
   		at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   		at java.lang.reflect.Method.invoke(Method.java:498)
   		at org.apache.spark.deploy.JavaMainApplication.start(SparkApplication.scala:52)
   		at org.apache.spark.deploy.SparkSubmit.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:849)
   		at org.apache.spark.deploy.SparkSubmit.doRunMain$1(SparkSubmit.scala:167)
   		at org.apache.spark.deploy.SparkSubmit.submit(SparkSubmit.scala:195)
   		at org.apache.spark.deploy.SparkSubmit.doSubmit(SparkSubmit.scala:86)
   		at org.apache.spark.deploy.SparkSubmit$$anon$2.doSubmit(SparkSubmit.scala:924)
   		at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:933)
   		at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
   	2022-09-30 16:44:26 INFO  SparkContext:54 - Invoking stop() from shutdown hook
   	2022-09-30 16:44:26 INFO  AbstractConnector:318 - Stopped Spark@52fc5eb1{HTTP/1.1,[http/1.1]}{0.0.0.0:4040}
   	2022-09-30 16:44:26 INFO  SparkUI:54 - Stopped Spark web UI at http://enoch:4040
   	2022-09-30 16:44:26 INFO  YarnClientSchedulerBackend:54 - Interrupting monitor thread
   	2022-09-30 16:44:26 INFO  YarnClientSchedulerBackend:54 - Shutting down all executors
   	2022-09-30 16:44:26 INFO  YarnSchedulerBackend$YarnDriverEndpoint:54 - Asking each executor to shut down
   	2022-09-30 16:44:26 INFO  SchedulerExtensionServices:54 - Stopping SchedulerExtensionServices
   	(serviceOption=None,
   	 services=List(),
   	 started=false)
   	2022-09-30 16:44:26 INFO  YarnClientSchedulerBackend:54 - Stopped
   	2022-09-30 16:44:26 INFO  MapOutputTrackerMasterEndpoint:54 - MapOutputTrackerMasterEndpoint stopped!
   	2022-09-30 16:44:26 INFO  MemoryStore:54 - MemoryStore cleared
   	2022-09-30 16:44:26 INFO  BlockManager:54 - BlockManager stopped
   	2022-09-30 16:44:26 INFO  BlockManagerMaster:54 - BlockManagerMaster stopped
   	2022-09-30 16:44:26 INFO  OutputCommitCoordinator$OutputCommitCoordinatorEndpoint:54 - OutputCommitCoordinator stopped!
   	2022-09-30 16:44:26 INFO  SparkContext:54 - Successfully stopped SparkContext
   	2022-09-30 16:44:26 INFO  ShutdownHookManager:54 - Shutdown hook called
   	2022-09-30 16:44:26 INFO  ShutdownHookManager:54 - Deleting directory /tmp/spark-0d51c855-2d04-4883-a8c7-7b0d2afc31f6
   	2022-09-30 16:44:26 INFO  ShutdownHookManager:54 - Deleting directory /tmp/spark-ed1d373e-5e95-4d3b-b97b-0fa14e0266c9
   [INFO] 2022-09-30 16:44:27.442 +0800 [taskAppId=TASK-20220930-7052280036032_2-376-859] TaskLogLogger-class org.apache.dolphinscheduler.plugin.task.dq.DataQualityTask:[57] - [WorkflowInstance-0][TaskInstance-0] - FINALIZE_SESSION`
   
   ### What you expected to happen
   
   How to start a data quality check task properly
   
   ### How to reproduce
   
   Perform data quality checks in a stand-alone manner
   
   ### Anything else
   
   _No response_
   
   ### Version
   
   3.0.x
   
   ### Are you willing to submit PR?
   
   - [X] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of Conduct](https://www.apache.org/foundation/policies/conduct)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


[GitHub] [dolphinscheduler] github-actions[bot] commented on issue #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing

Posted by GitBox <gi...@apache.org>.
github-actions[bot] commented on issue #12235:
URL: https://github.com/apache/dolphinscheduler/issues/12235#issuecomment-1296397896

   This issue has been automatically marked as stale because it has not had recent activity for 30 days. It will be closed in next 7 days if no further activity occurs.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


[GitHub] [dolphinscheduler] github-actions[bot] commented on issue #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing

Posted by GitBox <gi...@apache.org>.
github-actions[bot] commented on issue #12235:
URL: https://github.com/apache/dolphinscheduler/issues/12235#issuecomment-1263308796

   Thank you for your feedback, we have received your issue, Please wait patiently for a reply.
   * In order for us to understand your request as soon as possible, please provide detailed information、version or pictures.
   * If you haven't received a reply for a long time, you can [join our slack](https://s.apache.org/dolphinscheduler-slack) and send your question to channel `#troubleshooting`


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


[GitHub] [dolphinscheduler] github-actions[bot] commented on issue #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing

Posted by GitBox <gi...@apache.org>.
github-actions[bot] commented on issue #12235:
URL: https://github.com/apache/dolphinscheduler/issues/12235#issuecomment-1306413254

   This issue has been closed because it has not received response for too long time. You could reopen it if you encountered similar problems in the future.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org


[GitHub] [dolphinscheduler] github-actions[bot] closed issue #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing

Posted by GitBox <gi...@apache.org>.
github-actions[bot] closed issue #12235: [Bug] When I perform a quality check, the sql executed by the system always has a missing
URL: https://github.com/apache/dolphinscheduler/issues/12235


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@dolphinscheduler.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org