You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "Kaxil Naik (Jira)" <ji...@apache.org> on 2020/04/09 22:09:00 UTC

[jira] [Resolved] (AIRFLOW-6515) log level of INFO/WARN when ERROR happened

     [ https://issues.apache.org/jira/browse/AIRFLOW-6515?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Kaxil Naik resolved AIRFLOW-6515.
---------------------------------
    Fix Version/s: 1.10.11
       Resolution: Fixed

> log level of INFO/WARN when ERROR happened
> ------------------------------------------
>
>                 Key: AIRFLOW-6515
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-6515
>             Project: Apache Airflow
>          Issue Type: Bug
>          Components: logging
>    Affects Versions: 1.10.7
>            Reporter: t oo
>            Assignee: Will Hudgins
>            Priority: Major
>             Fix For: 1.10.11
>
>
> log level should be error on some (but there are false positives):
> grep -iE 'log\.(info|warn).*(error|exceptio|fail|unab|couldn|lost|gone|missing|not fou|abort|exit|could not)' -R *
> airflow/sensors/base_sensor_operator.py:        self.log.info("Success criteria met. Exiting.")
> airflow/logging_config.py:        log.warning('Unable to load the config, contains a configuration error.')
> airflow/operators/check_operator.py:            self.log.warning("The following %s tests out of %s failed:", j, n)
> airflow/operators/sql_to_gcs.py:            self.log.warning('Using default schema due to missing name or type. Please '
> airflow/operators/bash_operator.py:            self.log.info('Command exited with return code %s', self.sub_process.returncode)
> airflow/serialization/serialized_objects.py:            LOG.warning('Failed to stringify.', exc_info=True)
> airflow/providers/amazon/aws/operators/batch.py:            self.log.info("AWS Batch Job has failed")
> airflow/providers/amazon/aws/hooks/s3.py:            self.log.info(e.response["Error"]["Message"])
> airflow/providers/amazon/aws/hooks/s3.py:            self.log.info(e.response["Error"]["Message"])
> airflow/utils/dag_processing.py:        self.log.info("Exiting gracefully upon receiving signal %s", signum)
> airflow/utils/dag_processing.py:                self.log.info("Exiting dag parsing loop as all files "
> airflow/utils/dag_processing.py:            self.log.info("Failing jobs without heartbeat after %s", limit_dttm)
> airflow/utils/dag_processing.py:            self.log.info("Waiting up to %s seconds for processes to exit...", timeout)
> airflow/utils/helpers.py:        log.info("Process %s (%s) terminated with exit code %s", p, p.pid, p.returncode)
> airflow/models/dagrun.py:            self.log.info('Marking run %s failed', self)
> airflow/models/dagrun.py:            self.log.info('Deadlock; marking run %s failed', self)
> airflow/models/dagrun.py:                    self.log.warning("Failed to get task '{}' for dag '{}'. "
> airflow/gcp/sensors/gcs.py:            self.log.warning("FAILURE: Inactivity Period passed, not enough objects found in %s", path)
> airflow/gcp/operators/spanner.py:            self.log.info("The Cloud Spanner database was missing: "
> airflow/gcp/hooks/kubernetes_engine.py:            self.log.info('Assuming Success: %s', error.message)
> airflow/gcp/hooks/kubernetes_engine.py:            self.log.info('Assuming Success: %s', error.message)
> airflow/gcp/hooks/cloud_memorystore.py:        self.log.info("Failovering Instance: %s", name)
> airflow/gcp/hooks/cloud_memorystore.py:        self.log.info("Instance failovered: %s", name)
> airflow/gcp/hooks/bigquery.py:            self.log.info(error_msg)
> airflow/gcp/hooks/bigtable.py:            self.log.info("The instance '%s' does not exist in project '%s'. Exiting", instance_id,
> airflow/contrib/sensors/bash_sensor.py:                self.log.info("Command exited with return code %s", sp.returncode)
> airflow/contrib/sensors/ftp_sensor.py:                self.log.info('Ftp error encountered: %s', str(e))
> airflow/contrib/operators/azure_container_instances_operator.py:            self.log.info("Container had exit code: %s", exit_code)
> airflow/contrib/operators/azure_container_instances_operator.py:                    self.log.info("Container exited with detail_status %s", detail_status)
> airflow/contrib/operators/azure_container_instances_operator.py:                    self.log.info("Azure provision failure")
> airflow/contrib/operators/winrm_operator.py:            self.log.info("Hook not found, creating...")
> airflow/contrib/operators/docker_swarm_operator.py:                self.log.info('Service status before exiting: %s', status)
> airflow/contrib/auth/backends/ldap_auth.py:        log.warning("Unable to find group for %s %s", search_base, search_filter)
> airflow/contrib/auth/backends/ldap_auth.py:        log.warning("""Missing attribute "%s" when looked-up in Ldap database.
> airflow/contrib/auth/backends/ldap_auth.py:        log.warning("Parsing error when retrieving the user's group(s)."
> airflow/contrib/utils/sendgrid.py:        log.warning('Failed to send out email with subject %s, status code: %s',
> airflow/contrib/example_dags/example_kubernetes_operator.py:    log.warning("Could not import KubernetesPodOperator: " + str(e))
> airflow/contrib/hooks/gdrive_hook.py:                self.log.info("Not found %s directory", current_folder)
> airflow/contrib/hooks/sqoop_hook.py:        self.log.info("Command exited with return code %s", self.sp.returncode)
> airflow/contrib/hooks/salesforce_hook.py:            log.warning("Could not convert field to timestamps: %s", column.name)
> airflow/contrib/hooks/qubole_check_hook.py:            log.info("Qubole command not found")
> airflow/contrib/hooks/spark_submit_hook.py:                    self.log.info("Exception when attempting to kill Spark on K8s:")
> airflow/www/node_modules/eslint/lib/util/npm-util.js:        log.info("Could not read package.json file. Please check that the file contains valid JSON.");
> airflow/www/node_modules/eslint/lib/config/config-initializer.js:            log.info("Local ESLint installation not found.");
> grep: airflow/www/static/docs: No such file or directory
> airflow/www/api/experimental/endpoints.py:            _log.info(error_message)
> airflow/www/api/experimental/endpoints.py:        _log.info(error_message)
> airflow/www/api/experimental/endpoints.py:        _log.info(error_message)
> airflow/executors/kubernetes_executor.py:            self.log.info('Event: %s Failed', pod_id)
> airflow/executors/kubernetes_executor.py:            self.log.warning("could not get try_number as an int: %s", labels.get('try_number', '1'))
> airflow/executors/kubernetes_executor.py:                    self.log.warning('ApiException when attempting to run task, re-queueing. '
> airflow/executors/base_executor.py:            self.log.info("could not queue task %s", simple_task_instance.key)
> airflow/executors/debug_executor.py:                self.log.info("Setting %s to %s", ti.key, State.UPSTREAM_FAILED)
> airflow/executors/debug_executor.py:            self.log.info("Setting %s to %s", ti.key, State.UPSTREAM_FAILED)
> airflow/hooks/hive_hooks.py:                self.log.info("Could not connect to %s:%s", conn.host, conn.port)
> airflow/stats.py:    log.warning("Could not configure StatsClient: %s, using DummyStatsLogger instead.", e)
> airflow/task/task_runner/cgroup_task_runner.py:            self.log.warning("Task failed with return code of 137. This may indicate "
> airflow/jobs/local_task_job.py:                    self.log.info("Task exited with return code %s", return_code)
> airflow/jobs/scheduler_job.py:        self.log.info("Exiting gracefully upon receiving signal %s", signum)
> airflow/jobs/scheduler_job.py:                    self.log.warning("TaskInstance %s went missing from the database", ti)
> airflow/jobs/scheduler_job.py:            self.log.info("Exited execute loop")
> airflow/jobs/scheduler_job.py:                self.log.info("Exiting scheduler loop as all files"
> airflow/jobs/backfill_job.py:        self.log.info("Backfill done. Exiting.")
> airflow/kubernetes/pod_launcher.py:            self.log.info('Event with job id %s Failed', job_id)
> grep: logs/scheduler/latest: No such file or directory
> tests/sensors/test_timeout_sensor.py:        self.log.info("Success criteria met. Exiting.")
> tests/gcp/operators/test_cloud_sql_system_helper.py:            self.log.info("File doesn't exits. Creating dir...")
> tests/gcp/operators/test_cloud_sql_system_helper.py:                self.log.info("Error while creating dir.")
> tests/gcp/utils/gcp_authenticator.py:            self.log.info('Airflow DB Session error:' + str(ex))
> tests/gcp/utils/gcp_authenticator.py:            self.log.info('Airflow DB Session error:' + str(ex))
> tests/gcp/utils/gcp_authenticator.py:            self.log.info("The {} is missing".format(key_path))
> tests/contrib/utils/logging_command_executor.py:                self.log.warning("Error when executing %s", " ".join(cmd))
> tests/contrib/utils/logging_command_executor.py:            self.log.info("Error when executing '{}'".format(" ".join(cmd)))



--
This message was sent by Atlassian Jira
(v8.3.4#803005)