You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "kajanajustin (via GitHub)" <gi...@apache.org> on 2023/06/27 00:44:50 UTC
[GitHub] [airflow] kajanajustin opened a new issue, #32167: Getting *** on things that should be named airflow
kajanajustin opened a new issue, #32167:
URL: https://github.com/apache/airflow/issues/32167
### Apache Airflow version
Other Airflow 2 version (please specify below)
### What happened
Hi Team,
Currently, I am running docker-compose using the airflow:2.5.3-python3.8 image. Currently, when I go running my dag everything looks good except I am getting these ugly asterisks everywhere the word 'airflow' is supposed to be. I double-checked my config and paths and still have the same outcome. Please let me know if there is something I may be doing incorrectly.
here are the logs below
```
*** Reading local file: /opt/airflow/logs/dag_id=heartbeat/run_id=manual__2023-06-27T00:25:10.808687+00:00/task_id=heartbeat/attempt=1.log
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1090} INFO - Dependencies all met for dep_context=non-requeueable deps ti=<TaskInstance: heartbeat.heartbeat manual__2023-06-27T00:25:10.808687+00:00 [queued]>
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1090} INFO - Dependencies all met for dep_context=requeueable deps ti=<TaskInstance: heartbeat.heartbeat manual__2023-06-27T00:25:10.808687+00:00 [queued]>
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1288} INFO -
--------------------------------------------------------------------------------
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1289} INFO - Starting attempt 1 of 1
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1290} INFO -
--------------------------------------------------------------------------------
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1309} INFO - Executing <Task(PipelineOperator): heartbeat> on 2023-06-27 00:25:10.808687+00:00
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:55} INFO - Started process 21159 to run task
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:82} INFO - Running: ['***', 'tasks', 'run', 'heartbeat', 'heartbeat', 'manual__2023-06-27T00:25:10.808687+00:00', '--job-id', '26', '--raw', '--subdir', 'DAGS_FOLDER/heartbeat.py', '--cfg-path', '/tmp/tmpp2vbdho7']
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:83} INFO - Job 26: Subtask heartbeat
[2023-06-26, 20:25:23 EDT] {task_command.py:389} INFO - Running <TaskInstance: heartbeat.heartbeat manual__2023-06-27T00:25:10.808687+00:00 [running]> on host c8908875256e
[2023-06-26, 20:25:23 EDT] {taskinstance.py:1516} INFO - Exporting the following env vars:
AIRFLOW_CTX_DAG_OWNER=***
AIRFLOW_CTX_DAG_ID=heartbeat
AIRFLOW_CTX_TASK_ID=heartbeat
AIRFLOW_CTX_EXECUTION_DATE=2023-06-27T00:25:10.808687+00:00
AIRFLOW_CTX_TRY_NUMBER=1
AIRFLOW_CTX_DAG_RUN_ID=manual__2023-06-27T00:25:10.808687+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:514} INFO - Using common lib version [1.1.91].
[2023-06-26, 20:25:23 EDT] {operators.py:171} INFO - Building end_interval using data_interval_end: 2023-06-27T00:25:00+00:00, data_interval_type : SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:179} INFO - end_interval = 2023-06-27T00:25:00+00:00
[2023-06-26, 20:25:23 EDT] {logging_mixin.py:137} INFO - Run id for the Lineage API is being set to: 88562ae5-3511-37a6-a361-59ebb4b3e111
[2023-06-26, 20:25:23 EDT] {operators.py:139} INFO - Building start_interval using data_interval_start: 2023-06-27T00:20:00+00:00, prev_data_interval_end_success : None,data_interval_type : SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:166} INFO - start_interval = 2023-06-27T00:20:00+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:171} INFO - Building end_interval using data_interval_end: 2023-06-27T00:25:00+00:00, data_interval_type : SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:179} INFO - end_interval = 2023-06-27T00:25:00+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:611} INFO - Bash command configuration: /opt/***/run-pipeline-shared.sh general heartbeat/heartbeat.py
```
All of my other airflow servers don't have this issue, and I have checked to ensure that the `sensitive_var_conn_names = False` Not sure what else to do at this point in time....
### What you think should happen instead
Instead, the *** should be shown as `airflow`
### How to reproduce
Utilizing docker-compose and the airflow:2.5.3-python3.8 image, essentially use any basic dag and have airflow pick up on it then go to logs and see the asterisks
### Operating System
Debian GNU/Linux 11 (bullseye)
### Versions of Apache Airflow Providers
_No response_
### Deployment
Docker-Compose
### Deployment details
_No response_
### Anything else
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
[GitHub] [airflow] kajanajustin commented on issue #32167: Getting *** on things that should be named airflow
Posted by "kajanajustin (via GitHub)" <gi...@apache.org>.
kajanajustin commented on issue #32167:
URL: https://github.com/apache/airflow/issues/32167#issuecomment-1608538278
Here is my DAG
```
from datetime import timedelta
from datetime import datetime
from airflow import DAG
from edm_ngage.airflow_common import operators
doc_md = """
### Redshift Heartbeat
#### Documentation
- [README](https://github.ancestry.com/Ancestry/edm-ngage-airflow/tree/master/src/jobs/heartbeat)
- [Airflow Development Documentation](https://confluence.ancestry.int/display/DATAPLAT/AirFlow+Development)
"""
DEFAULT_ARGS = {
'retries': 0,
'retry_delay': timedelta(minutes=1),
'params': {'alert_contact':['@U02CJCB9N9J']}
}
dag = DAG(
'heartbeat',
default_args=DEFAULT_ARGS,
description='Heartbeat pipeline to show Airflow is running.',
schedule_interval='*/5 * * * *',
max_active_runs=1,
concurrency=1,
start_date=datetime(2021,10,21),
tags=['xxVERSION_TAGxx', 'from_redshift', 'ngage', 'idempotent', 'wont_run_again'],
catchup=False,
doc_md=doc_md
)
run_heartbeat = operators.PipelineOperator(
task_id = 'heartbeat',
lineage=False,
run_again=False,
data_interval_type="SINCE_LAST_SUCCESS",
venv_name='general',
alt_run_script='/opt/airflow/run-pipeline-shared.sh',
execution_timeout=timedelta(minutes=10),
dag=dag)
run_heartbeat
```
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
[GitHub] [airflow] potiuk closed issue #32167: Getting *** on things that should be named airflow
Posted by "potiuk (via GitHub)" <gi...@apache.org>.
potiuk closed issue #32167: Getting *** on things that should be named airflow
URL: https://github.com/apache/airflow/issues/32167
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org
[GitHub] [airflow] boring-cyborg[bot] commented on issue #32167: Getting *** on things that should be named airflow
Posted by "boring-cyborg[bot] (via GitHub)" <gi...@apache.org>.
boring-cyborg[bot] commented on issue #32167:
URL: https://github.com/apache/airflow/issues/32167#issuecomment-1608536440
Thanks for opening your first issue here! Be sure to follow the issue template! If you are willing to raise PR to address this issue please do so, no need to wait for approval.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org
For queries about this service, please contact Infrastructure at:
users@infra.apache.org