You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by GitBox <gi...@apache.org> on 2021/07/14 00:40:57 UTC

[GitHub] [airflow] daoudjahdou edited a comment on issue #15978: DAG getting stuck in "running" state indefinitely

daoudjahdou edited a comment on issue #15978:
URL: https://github.com/apache/airflow/issues/15978#issuecomment-878909135


   Just tried it on both 2.1.1 and 1.10.15 same behaviour, i even tried the example found here: 
   https://airflow.apache.org/docs/apache-airflow/1.10.15/_modules/airflow/example_dags/example_subdag_operator.html
   
   ```py
   from airflow.example_dags.subdags.subdag import subdag
   from airflow.models import DAG
   from airflow.operators.dummy_operator import DummyOperator
   from airflow.operators.subdag_operator import SubDagOperator
   from airflow.utils.dates import days_ago
   
   DAG_NAME = 'example_subdag_operator'
   
   args = {
       'owner': 'Airflow',
       'start_date': days_ago(2),
   }
   
   dag = DAG(
       dag_id=DAG_NAME,
       default_args=args,
       schedule_interval="@once",
       tags=['example']
   )
   
   start = DummyOperator(
       task_id='start',
       dag=dag,
   )
   
   section_1 = SubDagOperator(
       task_id='section-1',
       subdag=subdag(DAG_NAME, 'section-1', args),
       dag=dag,
   )
   
   some_other_task = DummyOperator(
       task_id='some-other-task',
       dag=dag,
   )
   
   section_2 = SubDagOperator(
       task_id='section-2',
       subdag=subdag(DAG_NAME, 'section-2', args),
       dag=dag,
   )
   
   end = DummyOperator(
       task_id='end',
       dag=dag,
   )
   
   start >> section_1 >> some_other_task >> section_2 >> end
   ```
   
   The log is the following:
   ```
     ____________       _____________
    ____    |__( )_________  __/__  /________      __
   ____  /| |_  /__  ___/_  /_ __  /_  __ \_ | /| / /
   ___  ___ |  / _  /   _  __/ _  / / /_/ /_ |/ |/ /
    _/_/  |_/_/  /_/    /_/    /_/  \____/____/|__/
   Starting flask
    * Serving Flask app "airflow.utils.serve_logs" (lazy loading)
    * Environment: production
      WARNING: This is a development server. Do not use it in a production deployment.
      Use a production WSGI server instead.
    * Debug mode: off
   [2021-07-13 08:57:29,682] {_internal.py:113} INFO -  * Running on http://0.0.0.0:8793/ (Press CTRL+C to quit)
   [2021-07-13 08:57:29,690] {scheduler_job.py:1270} INFO - Starting the scheduler
   [2021-07-13 08:57:29,691] {scheduler_job.py:1275} INFO - Processing each file at most -1 times
   [2021-07-13 08:57:29,696] {dag_processing.py:254} INFO - Launched DagFileProcessorManager with pid: 129
   [2021-07-13 08:57:29,697] {scheduler_job.py:1839} INFO - Resetting orphaned tasks for active dag runs
   [2021-07-13 08:57:29,701] {settings.py:52} INFO - Configured default timezone Timezone('UTC')
   [2021-07-13 08:57:29,711] {dag_processing.py:532} WARNING - Because we cannot use more than 1 thread (parsing_processes = 2 ) when using sqlite. So we set parallelism to 1.
   [2021-07-13 08:57:37,907] {scheduler_job.py:964} INFO - 1 tasks up for execution:
           <TaskInstance: example_subdag_operator.section-1 2021-07-11 00:00:00+00:00 [scheduled]>
   [2021-07-13 08:57:37,909] {scheduler_job.py:998} INFO - Figuring out tasks to run in Pool(name=default_pool) with 128 open slots and 1 task instances ready to be queued
   [2021-07-13 08:57:37,909] {scheduler_job.py:1025} INFO - DAG example_subdag_operator has 0/16 running and queued tasks
   [2021-07-13 08:57:37,909] {scheduler_job.py:1086} INFO - Setting the following tasks to queued state:
           <TaskInstance: example_subdag_operator.section-1 2021-07-11 00:00:00+00:00 [scheduled]>
   [2021-07-13 08:57:37,911] {scheduler_job.py:1128} INFO - Sending TaskInstanceKey(dag_id='example_subdag_operator', task_id='section-1', execution_date=datetime.datetime(2021, 7, 11, 0, 0, tzinfo=Timezone('UTC')), try_number=1) to executor with priority 4 and queue default
   [2021-07-13 08:57:37,911] {base_executor.py:82} INFO - Adding to queue: ['airflow', 'tasks', 'run', 'example_subdag_operator', 'section-1', '2021-07-11T00:00:00+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/dags/monitor.py']
   [2021-07-13 08:57:37,922] {sequential_executor.py:59} INFO - Executing command: ['airflow', 'tasks', 'run', 'example_subdag_operator', 'section-1', '2021-07-11T00:00:00+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/dags/monitor.py']
   [2021-07-13 08:57:39,332] {dagbag.py:496} INFO - Filling up the DagBag from /opt/airflow/dags/monitor.py
   Running <TaskInstance: example_subdag_operator.section-1 2021-07-11T00:00:00+00:00 [queued]> on host 969b77ead72f
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org