You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by GitBox <gi...@apache.org> on 2021/09/17 14:48:49 UTC

[GitHub] [airflow] tienhung2812 commented on issue #13542: Task stuck in "scheduled" or "queued" state, pool has all slots queued, nothing is executing

tienhung2812 commented on issue #13542:
URL: https://github.com/apache/airflow/issues/13542#issuecomment-921858288


   Hi, just got this issue again on `2.1.3` but the behavior quite different, all of schedulers stuck into a DAG check loop, but restart the scheduler will fix it, all other DAG as stuck in `scheduled`
   
   ```
    [scheduled]> since the number of tasks running or queued from DAG abc_dag is >= to the DAG's task concurrency limit of 16
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,511] {scheduler_job.py:410} INFO - DAG abc_dag has 16/16 running and queued tasks
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,512] {scheduler_job.py:417} INFO - Not executing <TaskInstance: abc_dag.aa.aa_20.collector_v2 2021-09-17 11:30:00+00:00 [scheduled]> since the number of tasks running or queued from DAG abc_dag is >= to the DAG's task concurrency limit of 16
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,512] {scheduler_job.py:410} INFO - DAG abc_dag has 16/16 running and queued tasks
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,513] {scheduler_job.py:417} INFO - Not executing <TaskInstance: abc_dag.aa.ff_46.collector_v2 2021-09-17 11:30:00+00:00 [scheduled]> since the number of tasks running or queued from DAG abc_dag is >= to the DAG's task concurrency limit of 16
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,513] {scheduler_job.py:410} INFO - DAG abc_dag has 16/16 running and queued tasks`
   airflow_scheduler.2.bj0t349vgttt@ip-172-30-2-178    | [2021-09-17 14:29:07,514] {scheduler_job.py:417} INFO - Not executing <TaskInstance: abc_dag.ff.gg_121.collector_v2 2021-09-17 11:30:00+00:00 [scheduled]> since the number of tasks running or queued from DAG abc_dag is >= to the DAG's task concurrency limit of 16
   ```
   
   The current system have about >= 150 active dags but we have a`abc_dag` have more than 500 tasks inside
   I have already increase the scheduler config to handling more task
   - max_tis_per_query: 512
   - max_dagruns_to_create_per_loop: 50
   - max_dagruns_per_loop_to_schedule: 70
   I have also add a scheduler healthcheck but it can not detect this issue 
   
   Environment:
   - Celery Executor
   - MySQL 8.0.23-14
   - Airflow: 2.1.3 
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscribe@airflow.apache.org

For queries about this service, please contact Infrastructure at:
users@infra.apache.org