You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by GitBox <gi...@apache.org> on 2021/03/29 13:08:28 UTC

[GitHub] [airflow] marclamberti commented on issue #12334: kubernetes executor pod not removed

marclamberti commented on issue #12334:
URL: https://github.com/apache/airflow/issues/12334#issuecomment-809362985


   I have this issue with Airflow 2.0.1 and the Helm chart.
   It happens since I enabled remote logging with S3 where the credentials are given through a Secret object.
   ```
   mydagdownloading.09c4d4061ada4c058dd91ac4885dd76f   1/1     Running   0          32m
   mydagdownloading.2bc28a8a155f4e29b234e1d153eef52e   1/1     Running   0          37m
   mydagdownloading.f25cdd3178604b3f8b343f6bb13d7840   1/1     Running   0          42m
   mydagdownloading.fcca60688a324a898014cb98bcbffbc2   1/1     Running   0          34m
   mydagprocessing.09d820cdb1f14a919d0e9a7661b295cf    1/1     Running   0          34m
   mydagprocessing.581978d0c2064f56bc6affebc499e489    1/1     Running   0          31m
   mydagprocessing.5d4b18ff4a8f4af4b2acff1595e156a3    1/1     Running   0          42m
   mydagprocessing.9efc24c0813a45789c93139c5ae4fc1d    1/1     Running   0          37m
   mydagstoring.3a9e3635a6294e57b605fd52c465825a       1/1     Running   0          41m
   mydagstoring.86ec9747f7824bcca684fe5e944ef1b4       1/1     Running   0          31m
   mydagstoring.eabf0e43467848e7988fca4e8bf45267       1/1     Running   0          34m
   mydagstoring.eb1533e230ed440e8b40c0c57b361a4b       1/1     Running   0          37m
   ```
   And whenever I try to get the logs I get this message:
   ```
   *** Failed to verify remote log exists s3://marcl-airflow/airflow/logs/my_dag/downloading/2021-03-29T12:33:27.660837+00:00/1.log.
   An error occurred (403) when calling the HeadObject operation: Forbidden
   *** Falling back to local log
   *** Trying to get logs (last 100 lines) from worker pod mydagdownloading.09c4d4061ada4c058dd91ac4885dd76f ***
   
   BACKEND=postgresql
   DB_HOST=airflow-postgresql.airflow.svc.cluster.local
   DB_PORT=5432
   
   [2021-03-29 12:33:34,965] {settings.py:210} DEBUG - Setting up DB connection pool (PID 8)
   [2021-03-29 12:33:34,966] {settings.py:281} DEBUG - settings.prepare_engine_args(): Using pool settings. pool_size=5, max_overflow=10, pool_recycle=1800, pid=8
   [2021-03-29 12:33:35,039] {cli_action_loggers.py:40} DEBUG - Adding <function default_action_log at 0x7f91156a5400> to pre execution callback
   [2021-03-29 12:33:37,012] {cli_action_loggers.py:66} DEBUG - Calling callbacks: [<function default_action_log at 0x7f91156a5400>]
   [2021-03-29 12:33:37,025] {settings.py:210} DEBUG - Setting up DB connection pool (PID 8)
   [2021-03-29 12:33:37,026] {settings.py:243} DEBUG - settings.prepare_engine_args(): Using NullPool
   [2021-03-29 12:33:37,027] {dagbag.py:448} INFO - Filling up the DagBag from /opt/airflow/dags/repo/my_dag.py
   [2021-03-29 12:33:37,027] {dagbag.py:287} DEBUG - Importing /opt/airflow/dags/repo/my_dag.py
   [2021-03-29 12:33:37,039] {dagbag.py:413} DEBUG - Loaded DAG <DAG: my_dag>
   [2021-03-29 12:33:37,057] {plugins_manager.py:270} DEBUG - Loading plugins
   [2021-03-29 12:33:37,057] {plugins_manager.py:207} DEBUG - Loading plugins from directory: /opt/airflow/plugins
   [2021-03-29 12:33:37,057] {plugins_manager.py:184} DEBUG - Loading plugins from entrypoints
   [2021-03-29 12:33:37,101] {plugins_manager.py:414} DEBUG - Integrate DAG plugins
   ```
   which I think is misleading as the user has all permissions on the bucket/objects. I tried from the worker pod and I can access the bucket and put files into it. But, the worker pods keep running and don't upload any log files.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
users@infra.apache.org