You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "Ben Storrie (JIRA)" <ji...@apache.org> on 2019/06/11 00:08:00 UTC

[jira] [Created] (AIRFLOW-4761) Airflow Task Clear function throws error

Ben Storrie created AIRFLOW-4761:
------------------------------------

             Summary: Airflow Task Clear function throws error
                 Key: AIRFLOW-4761
                 URL: https://issues.apache.org/jira/browse/AIRFLOW-4761
             Project: Apache Airflow
          Issue Type: Bug
          Components: DAG, DagRun
    Affects Versions: 1.10.3
         Environment: CentOS 7, Python 2.7.10
            Reporter: Ben Storrie


When using the airflow webserver to clear a task inside a dagrun, an error is thrown on certain types of tasks:

 
{code:java}
Traceback (most recent call last):
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py", line 2311, in wsgi_app
response = self.full_dispatch_request()
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py", line 1834, in full_dispatch_request
rv = self.handle_user_exception(e)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py", line 1737, in handle_user_exception
reraise(exc_type, exc_value, tb)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py", line 1832, in full_dispatch_request
rv = self.dispatch_request()
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask/app.py", line 1818, in dispatch_request
return self.view_functions[rule.endpoint](**req.view_args)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_admin/base.py", line 69, in inner
return self._run_view(f, *args, **kwargs)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_admin/base.py", line 368, in _run_view
return fn(self, *args, **kwargs)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/flask_login/utils.py", line 261, in decorated_view
return func(*args, **kwargs)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/utils.py", line 275, in wrapper
return f(*args, **kwargs)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/utils.py", line 322, in wrapper
return f(*args, **kwargs)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/www/views.py", line 1202, in clear
include_upstream=upstream)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py", line 3830, in sub_dag
dag = copy.deepcopy(self)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 174, in deepcopy
y = copier(memo)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py", line 3815, in __deepcopy__
setattr(result, k, copy.deepcopy(v, memo))
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 163, in deepcopy
y = copier(x, memo)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 257, in _deepcopy_dict
y[deepcopy(key, memo)] = deepcopy(value, memo)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 174, in deepcopy
y = copier(memo)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/site-packages/airflow/models/__init__.py", line 2492, in __deepcopy__
setattr(result, k, copy.copy(v))
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 96, in copy
return _reconstruct(x, rv, 0)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy.py", line 329, in _reconstruct
y = callable(*args)
File "/opt/my-miniconda/miniconda/envs/my-hadoop-airflow/lib/python2.7/copy_reg.py", line 93, in __newobj__
return cls.__new__(cls, *args)
TypeError: instancemethod expected at least 2 arguments, got 0{code}
 

I had expected AIRFLOW-2060 being resolved to resolve this on upgrade to 1.10.3:
{code:java}
(my-hadoop-airflow) [user@hostname ~]$ pip freeze | grep pendulum
pendulum==1.4.4{code}
 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)