You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by ka...@apache.org on 2020/08/06 12:43:40 UTC

[airflow] branch master updated: Changes to all the constructors to remove the args argument (#10163)

This is an automated email from the ASF dual-hosted git repository.

kaxilnaik pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/master by this push:
     new 24c8e4c  Changes to all the constructors to remove the args argument (#10163)
24c8e4c is described below

commit 24c8e4c2d6e359ecc2c7d6275dccc68de4a82832
Author: Leon Yuan <le...@gmail.com>
AuthorDate: Thu Aug 6 07:42:51 2020 -0500

    Changes to all the constructors to remove the args argument (#10163)
---
 airflow/operators/email.py                               |  4 ++--
 airflow/operators/sql.py                                 |  4 ----
 airflow/providers/amazon/aws/sensors/glue.py             |  2 +-
 .../providers/amazon/aws/transfers/google_api_to_s3.py   |  1 -
 airflow/providers/celery/sensors/celery_queue.py         |  3 +--
 .../cncf/kubernetes/sensors/spark_kubernetes.py          |  4 ++--
 airflow/providers/datadog/sensors/datadog.py             |  3 +--
 airflow/providers/ftp/sensors/ftp.py                     |  3 +--
 airflow/providers/google/ads/transfers/ads_to_gcs.py     |  3 +--
 airflow/providers/google/cloud/sensors/bigquery.py       |  4 ++--
 airflow/providers/google/cloud/sensors/bigquery_dts.py   |  3 +--
 airflow/providers/google/cloud/sensors/bigtable.py       |  3 +--
 .../cloud/sensors/cloud_storage_transfer_service.py      |  3 +--
 airflow/providers/google/cloud/sensors/gcs.py            | 16 ++++++++--------
 airflow/providers/google/cloud/sensors/pubsub.py         |  3 +--
 .../google/cloud/transfers/bigquery_to_bigquery.py       |  3 +--
 .../providers/google/cloud/transfers/bigquery_to_gcs.py  |  3 +--
 .../google/cloud/transfers/bigquery_to_mysql.py          |  3 +--
 .../providers/google/cloud/transfers/cassandra_to_gcs.py |  3 +--
 .../google/cloud/transfers/facebook_ads_to_gcs.py        |  3 +--
 .../providers/google/cloud/transfers/gcs_to_bigquery.py  |  4 ++--
 airflow/providers/google/cloud/transfers/gcs_to_gcs.py   |  3 +--
 airflow/providers/google/cloud/transfers/gcs_to_local.py |  3 +--
 airflow/providers/google/cloud/transfers/gcs_to_sftp.py  |  3 +--
 airflow/providers/google/cloud/transfers/local_to_gcs.py |  3 +--
 airflow/providers/google/cloud/transfers/mssql_to_gcs.py |  3 +--
 airflow/providers/google/cloud/transfers/mysql_to_gcs.py |  3 +--
 .../providers/google/cloud/transfers/postgres_to_gcs.py  |  3 +--
 .../providers/google/cloud/transfers/presto_to_gcs.py    |  3 +--
 airflow/providers/google/cloud/transfers/s3_to_gcs.py    |  2 --
 airflow/providers/google/cloud/transfers/sftp_to_gcs.py  |  3 +--
 airflow/providers/google/cloud/transfers/sql_to_gcs.py   |  3 +--
 .../marketing_platform/sensors/campaign_manager.py       |  3 +--
 airflow/providers/http/sensors/http.py                   |  5 ++---
 airflow/providers/imap/sensors/imap_attachment.py        |  3 +--
 airflow/providers/jira/sensors/jira.py                   |  3 +--
 .../providers/microsoft/azure/sensors/azure_cosmos.py    |  3 +--
 airflow/providers/microsoft/azure/sensors/wasb.py        |  6 ++----
 .../providers/microsoft/azure/transfers/file_to_wasb.py  |  3 +--
 .../azure/transfers/oracle_to_azure_data_lake.py         |  3 +--
 airflow/providers/mongo/sensors/mongo.py                 |  3 +--
 airflow/providers/mysql/transfers/presto_to_mysql.py     |  4 ++--
 airflow/providers/mysql/transfers/s3_to_mysql.py         |  4 ++--
 airflow/providers/mysql/transfers/vertica_to_mysql.py    |  4 ++--
 airflow/providers/oracle/transfers/oracle_to_oracle.py   |  4 ++--
 airflow/providers/qubole/operators/qubole.py             |  5 ++---
 airflow/providers/qubole/operators/qubole_check.py       |  5 ++---
 airflow/providers/qubole/sensors/qubole.py               | 12 ++++++------
 airflow/providers/redis/sensors/redis_key.py             |  4 ++--
 airflow/providers/redis/sensors/redis_pub_sub.py         |  4 ++--
 .../providers/salesforce/sensors/tableau_job_status.py   |  3 +--
 airflow/providers/sftp/sensors/sftp.py                   |  4 ++--
 airflow/providers/snowflake/transfers/s3_to_snowflake.py |  4 ++--
 .../providers/snowflake/transfers/snowflake_to_slack.py  |  4 ++--
 airflow/sensors/base_sensor_operator.py                  |  3 +--
 airflow/sensors/bash.py                                  |  4 ++--
 airflow/sensors/date_time_sensor.py                      |  4 ++--
 airflow/sensors/external_task_sensor.py                  |  6 ++----
 airflow/sensors/filesystem.py                            |  3 +--
 airflow/sensors/python.py                                |  4 ++--
 airflow/sensors/sql_sensor.py                            |  4 ++--
 airflow/sensors/time_delta_sensor.py                     |  4 ++--
 airflow/sensors/time_sensor.py                           |  4 ++--
 airflow/sensors/weekday_sensor.py                        |  4 ++--
 docs/howto/custom-operator.rst                           | 16 ++++++++--------
 docs/howto/define_extra_link.rst                         |  4 ++--
 tests/models/test_baseoperator.py                        |  4 ++--
 tests/sensors/test_timeout_sensor.py                     |  3 +--
 tests/test_utils/mock_operators.py                       |  8 ++++----
 tests/utils/test_decorators.py                           |  4 ++--
 70 files changed, 117 insertions(+), 165 deletions(-)

diff --git a/airflow/operators/email.py b/airflow/operators/email.py
index 8bd8956..411bb7d 100644
--- a/airflow/operators/email.py
+++ b/airflow/operators/email.py
@@ -61,8 +61,8 @@ class EmailOperator(BaseOperator):
             bcc: Optional[Union[List[str], str]] = None,
             mime_subtype: str = 'mixed',
             mime_charset: str = 'utf-8',
-            *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+            **kwargs) -> None:
+        super().__init__(**kwargs)
         self.to = to  # pylint: disable=invalid-name
         self.subject = subject
         self.html_content = html_content
diff --git a/airflow/operators/sql.py b/airflow/operators/sql.py
index ee2d547..2b260eb 100644
--- a/airflow/operators/sql.py
+++ b/airflow/operators/sql.py
@@ -157,7 +157,6 @@ class SQLValueCheckOperator(BaseOperator):
         pass_value: Any,
         tolerance: Any = None,
         conn_id: Optional[str] = None,
-        *args,
         **kwargs,
     ):
         super().__init__(**kwargs)
@@ -280,7 +279,6 @@ class SQLIntervalCheckOperator(BaseOperator):
         ratio_formula: Optional[str] = "max_over_min",
         ignore_zero: Optional[bool] = True,
         conn_id: Optional[str] = None,
-        *args,
         **kwargs,
     ):
         super().__init__(**kwargs)
@@ -422,7 +420,6 @@ class SQLThresholdCheckOperator(BaseOperator):
         min_threshold: Any,
         max_threshold: Any,
         conn_id: Optional[str] = None,
-        *args,
         **kwargs,
     ):
         super().__init__(**kwargs)
@@ -517,7 +514,6 @@ class BranchSQLOperator(BaseOperator, SkipMixin):
         conn_id: str = "default_conn_id",
         database: Optional[str] = None,
         parameters: Optional[Union[Mapping, Iterable]] = None,
-        *args,
         **kwargs,
     ) -> None:
         super().__init__(**kwargs)
diff --git a/airflow/providers/amazon/aws/sensors/glue.py b/airflow/providers/amazon/aws/sensors/glue.py
index 92c8dcf..4525602 100644
--- a/airflow/providers/amazon/aws/sensors/glue.py
+++ b/airflow/providers/amazon/aws/sensors/glue.py
@@ -40,7 +40,7 @@ class AwsGlueJobSensor(BaseSensorOperator):
                  run_id,
                  aws_conn_id='aws_default',
                  **kwargs):
-        super(AwsGlueJobSensor, self).__init__(**kwargs)
+        super().__init__(**kwargs)
         self.job_name = job_name
         self.run_id = run_id
         self.aws_conn_id = aws_conn_id
diff --git a/airflow/providers/amazon/aws/transfers/google_api_to_s3.py b/airflow/providers/amazon/aws/transfers/google_api_to_s3.py
index ddc81a4..741b8a0 100644
--- a/airflow/providers/amazon/aws/transfers/google_api_to_s3.py
+++ b/airflow/providers/amazon/aws/transfers/google_api_to_s3.py
@@ -91,7 +91,6 @@ class GoogleApiToS3Operator(BaseOperator):
         google_api_endpoint_path,
         google_api_endpoint_params,
         s3_destination_key,
-
         google_api_response_via_xcom=None,
         google_api_endpoint_params_via_xcom=None,
         google_api_endpoint_params_via_xcom_task_ids=None,
diff --git a/airflow/providers/celery/sensors/celery_queue.py b/airflow/providers/celery/sensors/celery_queue.py
index f9b09e5..484b7d9 100644
--- a/airflow/providers/celery/sensors/celery_queue.py
+++ b/airflow/providers/celery/sensors/celery_queue.py
@@ -40,10 +40,9 @@ class CeleryQueueSensor(BaseSensorOperator):
             self,
             celery_queue: str,
             target_task_id: Optional[str] = None,
-            *args,
             **kwargs) -> None:
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.celery_queue = celery_queue
         self.target_task_id = target_task_id
 
diff --git a/airflow/providers/cncf/kubernetes/sensors/spark_kubernetes.py b/airflow/providers/cncf/kubernetes/sensors/spark_kubernetes.py
index e6b6ac1..9b12863 100644
--- a/airflow/providers/cncf/kubernetes/sensors/spark_kubernetes.py
+++ b/airflow/providers/cncf/kubernetes/sensors/spark_kubernetes.py
@@ -48,8 +48,8 @@ class SparkKubernetesSensor(BaseSensorOperator):
                  application_name: str,
                  namespace: Optional[str] = None,
                  kubernetes_conn_id: str = 'kubernetes_default',
-                 *args, **kwargs):
-        super().__init__(*args, **kwargs)
+                 **kwargs):
+        super().__init__(**kwargs)
         self.application_name = application_name
         self.namespace = namespace
         self.kubernetes_conn_id = kubernetes_conn_id
diff --git a/airflow/providers/datadog/sensors/datadog.py b/airflow/providers/datadog/sensors/datadog.py
index a24f46e..c155e8e 100644
--- a/airflow/providers/datadog/sensors/datadog.py
+++ b/airflow/providers/datadog/sensors/datadog.py
@@ -48,9 +48,8 @@ class DatadogSensor(BaseSensorOperator):
             sources: Optional[str] = None,
             tags: Optional[List[str]] = None,
             response_check: Optional[Callable[[Dict[str, Any]], bool]] = None,
-            *args,
             **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.datadog_conn_id = datadog_conn_id
         self.from_seconds_ago = from_seconds_ago
         self.up_to_seconds_from_now = up_to_seconds_from_now
diff --git a/airflow/providers/ftp/sensors/ftp.py b/airflow/providers/ftp/sensors/ftp.py
index a36e558..cd43c7c 100644
--- a/airflow/providers/ftp/sensors/ftp.py
+++ b/airflow/providers/ftp/sensors/ftp.py
@@ -49,9 +49,8 @@ class FTPSensor(BaseSensorOperator):
             path: str,
             ftp_conn_id: str = 'ftp_default',
             fail_on_transient_errors: bool = True,
-            *args,
             **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         self.path = path
         self.ftp_conn_id = ftp_conn_id
diff --git a/airflow/providers/google/ads/transfers/ads_to_gcs.py b/airflow/providers/google/ads/transfers/ads_to_gcs.py
index 90a57b2..4ec25af 100644
--- a/airflow/providers/google/ads/transfers/ads_to_gcs.py
+++ b/airflow/providers/google/ads/transfers/ads_to_gcs.py
@@ -74,10 +74,9 @@ class GoogleAdsToGcsOperator(BaseOperator):
         google_ads_conn_id: str = "google_ads_default",
         page_size: int = 10000,
         gzip: bool = False,
-        *args,
         **kwargs,
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.client_ids = client_ids
         self.query = query
         self.attributes = attributes
diff --git a/airflow/providers/google/cloud/sensors/bigquery.py b/airflow/providers/google/cloud/sensors/bigquery.py
index c1f722a..25ea5f0 100644
--- a/airflow/providers/google/cloud/sensors/bigquery.py
+++ b/airflow/providers/google/cloud/sensors/bigquery.py
@@ -56,9 +56,9 @@ class BigQueryTableExistenceSensor(BaseSensorOperator):
                  table_id: str,
                  bigquery_conn_id: str = 'google_cloud_default',
                  delegate_to: Optional[str] = None,
-                 *args, **kwargs) -> None:
+                 **kwargs) -> None:
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.project_id = project_id
         self.dataset_id = dataset_id
         self.table_id = table_id
diff --git a/airflow/providers/google/cloud/sensors/bigquery_dts.py b/airflow/providers/google/cloud/sensors/bigquery_dts.py
index 06e3e59..086eb19 100644
--- a/airflow/providers/google/cloud/sensors/bigquery_dts.py
+++ b/airflow/providers/google/cloud/sensors/bigquery_dts.py
@@ -77,10 +77,9 @@ class BigQueryDataTransferServiceTransferRunSensor(BaseSensorOperator):
         retry: Optional[Retry] = None,
         request_timeout: Optional[float] = None,
         metadata: Optional[Sequence[Tuple[str, str]]] = None,
-        *args,
         **kwargs
     ):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.run_id = run_id
         self.transfer_config_id = transfer_config_id
         self.retry = retry
diff --git a/airflow/providers/google/cloud/sensors/bigtable.py b/airflow/providers/google/cloud/sensors/bigtable.py
index 11bc48e..bd0f1db 100644
--- a/airflow/providers/google/cloud/sensors/bigtable.py
+++ b/airflow/providers/google/cloud/sensors/bigtable.py
@@ -59,7 +59,6 @@ class BigtableTableReplicationCompletedSensor(BaseSensorOperator, BigtableValida
         table_id: str,
         project_id: Optional[str] = None,
         gcp_conn_id: str = 'google_cloud_default',
-        *args,
         **kwargs
     ) -> None:
         self.project_id = project_id
@@ -67,7 +66,7 @@ class BigtableTableReplicationCompletedSensor(BaseSensorOperator, BigtableValida
         self.table_id = table_id
         self.gcp_conn_id = gcp_conn_id
         self._validate_inputs()
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
     def poke(self, context):
         hook = BigtableHook(gcp_conn_id=self.gcp_conn_id)
diff --git a/airflow/providers/google/cloud/sensors/cloud_storage_transfer_service.py b/airflow/providers/google/cloud/sensors/cloud_storage_transfer_service.py
index d2153b0..04111b1 100644
--- a/airflow/providers/google/cloud/sensors/cloud_storage_transfer_service.py
+++ b/airflow/providers/google/cloud/sensors/cloud_storage_transfer_service.py
@@ -60,10 +60,9 @@ class CloudDataTransferServiceJobStatusSensor(BaseSensorOperator):
         expected_statuses: Union[Set[str], str],
         project_id: Optional[str] = None,
         gcp_conn_id: str = 'google_cloud_default',
-        *args,
         **kwargs
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.job_name = job_name
         self.expected_statuses = (
             {expected_statuses} if isinstance(expected_statuses, str) else expected_statuses
diff --git a/airflow/providers/google/cloud/sensors/gcs.py b/airflow/providers/google/cloud/sensors/gcs.py
index 07e4f9a..73aec47 100644
--- a/airflow/providers/google/cloud/sensors/gcs.py
+++ b/airflow/providers/google/cloud/sensors/gcs.py
@@ -55,9 +55,9 @@ class GCSObjectExistenceSensor(BaseSensorOperator):
                  object: str,  # pylint: disable=redefined-builtin
                  google_cloud_conn_id: str = 'google_cloud_default',
                  delegate_to: Optional[str] = None,
-                 *args, **kwargs) -> None:
+                 **kwargs) -> None:
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.bucket = bucket
         self.object = object
         self.google_cloud_conn_id = google_cloud_conn_id
@@ -111,9 +111,9 @@ class GCSObjectUpdateSensor(BaseSensorOperator):
                  ts_func: Callable = ts_function,
                  google_cloud_conn_id: str = 'google_cloud_default',
                  delegate_to: Optional[str] = None,
-                 *args, **kwargs) -> None:
+                 **kwargs) -> None:
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.bucket = bucket
         self.object = object
         self.ts_func = ts_func
@@ -158,8 +158,8 @@ class GCSObjectsWtihPrefixExistenceSensor(BaseSensorOperator):
                  prefix: str,
                  google_cloud_conn_id: str = 'google_cloud_default',
                  delegate_to: Optional[str] = None,
-                 *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+                 **kwargs) -> None:
+        super().__init__(**kwargs)
         self.bucket = bucket
         self.prefix = prefix
         self.google_cloud_conn_id = google_cloud_conn_id
@@ -239,9 +239,9 @@ class GCSUploadSessionCompleteSensor(BaseSensorOperator):
                  allow_delete: bool = True,
                  google_cloud_conn_id: str = 'google_cloud_default',
                  delegate_to: Optional[str] = None,
-                 *args, **kwargs) -> None:
+                 **kwargs) -> None:
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         self.bucket = bucket
         self.prefix = prefix
diff --git a/airflow/providers/google/cloud/sensors/pubsub.py b/airflow/providers/google/cloud/sensors/pubsub.py
index d30320a..45c077f 100644
--- a/airflow/providers/google/cloud/sensors/pubsub.py
+++ b/airflow/providers/google/cloud/sensors/pubsub.py
@@ -103,7 +103,6 @@ class PubSubPullSensor(BaseSensorOperator):
             messages_callback: Optional[Callable[[List[ReceivedMessage], Dict[str, Any]], Any]] = None,
             delegate_to: Optional[str] = None,
             project: Optional[str] = None,
-            *args,
             **kwargs
     ) -> None:
         # To preserve backward compatibility
@@ -126,7 +125,7 @@ class PubSubPullSensor(BaseSensorOperator):
                 stacklevel=2
             )
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.gcp_conn_id = gcp_conn_id
         self.delegate_to = delegate_to
         self.project_id = project_id
diff --git a/airflow/providers/google/cloud/transfers/bigquery_to_bigquery.py b/airflow/providers/google/cloud/transfers/bigquery_to_bigquery.py
index f164cf2..a26ac29 100644
--- a/airflow/providers/google/cloud/transfers/bigquery_to_bigquery.py
+++ b/airflow/providers/google/cloud/transfers/bigquery_to_bigquery.py
@@ -86,9 +86,8 @@ class BigQueryToBigQueryOperator(BaseOperator):
                  labels: Optional[Dict] = None,
                  encryption_configuration: Optional[Dict] = None,
                  location: Optional[str] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         if bigquery_conn_id:
             warnings.warn(
diff --git a/airflow/providers/google/cloud/transfers/bigquery_to_gcs.py b/airflow/providers/google/cloud/transfers/bigquery_to_gcs.py
index 2c7002e..6b40088 100644
--- a/airflow/providers/google/cloud/transfers/bigquery_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/bigquery_to_gcs.py
@@ -85,9 +85,8 @@ class BigQueryToGCSOperator(BaseOperator):
                  delegate_to: Optional[str] = None,
                  labels: Optional[Dict] = None,
                  location: Optional[str] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         if bigquery_conn_id:
             warnings.warn(
diff --git a/airflow/providers/google/cloud/transfers/bigquery_to_mysql.py b/airflow/providers/google/cloud/transfers/bigquery_to_mysql.py
index ab0174d..c8e19bb 100644
--- a/airflow/providers/google/cloud/transfers/bigquery_to_mysql.py
+++ b/airflow/providers/google/cloud/transfers/bigquery_to_mysql.py
@@ -89,9 +89,8 @@ class BigQueryToMySqlOperator(BaseOperator):
                  replace: bool = False,
                  batch_size: int = 1000,
                  location: Optional[str] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.selected_fields = selected_fields
         self.gcp_conn_id = gcp_conn_id
         self.mysql_conn_id = mysql_conn_id
diff --git a/airflow/providers/google/cloud/transfers/cassandra_to_gcs.py b/airflow/providers/google/cloud/transfers/cassandra_to_gcs.py
index 16f7e2b..96ea6e4 100644
--- a/airflow/providers/google/cloud/transfers/cassandra_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/cassandra_to_gcs.py
@@ -93,9 +93,8 @@ class CassandraToGCSOperator(BaseOperator):
                  gcp_conn_id: str = 'google_cloud_default',
                  google_cloud_storage_conn_id: Optional[str] = None,
                  delegate_to: Optional[str] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         if google_cloud_storage_conn_id:
             warnings.warn(
diff --git a/airflow/providers/google/cloud/transfers/facebook_ads_to_gcs.py b/airflow/providers/google/cloud/transfers/facebook_ads_to_gcs.py
index 0b9140e..31abfbc 100644
--- a/airflow/providers/google/cloud/transfers/facebook_ads_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/facebook_ads_to_gcs.py
@@ -81,10 +81,9 @@ class FacebookAdsReportToGcsOperator(BaseOperator):
         api_version: str = "v6.0",
         gcp_conn_id: str = "google_cloud_default",
         facebook_conn_id: str = "facebook_default",
-        *args,
         **kwargs,
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.bucket_name = bucket_name
         self.object_name = object_name
         self.gcp_conn_id = gcp_conn_id
diff --git a/airflow/providers/google/cloud/transfers/gcs_to_bigquery.py b/airflow/providers/google/cloud/transfers/gcs_to_bigquery.py
index 241481e..58e69b1 100644
--- a/airflow/providers/google/cloud/transfers/gcs_to_bigquery.py
+++ b/airflow/providers/google/cloud/transfers/gcs_to_bigquery.py
@@ -187,9 +187,9 @@ class GCSToBigQueryOperator(BaseOperator):
                  autodetect=True,
                  encryption_configuration=None,
                  location=None,
-                 *args, **kwargs):
+                 **kwargs):
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         # GCS config
         if src_fmt_configs is None:
diff --git a/airflow/providers/google/cloud/transfers/gcs_to_gcs.py b/airflow/providers/google/cloud/transfers/gcs_to_gcs.py
index 70b6288..74bb0e6 100644
--- a/airflow/providers/google/cloud/transfers/gcs_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/gcs_to_gcs.py
@@ -185,9 +185,8 @@ class GCSToGCSOperator(BaseOperator):
                  last_modified_time=None,
                  maximum_modified_time=None,
                  is_older_than=None,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if google_cloud_storage_conn_id:
             warnings.warn(
                 "The google_cloud_storage_conn_id parameter has been deprecated. You should pass "
diff --git a/airflow/providers/google/cloud/transfers/gcs_to_local.py b/airflow/providers/google/cloud/transfers/gcs_to_local.py
index 9c564bd..36a57b8 100644
--- a/airflow/providers/google/cloud/transfers/gcs_to_local.py
+++ b/airflow/providers/google/cloud/transfers/gcs_to_local.py
@@ -75,7 +75,6 @@ class GCSToLocalFilesystemOperator(BaseOperator):
                  gcp_conn_id: str = 'google_cloud_default',
                  google_cloud_storage_conn_id: Optional[str] = None,
                  delegate_to: Optional[str] = None,
-                 *args,
                  **kwargs) -> None:
         # To preserve backward compatibility
         # TODO: Remove one day
@@ -95,7 +94,7 @@ class GCSToLocalFilesystemOperator(BaseOperator):
                 "the gcp_conn_id parameter.", DeprecationWarning, stacklevel=3)
             gcp_conn_id = google_cloud_storage_conn_id
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.bucket = bucket
         self.object = object_name
         self.filename = filename  # noqa
diff --git a/airflow/providers/google/cloud/transfers/gcs_to_sftp.py b/airflow/providers/google/cloud/transfers/gcs_to_sftp.py
index dfa8352..ffbee76 100644
--- a/airflow/providers/google/cloud/transfers/gcs_to_sftp.py
+++ b/airflow/providers/google/cloud/transfers/gcs_to_sftp.py
@@ -81,10 +81,9 @@ class GCSToSFTPOperator(BaseOperator):
         gcp_conn_id: str = "google_cloud_default",
         sftp_conn_id: str = "ssh_default",
         delegate_to: Optional[str] = None,
-        *args,
         **kwargs
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         self.source_bucket = source_bucket
         self.source_object = source_object
diff --git a/airflow/providers/google/cloud/transfers/local_to_gcs.py b/airflow/providers/google/cloud/transfers/local_to_gcs.py
index 1995261..bbbfa63 100644
--- a/airflow/providers/google/cloud/transfers/local_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/local_to_gcs.py
@@ -69,9 +69,8 @@ class LocalFilesystemToGCSOperator(BaseOperator):
                  mime_type='application/octet-stream',
                  delegate_to=None,
                  gzip=False,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         if google_cloud_storage_conn_id:
             warnings.warn(
diff --git a/airflow/providers/google/cloud/transfers/mssql_to_gcs.py b/airflow/providers/google/cloud/transfers/mssql_to_gcs.py
index d6d1c80..14317c0 100644
--- a/airflow/providers/google/cloud/transfers/mssql_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/mssql_to_gcs.py
@@ -60,9 +60,8 @@ class MSSQLToGCSOperator(BaseSQLToGCSOperator):
     @apply_defaults
     def __init__(self,
                  mssql_conn_id='mssql_default',
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.mssql_conn_id = mssql_conn_id
 
     def query(self):
diff --git a/airflow/providers/google/cloud/transfers/mysql_to_gcs.py b/airflow/providers/google/cloud/transfers/mysql_to_gcs.py
index 8ab7539..8bc5cb9 100644
--- a/airflow/providers/google/cloud/transfers/mysql_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/mysql_to_gcs.py
@@ -65,9 +65,8 @@ class MySQLToGCSOperator(BaseSQLToGCSOperator):
     def __init__(self,
                  mysql_conn_id='mysql_default',
                  ensure_utc=False,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.mysql_conn_id = mysql_conn_id
         self.ensure_utc = ensure_utc
 
diff --git a/airflow/providers/google/cloud/transfers/postgres_to_gcs.py b/airflow/providers/google/cloud/transfers/postgres_to_gcs.py
index 1f0aabf..98012a4 100644
--- a/airflow/providers/google/cloud/transfers/postgres_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/postgres_to_gcs.py
@@ -60,9 +60,8 @@ class PostgresToGCSOperator(BaseSQLToGCSOperator):
     @apply_defaults
     def __init__(self,
                  postgres_conn_id='postgres_default',
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.postgres_conn_id = postgres_conn_id
 
     def query(self):
diff --git a/airflow/providers/google/cloud/transfers/presto_to_gcs.py b/airflow/providers/google/cloud/transfers/presto_to_gcs.py
index 855b99a..047641f 100644
--- a/airflow/providers/google/cloud/transfers/presto_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/presto_to_gcs.py
@@ -182,10 +182,9 @@ class PrestoToGCSOperator(BaseSQLToGCSOperator):
     def __init__(
         self,
         presto_conn_id: str = "presto_default",
-        *args,
         **kwargs
     ):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.presto_conn_id = presto_conn_id
 
     def query(self):
diff --git a/airflow/providers/google/cloud/transfers/s3_to_gcs.py b/airflow/providers/google/cloud/transfers/s3_to_gcs.py
index 009f4e6..7502faa 100644
--- a/airflow/providers/google/cloud/transfers/s3_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/s3_to_gcs.py
@@ -105,7 +105,6 @@ class S3ToGCSOperator(S3ListOperator):
                  delegate_to=None,
                  replace=False,
                  gzip=False,
-                 *args,
                  **kwargs):
 
         super().__init__(
@@ -113,7 +112,6 @@ class S3ToGCSOperator(S3ListOperator):
             prefix=prefix,
             delimiter=delimiter,
             aws_conn_id=aws_conn_id,
-            *args,
             **kwargs)
 
         if dest_gcs_conn_id:
diff --git a/airflow/providers/google/cloud/transfers/sftp_to_gcs.py b/airflow/providers/google/cloud/transfers/sftp_to_gcs.py
index 38ef170..979f48e 100644
--- a/airflow/providers/google/cloud/transfers/sftp_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/sftp_to_gcs.py
@@ -84,10 +84,9 @@ class SFTPToGCSOperator(BaseOperator):
         mime_type: str = "application/octet-stream",
         gzip: bool = False,
         move_object: bool = False,
-        *args,
         **kwargs
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         self.source_path = source_path
         self.destination_path = self._set_destination_path(destination_path)
diff --git a/airflow/providers/google/cloud/transfers/sql_to_gcs.py b/airflow/providers/google/cloud/transfers/sql_to_gcs.py
index 216c1fa..3a50c6e 100644
--- a/airflow/providers/google/cloud/transfers/sql_to_gcs.py
+++ b/airflow/providers/google/cloud/transfers/sql_to_gcs.py
@@ -93,9 +93,8 @@ class BaseSQLToGCSOperator(BaseOperator):
                  gcp_conn_id='google_cloud_default',
                  google_cloud_storage_conn_id=None,
                  delegate_to=None,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         if google_cloud_storage_conn_id:
             warnings.warn(
diff --git a/airflow/providers/google/marketing_platform/sensors/campaign_manager.py b/airflow/providers/google/marketing_platform/sensors/campaign_manager.py
index b721c2f..007eb8c 100644
--- a/airflow/providers/google/marketing_platform/sensors/campaign_manager.py
+++ b/airflow/providers/google/marketing_platform/sensors/campaign_manager.py
@@ -77,10 +77,9 @@ class GoogleCampaignManagerReportSensor(BaseSensorOperator):
         delegate_to: Optional[str] = None,
         mode: str = "reschedule",
         poke_interval: int = 60 * 5,
-        *args,
         **kwargs
     ):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.mode = mode
         self.poke_interval = poke_interval
         self.profile_id = profile_id
diff --git a/airflow/providers/http/sensors/http.py b/airflow/providers/http/sensors/http.py
index 05a9bff..22c97b5 100644
--- a/airflow/providers/http/sensors/http.py
+++ b/airflow/providers/http/sensors/http.py
@@ -78,9 +78,8 @@ class HttpSensor(BaseSensorOperator):
                  headers: Optional[Dict[str, Any]] = None,
                  response_check: Optional[Callable[..., Any]] = None,
                  extra_options: Optional[Dict[str, Any]] = None,
-                 *args: Any, **kwargs: Any
-                 ) -> None:
-        super().__init__(*args, **kwargs)
+                 **kwargs: Any) -> None:
+        super().__init__(**kwargs)
         self.endpoint = endpoint
         self.http_conn_id = http_conn_id
         self.request_params = request_params or {}
diff --git a/airflow/providers/imap/sensors/imap_attachment.py b/airflow/providers/imap/sensors/imap_attachment.py
index c02badd..f62bbc4 100644
--- a/airflow/providers/imap/sensors/imap_attachment.py
+++ b/airflow/providers/imap/sensors/imap_attachment.py
@@ -50,9 +50,8 @@ class ImapAttachmentSensor(BaseSensorOperator):
                  mail_folder='INBOX',
                  mail_filter='All',
                  conn_id='imap_default',
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
         self.attachment_name = attachment_name
         self.check_regex = check_regex
diff --git a/airflow/providers/jira/sensors/jira.py b/airflow/providers/jira/sensors/jira.py
index 00d673b..8ae4979 100644
--- a/airflow/providers/jira/sensors/jira.py
+++ b/airflow/providers/jira/sensors/jira.py
@@ -44,9 +44,8 @@ class JiraSensor(BaseSensorOperator):
                  jira_conn_id: str = 'jira_default',
                  method_params: Optional[dict] = None,
                  result_processor: Optional[Callable] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.jira_conn_id = jira_conn_id
         self.result_processor = None
         if result_processor is not None:
diff --git a/airflow/providers/microsoft/azure/sensors/azure_cosmos.py b/airflow/providers/microsoft/azure/sensors/azure_cosmos.py
index d7f2616..6bf2c96 100644
--- a/airflow/providers/microsoft/azure/sensors/azure_cosmos.py
+++ b/airflow/providers/microsoft/azure/sensors/azure_cosmos.py
@@ -51,9 +51,8 @@ class AzureCosmosDocumentSensor(BaseSensorOperator):
             collection_name: str,
             document_id: str,
             azure_cosmos_conn_id: str = "azure_cosmos_default",
-            *args,
             **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.azure_cosmos_conn_id = azure_cosmos_conn_id
         self.database_name = database_name
         self.collection_name = collection_name
diff --git a/airflow/providers/microsoft/azure/sensors/wasb.py b/airflow/providers/microsoft/azure/sensors/wasb.py
index f689ed0..92ff7fc 100644
--- a/airflow/providers/microsoft/azure/sensors/wasb.py
+++ b/airflow/providers/microsoft/azure/sensors/wasb.py
@@ -46,9 +46,8 @@ class WasbBlobSensor(BaseSensorOperator):
                  blob_name: str,
                  wasb_conn_id: str = 'wasb_default',
                  check_options: Optional[dict] = None,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if check_options is None:
             check_options = {}
         self.wasb_conn_id = wasb_conn_id
@@ -88,9 +87,8 @@ class WasbPrefixSensor(BaseSensorOperator):
                  prefix: str,
                  wasb_conn_id: str = 'wasb_default',
                  check_options: Optional[dict] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if check_options is None:
             check_options = {}
         self.wasb_conn_id = wasb_conn_id
diff --git a/airflow/providers/microsoft/azure/transfers/file_to_wasb.py b/airflow/providers/microsoft/azure/transfers/file_to_wasb.py
index 64cdbc0..610b9f5 100644
--- a/airflow/providers/microsoft/azure/transfers/file_to_wasb.py
+++ b/airflow/providers/microsoft/azure/transfers/file_to_wasb.py
@@ -48,9 +48,8 @@ class FileToWasbOperator(BaseOperator):
                  blob_name: str,
                  wasb_conn_id: str = 'wasb_default',
                  load_options: Optional[dict] = None,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if load_options is None:
             load_options = {}
         self.file_path = file_path
diff --git a/airflow/providers/microsoft/azure/transfers/oracle_to_azure_data_lake.py b/airflow/providers/microsoft/azure/transfers/oracle_to_azure_data_lake.py
index 65c128c..a506e42 100644
--- a/airflow/providers/microsoft/azure/transfers/oracle_to_azure_data_lake.py
+++ b/airflow/providers/microsoft/azure/transfers/oracle_to_azure_data_lake.py
@@ -73,9 +73,8 @@ class OracleToAzureDataLakeOperator(BaseOperator):
             encoding: str = "utf-8",
             quotechar: str = '"',
             quoting: str = csv.QUOTE_MINIMAL,
-            *args,
             **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if sql_params is None:
             sql_params = {}
         self.filename = filename
diff --git a/airflow/providers/mongo/sensors/mongo.py b/airflow/providers/mongo/sensors/mongo.py
index b078699..25684ba 100644
--- a/airflow/providers/mongo/sensors/mongo.py
+++ b/airflow/providers/mongo/sensors/mongo.py
@@ -45,9 +45,8 @@ class MongoSensor(BaseSensorOperator):
                  collection: str,
                  query: dict,
                  mongo_conn_id: str = "mongo_default",
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.mongo_conn_id = mongo_conn_id
         self.collection = collection
         self.query = query
diff --git a/airflow/providers/mysql/transfers/presto_to_mysql.py b/airflow/providers/mysql/transfers/presto_to_mysql.py
index 156ff31..3be4d93 100644
--- a/airflow/providers/mysql/transfers/presto_to_mysql.py
+++ b/airflow/providers/mysql/transfers/presto_to_mysql.py
@@ -56,8 +56,8 @@ class PrestoToMySqlOperator(BaseOperator):
                  presto_conn_id: str = 'presto_default',
                  mysql_conn_id: str = 'mysql_default',
                  mysql_preoperator: Optional[str] = None,
-                 *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+                 **kwargs) -> None:
+        super().__init__(**kwargs)
         self.sql = sql
         self.mysql_table = mysql_table
         self.mysql_conn_id = mysql_conn_id
diff --git a/airflow/providers/mysql/transfers/s3_to_mysql.py b/airflow/providers/mysql/transfers/s3_to_mysql.py
index 53cb144..94638e6 100644
--- a/airflow/providers/mysql/transfers/s3_to_mysql.py
+++ b/airflow/providers/mysql/transfers/s3_to_mysql.py
@@ -58,8 +58,8 @@ class S3ToMySqlOperator(BaseOperator):
                  mysql_extra_options: Optional[str] = None,
                  aws_conn_id: str = 'aws_default',
                  mysql_conn_id: str = 'mysql_default',
-                 *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+                 **kwargs) -> None:
+        super().__init__(**kwargs)
         self.s3_source_key = s3_source_key
         self.mysql_table = mysql_table
         self.mysql_duplicate_key_handling = mysql_duplicate_key_handling
diff --git a/airflow/providers/mysql/transfers/vertica_to_mysql.py b/airflow/providers/mysql/transfers/vertica_to_mysql.py
index 0e6b8a6..ee72246 100644
--- a/airflow/providers/mysql/transfers/vertica_to_mysql.py
+++ b/airflow/providers/mysql/transfers/vertica_to_mysql.py
@@ -72,8 +72,8 @@ class VerticaToMySqlOperator(BaseOperator):
             mysql_preoperator=None,
             mysql_postoperator=None,
             bulk_load=False,
-            *args, **kwargs):
-        super().__init__(*args, **kwargs)
+            **kwargs):
+        super().__init__(**kwargs)
         self.sql = sql
         self.mysql_table = mysql_table
         self.mysql_conn_id = mysql_conn_id
diff --git a/airflow/providers/oracle/transfers/oracle_to_oracle.py b/airflow/providers/oracle/transfers/oracle_to_oracle.py
index 6f845d3..42b2516 100644
--- a/airflow/providers/oracle/transfers/oracle_to_oracle.py
+++ b/airflow/providers/oracle/transfers/oracle_to_oracle.py
@@ -53,8 +53,8 @@ class OracleToOracleOperator(BaseOperator):
             source_sql,
             source_sql_params=None,
             rows_chunk=5000,
-            *args, **kwargs):
-        super().__init__(*args, **kwargs)
+            **kwargs):
+        super().__init__(**kwargs)
         if source_sql_params is None:
             source_sql_params = {}
         self.oracle_destination_conn_id = oracle_destination_conn_id
diff --git a/airflow/providers/qubole/operators/qubole.py b/airflow/providers/qubole/operators/qubole.py
index 84afd5e..acfd9f2 100644
--- a/airflow/providers/qubole/operators/qubole.py
+++ b/airflow/providers/qubole/operators/qubole.py
@@ -184,8 +184,7 @@ class QuboleOperator(BaseOperator):
     )
 
     @apply_defaults
-    def __init__(self, qubole_conn_id="qubole_default", *args, **kwargs):
-        self.args = args
+    def __init__(self, qubole_conn_id="qubole_default", **kwargs):
         self.kwargs = kwargs
         self.kwargs['qubole_conn_id'] = qubole_conn_id
         self.hook = None
@@ -226,7 +225,7 @@ class QuboleOperator(BaseOperator):
 
     def get_hook(self):
         """Reinitialising the hook, as some template fields might have changed"""
-        return QuboleHook(*self.args, **self.kwargs)
+        return QuboleHook(**self.kwargs)
 
     def __getattribute__(self, name):
         if name in QuboleOperator.template_fields:
diff --git a/airflow/providers/qubole/operators/qubole_check.py b/airflow/providers/qubole/operators/qubole_check.py
index 5ed840e..ba70f07 100644
--- a/airflow/providers/qubole/operators/qubole_check.py
+++ b/airflow/providers/qubole/operators/qubole_check.py
@@ -101,7 +101,7 @@ class QuboleCheckOperator(CheckOperator, QuboleOperator):
         if hasattr(self, 'hook') and (self.hook is not None):
             return self.hook
         else:
-            return QuboleCheckHook(context=context, *self.args, **self.kwargs)
+            return QuboleCheckHook(context=context, **self.kwargs)
 
     def __getattribute__(self, name):
         if name in QuboleCheckOperator.template_fields:
@@ -162,7 +162,7 @@ class QuboleValueCheckOperator(ValueCheckOperator, QuboleOperator):
     ui_fgcolor = '#000'
 
     @apply_defaults
-    def __init__(self, pass_value, tolerance=None, results_parser_callable=None,
+    def __init__(self, *, pass_value, tolerance=None, results_parser_callable=None,
                  qubole_conn_id="qubole_default", **kwargs):
 
         sql = get_sql_from_qbol_cmd(kwargs)
@@ -191,7 +191,6 @@ class QuboleValueCheckOperator(ValueCheckOperator, QuboleOperator):
         else:
             return QuboleCheckHook(
                 context=context,
-                *self.args,
                 results_parser_callable=self.results_parser_callable,
                 **self.kwargs
             )
diff --git a/airflow/providers/qubole/sensors/qubole.py b/airflow/providers/qubole/sensors/qubole.py
index fae8c28..c0ad7fb 100644
--- a/airflow/providers/qubole/sensors/qubole.py
+++ b/airflow/providers/qubole/sensors/qubole.py
@@ -35,7 +35,7 @@ class QuboleSensor(BaseSensorOperator):
     template_ext = ('.txt',)
 
     @apply_defaults
-    def __init__(self, data, qubole_conn_id="qubole_default", *args, **kwargs):
+    def __init__(self, data, qubole_conn_id="qubole_default", **kwargs):
         self.data = data
         self.qubole_conn_id = qubole_conn_id
 
@@ -44,7 +44,7 @@ class QuboleSensor(BaseSensorOperator):
                                    "task '{0}' in dag '{1}'."
                                    .format(kwargs['task_id'], kwargs['dag'].dag_id))
 
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
     def poke(self, context):
 
@@ -83,9 +83,9 @@ class QuboleFileSensor(QuboleSensor):
     """
 
     @apply_defaults
-    def __init__(self, *args, **kwargs):
+    def __init__(self, **kwargs):
         self.sensor_class = FileSensor
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
 
 class QubolePartitionSensor(QuboleSensor):
@@ -106,6 +106,6 @@ class QubolePartitionSensor(QuboleSensor):
     """
 
     @apply_defaults
-    def __init__(self, *args, **kwargs):
+    def __init__(self, **kwargs):
         self.sensor_class = PartitionSensor
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
diff --git a/airflow/providers/redis/sensors/redis_key.py b/airflow/providers/redis/sensors/redis_key.py
index 59167e9..744b074 100644
--- a/airflow/providers/redis/sensors/redis_key.py
+++ b/airflow/providers/redis/sensors/redis_key.py
@@ -30,8 +30,8 @@ class RedisKeySensor(BaseSensorOperator):
     ui_color = '#f0eee4'
 
     @apply_defaults
-    def __init__(self, key: str, redis_conn_id: str, *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+    def __init__(self, key: str, redis_conn_id: str, **kwargs) -> None:
+        super().__init__(**kwargs)
         self.redis_conn_id = redis_conn_id
         self.key = key
 
diff --git a/airflow/providers/redis/sensors/redis_pub_sub.py b/airflow/providers/redis/sensors/redis_pub_sub.py
index 9be2b56..b9d5028 100644
--- a/airflow/providers/redis/sensors/redis_pub_sub.py
+++ b/airflow/providers/redis/sensors/redis_pub_sub.py
@@ -36,8 +36,8 @@ class RedisPubSubSensor(BaseSensorOperator):
     ui_color = '#f0eee4'
 
     @apply_defaults
-    def __init__(self, channels: Union[List[str], str], redis_conn_id: str, *args, **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+    def __init__(self, channels: Union[List[str], str], redis_conn_id: str, **kwargs) -> None:
+        super().__init__(**kwargs)
         self.channels = channels
         self.redis_conn_id = redis_conn_id
         self.pubsub = RedisHook(redis_conn_id=self.redis_conn_id).get_conn().pubsub()
diff --git a/airflow/providers/salesforce/sensors/tableau_job_status.py b/airflow/providers/salesforce/sensors/tableau_job_status.py
index b3873d5..424f172 100644
--- a/airflow/providers/salesforce/sensors/tableau_job_status.py
+++ b/airflow/providers/salesforce/sensors/tableau_job_status.py
@@ -50,9 +50,8 @@ class TableauJobStatusSensor(BaseSensorOperator):
                  job_id: str,
                  site_id: Optional[str] = None,
                  tableau_conn_id: str = 'tableau_default',
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.tableau_conn_id = tableau_conn_id
         self.job_id = job_id
         self.site_id = site_id
diff --git a/airflow/providers/sftp/sensors/sftp.py b/airflow/providers/sftp/sensors/sftp.py
index 2973d54..15592e3 100644
--- a/airflow/providers/sftp/sensors/sftp.py
+++ b/airflow/providers/sftp/sensors/sftp.py
@@ -37,8 +37,8 @@ class SFTPSensor(BaseSensorOperator):
     template_fields = ('path',)
 
     @apply_defaults
-    def __init__(self, path, sftp_conn_id='sftp_default', *args, **kwargs):
-        super().__init__(*args, **kwargs)
+    def __init__(self, path, sftp_conn_id='sftp_default', **kwargs):
+        super().__init__(**kwargs)
         self.path = path
         self.hook = None
         self.sftp_conn_id = sftp_conn_id
diff --git a/airflow/providers/snowflake/transfers/s3_to_snowflake.py b/airflow/providers/snowflake/transfers/s3_to_snowflake.py
index bbd6192..5a4a147 100644
--- a/airflow/providers/snowflake/transfers/s3_to_snowflake.py
+++ b/airflow/providers/snowflake/transfers/s3_to_snowflake.py
@@ -55,8 +55,8 @@ class S3ToSnowflakeOperator(BaseOperator):
                  columns_array=None,
                  autocommit=True,
                  snowflake_conn_id='snowflake_default',
-                 *args, **kwargs):
-        super().__init__(*args, **kwargs)
+                 **kwargs):
+        super().__init__(**kwargs)
         self.s3_keys = s3_keys
         self.table = table
         self.stage = stage
diff --git a/airflow/providers/snowflake/transfers/snowflake_to_slack.py b/airflow/providers/snowflake/transfers/snowflake_to_slack.py
index 9714695..2a5682d 100644
--- a/airflow/providers/snowflake/transfers/snowflake_to_slack.py
+++ b/airflow/providers/snowflake/transfers/snowflake_to_slack.py
@@ -80,9 +80,9 @@ class SnowflakeToSlackOperator(BaseOperator):
         schema: Optional[str] = None,
         role: Optional[str] = None,
         slack_token: Optional[str] = None,
-        *args, **kwargs
+        **kwargs
     ) -> None:
-        super(SnowflakeToSlackOperator, self).__init__(*args, **kwargs)
+        super(SnowflakeToSlackOperator, self).__init__(**kwargs)
 
         self.snowflake_conn_id = snowflake_conn_id
         self.sql = sql
diff --git a/airflow/sensors/base_sensor_operator.py b/airflow/sensors/base_sensor_operator.py
index f111996..3c74f3f 100644
--- a/airflow/sensors/base_sensor_operator.py
+++ b/airflow/sensors/base_sensor_operator.py
@@ -74,9 +74,8 @@ class BaseSensorOperator(BaseOperator, SkipMixin):
                  soft_fail: bool = False,
                  mode: str = 'poke',
                  exponential_backoff: bool = False,
-                 *args,
                  **kwargs) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.poke_interval = poke_interval
         self.soft_fail = soft_fail
         self.timeout = timeout
diff --git a/airflow/sensors/bash.py b/airflow/sensors/bash.py
index 4183773..20d8a81 100644
--- a/airflow/sensors/bash.py
+++ b/airflow/sensors/bash.py
@@ -49,8 +49,8 @@ class BashSensor(BaseSensorOperator):
                  bash_command,
                  env=None,
                  output_encoding='utf-8',
-                 *args, **kwargs):
-        super().__init__(*args, **kwargs)
+                 **kwargs):
+        super().__init__(**kwargs)
         self.bash_command = bash_command
         self.env = env
         self.output_encoding = output_encoding
diff --git a/airflow/sensors/date_time_sensor.py b/airflow/sensors/date_time_sensor.py
index 4af3bd2..d188b7f 100644
--- a/airflow/sensors/date_time_sensor.py
+++ b/airflow/sensors/date_time_sensor.py
@@ -58,9 +58,9 @@ class DateTimeSensor(BaseSensorOperator):
 
     @apply_defaults
     def __init__(
-        self, target_time: Union[str, datetime.datetime], *args, **kwargs
+        self, target_time: Union[str, datetime.datetime], **kwargs
     ) -> None:
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         if isinstance(target_time, datetime.datetime):
             self.target_time = target_time.isoformat()
         elif isinstance(target_time, str):
diff --git a/airflow/sensors/external_task_sensor.py b/airflow/sensors/external_task_sensor.py
index d9b47b5..63cedd1 100644
--- a/airflow/sensors/external_task_sensor.py
+++ b/airflow/sensors/external_task_sensor.py
@@ -74,9 +74,8 @@ class ExternalTaskSensor(BaseSensorOperator):
                  execution_delta=None,
                  execution_date_fn=None,
                  check_existence=False,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.allowed_states = allowed_states or [State.SUCCESS]
         self.failed_states = failed_states or []
 
@@ -248,9 +247,8 @@ class ExternalTaskMarker(DummyOperator):
                  external_task_id,
                  execution_date: Optional[Union[str, datetime.datetime]] = "{{ execution_date.isoformat() }}",
                  recursion_depth: int = 10,
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.external_dag_id = external_dag_id
         self.external_task_id = external_task_id
         if isinstance(execution_date, datetime.datetime):
diff --git a/airflow/sensors/filesystem.py b/airflow/sensors/filesystem.py
index 88554be..a06952c 100644
--- a/airflow/sensors/filesystem.py
+++ b/airflow/sensors/filesystem.py
@@ -46,9 +46,8 @@ class FileSensor(BaseSensorOperator):
     def __init__(self,
                  filepath,
                  fs_conn_id='fs_default',
-                 *args,
                  **kwargs):
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
         self.filepath = filepath
         self.fs_conn_id = fs_conn_id
 
diff --git a/airflow/sensors/python.py b/airflow/sensors/python.py
index 16621c1..fd31a2f 100644
--- a/airflow/sensors/python.py
+++ b/airflow/sensors/python.py
@@ -55,8 +55,8 @@ class PythonSensor(BaseSensorOperator):
             op_args: Optional[List] = None,
             op_kwargs: Optional[Dict] = None,
             templates_dict: Optional[Dict] = None,
-            *args, **kwargs):
-        super().__init__(*args, **kwargs)
+            **kwargs):
+        super().__init__(**kwargs)
         self.python_callable = python_callable
         self.op_args = op_args or []
         self.op_kwargs = op_kwargs or {}
diff --git a/airflow/sensors/sql_sensor.py b/airflow/sensors/sql_sensor.py
index 2d96784..70a8301 100644
--- a/airflow/sensors/sql_sensor.py
+++ b/airflow/sensors/sql_sensor.py
@@ -57,14 +57,14 @@ class SqlSensor(BaseSensorOperator):
 
     @apply_defaults
     def __init__(self, conn_id, sql, parameters=None, success=None, failure=None, fail_on_empty=False,
-                 *args, **kwargs):
+                 **kwargs):
         self.conn_id = conn_id
         self.sql = sql
         self.parameters = parameters
         self.success = success
         self.failure = failure
         self.fail_on_empty = fail_on_empty
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
     def _get_hook(self):
         conn = BaseHook.get_connection(self.conn_id)
diff --git a/airflow/sensors/time_delta_sensor.py b/airflow/sensors/time_delta_sensor.py
index 9b5b333..68f29ce 100644
--- a/airflow/sensors/time_delta_sensor.py
+++ b/airflow/sensors/time_delta_sensor.py
@@ -33,8 +33,8 @@ class TimeDeltaSensor(BaseSensorOperator):
     """
 
     @apply_defaults
-    def __init__(self, delta, *args, **kwargs):
-        super().__init__(*args, **kwargs)
+    def __init__(self, delta, **kwargs):
+        super().__init__(**kwargs)
         self.delta = delta
 
     def poke(self, context):
diff --git a/airflow/sensors/time_sensor.py b/airflow/sensors/time_sensor.py
index eb6846b..608a30c 100644
--- a/airflow/sensors/time_sensor.py
+++ b/airflow/sensors/time_sensor.py
@@ -30,8 +30,8 @@ class TimeSensor(BaseSensorOperator):
     """
 
     @apply_defaults
-    def __init__(self, target_time, *args, **kwargs):
-        super().__init__(*args, **kwargs)
+    def __init__(self, target_time, **kwargs):
+        super().__init__(**kwargs)
         self.target_time = target_time
 
     def poke(self, context):
diff --git a/airflow/sensors/weekday_sensor.py b/airflow/sensors/weekday_sensor.py
index 90b3834..e07fdd6 100644
--- a/airflow/sensors/weekday_sensor.py
+++ b/airflow/sensors/weekday_sensor.py
@@ -75,8 +75,8 @@ class DayOfWeekSensor(BaseSensorOperator):
     @apply_defaults
     def __init__(self, week_day,
                  use_task_execution_day=False,
-                 *args, **kwargs):
-        super().__init__(*args, **kwargs)
+                 **kwargs):
+        super().__init__(**kwargs)
         self.week_day = week_day
         self.use_task_execution_day = use_task_execution_day
         if isinstance(self.week_day, str):
diff --git a/docs/howto/custom-operator.rst b/docs/howto/custom-operator.rst
index 153f751..94c8240 100644
--- a/docs/howto/custom-operator.rst
+++ b/docs/howto/custom-operator.rst
@@ -47,8 +47,8 @@ Let's implement an example ``HelloOperator`` in a new file ``hello_operator.py``
             def __init__(
                     self,
                     name: str,
-                    *args, **kwargs) -> None:
-                super().__init__(*args, **kwargs)
+                    **kwargs) -> None:
+                super().__init__(**kwargs)
                 self.name = name
 
             def execute(self, context):
@@ -104,8 +104,8 @@ Let's extend our previous example to fetch name from MySQL:
                     name: str,
                     mysql_conn_id: str,
                     database: str,
-                    *args, **kwargs) -> None:
-                super().__init__(*args, **kwargs)
+                    **kwargs) -> None:
+                super().__init__(**kwargs)
                 self.name = name
                 self.mysql_conn_id = mysql_conn_id
                 self.database = database
@@ -157,8 +157,8 @@ the operator.
             def __init__(
                     self,
                     name: str,
-                    *args, **kwargs) -> None:
-                super().__init__(*args, **kwargs)
+                    **kwargs) -> None:
+                super().__init__(**kwargs)
                 self.name = name
 
             def execute(self, context):
@@ -193,8 +193,8 @@ with actual value. Note that Jinja substitutes the operator attributes and not t
             def __init__(
                     self,
                     name: str,
-                    *args, **kwargs) -> None:
-                super().__init__(*args, **kwargs)
+                    **kwargs) -> None:
+                super().__init__(**kwargs)
                 self.guest_name = name
 
 In the example, the ``template_fields`` should be ``['guest_name']`` and not  ``['name']``
diff --git a/docs/howto/define_extra_link.rst b/docs/howto/define_extra_link.rst
index 47bb5ed..bded978 100644
--- a/docs/howto/define_extra_link.rst
+++ b/docs/howto/define_extra_link.rst
@@ -48,8 +48,8 @@ The following code shows how to add extra links to an operator:
         )
 
         @apply_defaults
-        def __init__(self, *args, **kwargs):
-            super().__init__(*args, **kwargs)
+        def __init__(self, **kwargs):
+            super().__init__(**kwargs)
 
         def execute(self, context):
             self.log.info("Hello World!")
diff --git a/tests/models/test_baseoperator.py b/tests/models/test_baseoperator.py
index fa6804b..89d6605 100644
--- a/tests/models/test_baseoperator.py
+++ b/tests/models/test_baseoperator.py
@@ -363,8 +363,8 @@ class CustomOp(DummyOperator):
     template_fields = ("field", "field2")
 
     @apply_defaults
-    def __init__(self, field=None, field2=None, *args, **kwargs):
-        super().__init__(*args, **kwargs)
+    def __init__(self, field=None, field2=None, **kwargs):
+        super().__init__(**kwargs)
         self.field = field
         self.field2 = field2
 
diff --git a/tests/sensors/test_timeout_sensor.py b/tests/sensors/test_timeout_sensor.py
index 810a6e1..70228dd 100644
--- a/tests/sensors/test_timeout_sensor.py
+++ b/tests/sensors/test_timeout_sensor.py
@@ -43,10 +43,9 @@ class TimeoutTestSensor(BaseSensorOperator):
     @apply_defaults
     def __init__(self,
                  return_value=False,
-                 *args,
                  **kwargs):
         self.return_value = return_value
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
     def poke(self, context):
         return self.return_value
diff --git a/tests/test_utils/mock_operators.py b/tests/test_utils/mock_operators.py
index 6a25d9c..36bca43 100644
--- a/tests/test_utils/mock_operators.py
+++ b/tests/test_utils/mock_operators.py
@@ -121,8 +121,8 @@ class CustomOperator(BaseOperator):
         )
 
     @apply_defaults
-    def __init__(self, bash_command=None, *args, **kwargs):
-        super(CustomOperator, self).__init__(*args, **kwargs)
+    def __init__(self, bash_command=None, **kwargs):
+        super(CustomOperator, self).__init__(**kwargs)
         self.bash_command = bash_command
 
     def execute(self, context):
@@ -170,9 +170,9 @@ class MockHiveOperator(HiveOperator):
 
 class DeprecatedOperator(BaseOperator):
     @apply_defaults
-    def __init__(self, *args, **kwargs):
+    def __init__(self, **kwargs):
         warnings.warn("This operator is deprecated.", DeprecationWarning, stacklevel=4)
-        super().__init__(*args, **kwargs)
+        super().__init__(**kwargs)
 
     def execute(self, context):
         pass
diff --git a/tests/utils/test_decorators.py b/tests/utils/test_decorators.py
index d2d5939..94920b9 100644
--- a/tests/utils/test_decorators.py
+++ b/tests/utils/test_decorators.py
@@ -31,8 +31,8 @@ class DummyClass:
 
 class DummySubClass(DummyClass):
     @apply_defaults
-    def __init__(self, test_sub_param, *args, **kwargs):
-        super().__init__(*args, **kwargs)
+    def __init__(self, test_sub_param, **kwargs):
+        super().__init__(**kwargs)
         self.test_sub_param = test_sub_param