You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by el...@apache.org on 2022/07/08 20:45:23 UTC

[airflow] branch main updated: Convert RDS Event and Snapshot Sample DAGs to System Tests (#24932)

This is an automated email from the ASF dual-hosted git repository.

eladkal pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/main by this push:
     new 109acc75c4 Convert RDS Event and Snapshot Sample DAGs to System Tests (#24932)
109acc75c4 is described below

commit 109acc75c4b94a1155b0469146ca5b5b5d6040f1
Author: D. Ferruzzi <fe...@amazon.com>
AuthorDate: Fri Jul 8 20:45:00 2022 +0000

    Convert RDS Event and Snapshot Sample DAGs to System Tests (#24932)
---
 .../amazon/aws/example_dags/example_rds_event.py   |  58 --------
 .../aws/example_dags/example_rds_snapshot.py       |  76 ----------
 .../operators/rds.rst                              |  12 +-
 .../providers/amazon/aws/example_rds_event.py      | 134 ++++++++++++++++++
 .../providers/amazon/aws/example_rds_snapshot.py   | 155 +++++++++++++++++++++
 5 files changed, 295 insertions(+), 140 deletions(-)

diff --git a/airflow/providers/amazon/aws/example_dags/example_rds_event.py b/airflow/providers/amazon/aws/example_dags/example_rds_event.py
deleted file mode 100644
index 4ec8b6f5be..0000000000
--- a/airflow/providers/amazon/aws/example_dags/example_rds_event.py
+++ /dev/null
@@ -1,58 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-from datetime import datetime
-from os import getenv
-
-from airflow import DAG
-from airflow.models.baseoperator import chain
-from airflow.providers.amazon.aws.operators.rds import (
-    RdsCreateEventSubscriptionOperator,
-    RdsDeleteEventSubscriptionOperator,
-)
-
-SUBSCRIPTION_NAME = getenv("SUBSCRIPTION_NAME", "subscription-name")
-SNS_TOPIC_ARN = getenv("SNS_TOPIC_ARN", "arn:aws:sns:<region>:<account number>:MyTopic")
-RDS_DB_IDENTIFIER = getenv("RDS_DB_IDENTIFIER", "database-identifier")
-
-with DAG(
-    dag_id='example_rds_event',
-    schedule_interval=None,
-    start_date=datetime(2021, 1, 1),
-    tags=['example'],
-    catchup=False,
-) as dag:
-    # [START howto_operator_rds_create_event_subscription]
-    create_subscription = RdsCreateEventSubscriptionOperator(
-        task_id='create_subscription',
-        subscription_name=SUBSCRIPTION_NAME,
-        sns_topic_arn=SNS_TOPIC_ARN,
-        source_type='db-instance',
-        source_ids=[RDS_DB_IDENTIFIER],
-        event_categories=['availability'],
-    )
-    # [END howto_operator_rds_create_event_subscription]
-
-    # [START howto_operator_rds_delete_event_subscription]
-    delete_subscription = RdsDeleteEventSubscriptionOperator(
-        task_id='delete_subscription',
-        subscription_name=SUBSCRIPTION_NAME,
-    )
-    # [END howto_operator_rds_delete_event_subscription]
-
-    chain(create_subscription, delete_subscription)
diff --git a/airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py b/airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
deleted file mode 100644
index f7e1d02e07..0000000000
--- a/airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
+++ /dev/null
@@ -1,76 +0,0 @@
-#
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-from datetime import datetime
-from os import getenv
-
-from airflow import DAG
-from airflow.models.baseoperator import chain
-from airflow.providers.amazon.aws.operators.rds import (
-    RdsCopyDbSnapshotOperator,
-    RdsCreateDbSnapshotOperator,
-    RdsDeleteDbSnapshotOperator,
-)
-from airflow.providers.amazon.aws.sensors.rds import RdsSnapshotExistenceSensor
-
-RDS_DB_IDENTIFIER = getenv("RDS_DB_IDENTIFIER", "database-identifier")
-RDS_DB_SNAPSHOT_IDENTIFIER = getenv("RDS_DB_SNAPSHOT_IDENTIFIER", "database-1-snap")
-
-with DAG(
-    dag_id='example_rds_snapshot',
-    schedule_interval=None,
-    start_date=datetime(2021, 1, 1),
-    tags=['example'],
-    catchup=False,
-) as dag:
-    # [START howto_operator_rds_create_db_snapshot]
-    create_snapshot = RdsCreateDbSnapshotOperator(
-        task_id='create_snapshot',
-        db_type='instance',
-        db_identifier=RDS_DB_IDENTIFIER,
-        db_snapshot_identifier=RDS_DB_SNAPSHOT_IDENTIFIER,
-    )
-    # [END howto_operator_rds_create_db_snapshot]
-
-    # [START howto_sensor_rds_snapshot_existence]
-    snapshot_sensor = RdsSnapshotExistenceSensor(
-        task_id='snapshot_sensor',
-        db_type='instance',
-        db_snapshot_identifier=RDS_DB_IDENTIFIER,
-        target_statuses=['available'],
-    )
-    # [END howto_sensor_rds_snapshot_existence]
-
-    # [START howto_operator_rds_copy_snapshot]
-    copy_snapshot = RdsCopyDbSnapshotOperator(
-        task_id='copy_snapshot',
-        db_type='instance',
-        source_db_snapshot_identifier=RDS_DB_IDENTIFIER,
-        target_db_snapshot_identifier=f'{RDS_DB_IDENTIFIER}-copy',
-    )
-    # [END howto_operator_rds_copy_snapshot]
-
-    # [START howto_operator_rds_delete_snapshot]
-    delete_snapshot = RdsDeleteDbSnapshotOperator(
-        task_id='delete_snapshot',
-        db_type='instance',
-        db_snapshot_identifier=RDS_DB_IDENTIFIER,
-    )
-    # [END howto_operator_rds_delete_snapshot]
-
-    chain(create_snapshot, snapshot_sensor, copy_snapshot, delete_snapshot)
diff --git a/docs/apache-airflow-providers-amazon/operators/rds.rst b/docs/apache-airflow-providers-amazon/operators/rds.rst
index 434804d910..022b9ca8f1 100644
--- a/docs/apache-airflow-providers-amazon/operators/rds.rst
+++ b/docs/apache-airflow-providers-amazon/operators/rds.rst
@@ -41,7 +41,7 @@ To create a snapshot of an Amazon RDS database instance or cluster you can use
 :class:`~airflow.providers.amazon.aws.operators.rds.RDSCreateDBSnapshotOperator`.
 The source database instance must be in the ``available`` or ``storage-optimization`` state.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_snapshot.py
     :language: python
     :dedent: 4
     :start-after: [START howto_operator_rds_create_db_snapshot]
@@ -56,7 +56,7 @@ To copy a snapshot of an Amazon RDS database instance or cluster you can use
 :class:`~airflow.providers.amazon.aws.operators.rds.RDSCopyDBSnapshotOperator`.
 The source database snapshot must be in the ``available`` state.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_snapshot.py
     :language: python
     :dedent: 4
     :start-after: [START howto_operator_rds_copy_snapshot]
@@ -71,7 +71,7 @@ To delete a snapshot of an Amazon RDS database instance or cluster you can use
 :class:`~airflow.providers.amazon.aws.operators.rds.RDSDeleteDBSnapshotOperator`.
 The database snapshot must be in the ``available`` state to be deleted.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_snapshot.py
     :language: python
     :dedent: 4
     :start-after: [START howto_operator_rds_delete_snapshot]
@@ -118,7 +118,7 @@ This action requires an Amazon SNS topic Amazon Resource Name (ARN).
 Amazon RDS event notification is only available for not encrypted SNS topics.
 If you specify an encrypted SNS topic, event notifications are not sent for the topic.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_event.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_event.py
     :language: python
     :dedent: 4
     :start-after: [START howto_operator_rds_create_event_subscription]
@@ -132,7 +132,7 @@ Unsubscribe to an Amazon RDS event notification
 To delete an Amazon RDS event subscription you can use
 :class:`~airflow.providers.amazon.aws.operators.rds.RDSDeleteEventSubscriptionOperator`.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_event.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_event.py
     :language: python
     :dedent: 4
     :start-after: [START howto_operator_rds_delete_event_subscription]
@@ -178,7 +178,7 @@ To wait for an Amazon RDS snapshot with specific statuses you can use
 :class:`~airflow.providers.amazon.aws.sensors.rds.RdsSnapshotExistenceSensor`.
 By default, the sensor waits for the existence of a snapshot with status ``available``.
 
-.. exampleinclude:: /../../airflow/providers/amazon/aws/example_dags/example_rds_snapshot.py
+.. exampleinclude:: /../../tests/system/providers/amazon/aws/example_rds_snapshot.py
     :language: python
     :dedent: 4
     :start-after: [START howto_sensor_rds_snapshot_existence]
diff --git a/tests/system/providers/amazon/aws/example_rds_event.py b/tests/system/providers/amazon/aws/example_rds_event.py
new file mode 100644
index 0000000000..cb19cce09b
--- /dev/null
+++ b/tests/system/providers/amazon/aws/example_rds_event.py
@@ -0,0 +1,134 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from datetime import datetime
+
+import boto3
+
+from airflow import DAG
+from airflow.decorators import task
+from airflow.models.baseoperator import chain
+from airflow.providers.amazon.aws.hooks.rds import RdsHook
+from airflow.providers.amazon.aws.operators.rds import (
+    RdsCreateEventSubscriptionOperator,
+    RdsDeleteEventSubscriptionOperator,
+)
+from airflow.utils.trigger_rule import TriggerRule
+from tests.system.providers.amazon.aws.utils import ENV_ID_KEY, SystemTestContextBuilder
+
+DAG_ID = 'example_rds_event'
+
+sys_test_context_task = SystemTestContextBuilder().build()
+
+
+@task
+def create_sns_topic(env_id) -> str:
+    return boto3.client('sns').create_topic(Name=f'{env_id}-topic')['TopicArn']
+
+
+@task
+def create_rds_instance(db_name, instance_name) -> None:
+    rds_client = RdsHook().get_conn()
+    rds_client.create_db_instance(
+        DBName=db_name,
+        DBInstanceIdentifier=instance_name,
+        AllocatedStorage=20,
+        DBInstanceClass='db.t3.micro',
+        Engine='postgres',
+        MasterUsername='username',
+        # NEVER store your production password in plaintext in a DAG like this.
+        # Use Airflow Secrets or a secret manager for this in production.
+        MasterUserPassword='rds_password',
+    )
+
+    rds_client.get_waiter('db_instance_available').wait(DBInstanceIdentifier=instance_name)
+
+
+@task(trigger_rule=TriggerRule.ALL_DONE)
+def delete_db_instance(instance_name) -> None:
+    rds_client = RdsHook().get_conn()
+    rds_client.delete_db_instance(
+        DBInstanceIdentifier=instance_name,
+        SkipFinalSnapshot=True,
+    )
+
+    rds_client.get_waiter('db_instance_deleted').wait(DBInstanceIdentifier=instance_name)
+
+
+@task(trigger_rule=TriggerRule.ALL_DONE)
+def delete_sns_topic(topic_arn) -> None:
+    boto3.client('sns').delete_topic(TopicArn=topic_arn)
+
+
+with DAG(
+    dag_id=DAG_ID,
+    schedule_interval='@once',
+    start_date=datetime(2021, 1, 1),
+    tags=['example'],
+    catchup=False,
+) as dag:
+    test_context = sys_test_context_task()
+
+    rds_db_name = f'{test_context[ENV_ID_KEY]}_db'
+    rds_instance_name = f'{test_context[ENV_ID_KEY]}-instance'
+    rds_subscription_name = f'{test_context[ENV_ID_KEY]}-subscription'
+
+    sns_topic = create_sns_topic(test_context[ENV_ID_KEY])
+
+    # [START howto_operator_rds_create_event_subscription]
+    create_subscription = RdsCreateEventSubscriptionOperator(
+        task_id='create_subscription',
+        subscription_name=rds_subscription_name,
+        sns_topic_arn=sns_topic,
+        source_type='db-instance',
+        source_ids=[rds_instance_name],
+        event_categories=['availability'],
+    )
+    # [END howto_operator_rds_create_event_subscription]
+
+    # [START howto_operator_rds_delete_event_subscription]
+    delete_subscription = RdsDeleteEventSubscriptionOperator(
+        task_id='delete_subscription',
+        subscription_name=rds_subscription_name,
+    )
+    # [END howto_operator_rds_delete_event_subscription]
+
+    chain(
+        # TEST SETUP
+        test_context,
+        sns_topic,
+        create_rds_instance(rds_db_name, rds_instance_name),
+        # TEST BODY
+        create_subscription,
+        delete_subscription,
+        # TEST TEARDOWN
+        delete_db_instance(rds_instance_name),
+        delete_sns_topic(sns_topic),
+    )
+
+    from tests.system.utils.watcher import watcher
+
+    # This test needs watcher in order to properly mark success/failure
+    # when "tearDown" task with trigger rule is part of the DAG
+    list(dag.tasks) >> watcher()
+
+
+from tests.system.utils import get_test_run  # noqa: E402
+
+# Needed to run the example DAG with pytest (see: tests/system/README.md#run_via_pytest)
+test_run = get_test_run(dag)
diff --git a/tests/system/providers/amazon/aws/example_rds_snapshot.py b/tests/system/providers/amazon/aws/example_rds_snapshot.py
new file mode 100644
index 0000000000..2bde55287b
--- /dev/null
+++ b/tests/system/providers/amazon/aws/example_rds_snapshot.py
@@ -0,0 +1,155 @@
+#
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+from datetime import datetime
+
+from airflow import DAG
+from airflow.decorators import task
+from airflow.models.baseoperator import chain
+from airflow.providers.amazon.aws.hooks.rds import RdsHook
+from airflow.providers.amazon.aws.operators.rds import (
+    RdsCopyDbSnapshotOperator,
+    RdsCreateDbSnapshotOperator,
+    RdsDeleteDbSnapshotOperator,
+)
+from airflow.providers.amazon.aws.sensors.rds import RdsSnapshotExistenceSensor
+from airflow.utils.trigger_rule import TriggerRule
+from tests.system.providers.amazon.aws.utils import ENV_ID_KEY, SystemTestContextBuilder
+
+DAG_ID = 'example_rds_snapshot'
+
+sys_test_context_task = SystemTestContextBuilder().build()
+
+
+@task
+def create_rds_instance(db_name, instance_name) -> None:
+    rds_client = RdsHook().get_conn()
+    rds_client.create_db_instance(
+        DBName=db_name,
+        DBInstanceIdentifier=instance_name,
+        AllocatedStorage=20,
+        DBInstanceClass='db.t3.micro',
+        Engine='postgres',
+        MasterUsername='username',
+        # NEVER store your production password in plaintext in a DAG like this.
+        # Use Airflow Secrets or a secret manager for this in production.
+        MasterUserPassword='rds_password',
+    )
+
+    rds_client.get_waiter('db_instance_available').wait(DBInstanceIdentifier=instance_name)
+
+
+@task(trigger_rule=TriggerRule.ALL_DONE)
+def delete_rds_instance(instance_name) -> None:
+    rds_client = RdsHook().get_conn()
+    rds_client.delete_db_instance(
+        DBInstanceIdentifier=instance_name,
+        SkipFinalSnapshot=True,
+    )
+
+    rds_client.get_waiter('db_instance_deleted').wait(DBInstanceIdentifier=instance_name)
+
+
+with DAG(
+    dag_id=DAG_ID,
+    schedule_interval='@once',
+    start_date=datetime(2021, 1, 1),
+    tags=['example'],
+    catchup=False,
+) as dag:
+    test_context = sys_test_context_task()
+
+    rds_db_name = f'{test_context[ENV_ID_KEY]}_db'
+    rds_instance_name = f'{test_context[ENV_ID_KEY]}-instance'
+    rds_snapshot_name = f'{test_context[ENV_ID_KEY]}-snapshot'
+    rds_snapshot_copy_name = f'{rds_snapshot_name}-copy'
+
+    # [START howto_operator_rds_create_db_snapshot]
+    create_snapshot = RdsCreateDbSnapshotOperator(
+        task_id='create_snapshot',
+        db_type='instance',
+        db_identifier=rds_instance_name,
+        db_snapshot_identifier=rds_snapshot_name,
+    )
+    # [END howto_operator_rds_create_db_snapshot]
+
+    # [START howto_sensor_rds_snapshot_existence]
+    snapshot_sensor = RdsSnapshotExistenceSensor(
+        task_id='snapshot_sensor',
+        db_type='instance',
+        db_snapshot_identifier=rds_snapshot_name,
+        target_statuses=['available'],
+    )
+    # [END howto_sensor_rds_snapshot_existence]
+
+    # [START howto_operator_rds_copy_snapshot]
+    copy_snapshot = RdsCopyDbSnapshotOperator(
+        task_id='copy_snapshot',
+        db_type='instance',
+        source_db_snapshot_identifier=rds_snapshot_name,
+        target_db_snapshot_identifier=rds_snapshot_copy_name,
+    )
+    # [END howto_operator_rds_copy_snapshot]
+
+    # [START howto_operator_rds_delete_snapshot]
+    delete_snapshot = RdsDeleteDbSnapshotOperator(
+        task_id='delete_snapshot',
+        db_type='instance',
+        db_snapshot_identifier=rds_snapshot_name,
+    )
+    # [END howto_operator_rds_delete_snapshot]
+
+    snapshot_copy_sensor = RdsSnapshotExistenceSensor(
+        task_id='snapshot_copy_sensor',
+        db_type='instance',
+        db_snapshot_identifier=rds_snapshot_copy_name,
+        target_statuses=['available'],
+    )
+
+    delete_snapshot_copy = RdsDeleteDbSnapshotOperator(
+        task_id='delete_snapshot_copy',
+        db_type='instance',
+        db_snapshot_identifier=rds_snapshot_copy_name,
+    )
+
+    chain(
+        # TEST SETUP
+        test_context,
+        create_rds_instance(rds_db_name, rds_instance_name),
+        # TEST BODY
+        create_snapshot,
+        snapshot_sensor,
+        copy_snapshot,
+        delete_snapshot,
+        # TEST TEARDOWN
+        snapshot_copy_sensor,
+        delete_snapshot_copy,
+        delete_rds_instance(rds_instance_name),
+    )
+
+    from tests.system.utils.watcher import watcher
+
+    # This test needs watcher in order to properly mark success/failure
+    # when "tearDown" task with trigger rule is part of the DAG
+    list(dag.tasks) >> watcher()
+
+
+from tests.system.utils import get_test_run  # noqa: E402
+
+# Needed to run the example DAG with pytest (see: tests/system/README.md#run_via_pytest)
+test_run = get_test_run(dag)