You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by po...@apache.org on 2022/06/20 09:05:43 UTC

[airflow] branch main updated: Migrate Google example DAG oracle_to_gcs to new design AIP-47 (#24542)

This is an automated email from the ASF dual-hosted git repository.

potiuk pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/main by this push:
     new 88ddf65708 Migrate Google example DAG oracle_to_gcs to new design AIP-47 (#24542)
88ddf65708 is described below

commit 88ddf65708ce65bc166e7e0d0bd21c3896dc3053
Author: Chenglong Yan <al...@gmail.com>
AuthorDate: Mon Jun 20 17:05:35 2022 +0800

    Migrate Google example DAG oracle_to_gcs to new design AIP-47 (#24542)
    
    related: #22447, #22430
---
 .../cloud/example_dags/example_oracle_to_gcs.py    | 40 ------------
 .../operators/transfer/oracle_to_gcs.rst           |  2 +-
 .../google/cloud/gcs/example_oracle_to_gcs.py      | 73 ++++++++++++++++++++++
 3 files changed, 74 insertions(+), 41 deletions(-)

diff --git a/airflow/providers/google/cloud/example_dags/example_oracle_to_gcs.py b/airflow/providers/google/cloud/example_dags/example_oracle_to_gcs.py
deleted file mode 100644
index 2d5d5c59bf..0000000000
--- a/airflow/providers/google/cloud/example_dags/example_oracle_to_gcs.py
+++ /dev/null
@@ -1,40 +0,0 @@
-# Licensed to the Apache Software Foundation (ASF) under one
-# or more contributor license agreements.  See the NOTICE file
-# distributed with this work for additional information
-# regarding copyright ownership.  The ASF licenses this file
-# to you under the Apache License, Version 2.0 (the
-# "License"); you may not use this file except in compliance
-# with the License.  You may obtain a copy of the License at
-#
-#   http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing,
-# software distributed under the License is distributed on an
-# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
-# KIND, either express or implied.  See the License for the
-# specific language governing permissions and limitations
-# under the License.
-
-import os
-from datetime import datetime
-
-from airflow import models
-from airflow.providers.google.cloud.transfers.oracle_to_gcs import OracleToGCSOperator
-
-GCS_BUCKET = os.environ.get("GCP_GCS_BUCKET", "example-airflow-oracle-gcs")
-FILENAME = 'test_file'
-
-SQL_QUERY = "SELECT * from test_table"
-
-with models.DAG(
-    'example_oracle_to_gcs',
-    schedule_interval=None,
-    start_date=datetime(2021, 1, 1),
-    catchup=False,
-    tags=['example'],
-) as dag:
-    # [START howto_operator_oracle_to_gcs]
-    upload = OracleToGCSOperator(
-        task_id='oracle_to_gcs', sql=SQL_QUERY, bucket=GCS_BUCKET, filename=FILENAME, export_format='csv'
-    )
-    # [END howto_operator_oracle_to_gcs]
diff --git a/docs/apache-airflow-providers-google/operators/transfer/oracle_to_gcs.rst b/docs/apache-airflow-providers-google/operators/transfer/oracle_to_gcs.rst
index 2c23209ae4..0b0b00379f 100644
--- a/docs/apache-airflow-providers-google/operators/transfer/oracle_to_gcs.rst
+++ b/docs/apache-airflow-providers-google/operators/transfer/oracle_to_gcs.rst
@@ -38,7 +38,7 @@ When you use this operator, you can optionally compress the data being uploaded
 
 Below is an example of using this operator to upload data to GCS.
 
-.. exampleinclude:: /../../airflow/providers/google/cloud/example_dags/example_oracle_to_gcs.py
+.. exampleinclude:: /../../tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py
     :language: python
     :dedent: 0
     :start-after: [START howto_operator_oracle_to_gcs]
diff --git a/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py b/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py
new file mode 100644
index 0000000000..5c32b1b400
--- /dev/null
+++ b/tests/system/providers/google/cloud/gcs/example_oracle_to_gcs.py
@@ -0,0 +1,73 @@
+# Licensed to the Apache Software Foundation (ASF) under one
+# or more contributor license agreements.  See the NOTICE file
+# distributed with this work for additional information
+# regarding copyright ownership.  The ASF licenses this file
+# to you under the Apache License, Version 2.0 (the
+# "License"); you may not use this file except in compliance
+# with the License.  You may obtain a copy of the License at
+#
+#   http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing,
+# software distributed under the License is distributed on an
+# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+# KIND, either express or implied.  See the License for the
+# specific language governing permissions and limitations
+# under the License.
+
+import os
+from datetime import datetime
+
+from airflow import models
+from airflow.providers.google.cloud.operators.gcs import GCSCreateBucketOperator, GCSDeleteBucketOperator
+from airflow.providers.google.cloud.transfers.oracle_to_gcs import OracleToGCSOperator
+from airflow.utils.trigger_rule import TriggerRule
+
+ENV_ID = os.environ.get("SYSTEM_TESTS_ENV_ID")
+PROJECT_ID = os.environ.get("SYSTEM_TESTS_GCP_PROJECT")
+DAG_ID = "example_oracle_to_gcs"
+
+BUCKET_NAME = f"bucket_{DAG_ID}_{ENV_ID}"
+FILENAME = 'test_file'
+SQL_QUERY = "SELECT * from test_table"
+
+with models.DAG(
+    DAG_ID,
+    schedule_interval='@once',
+    start_date=datetime(2021, 1, 1),
+    catchup=False,
+    tags=['example', 'oracle'],
+) as dag:
+    create_bucket = GCSCreateBucketOperator(
+        task_id="create_bucket", bucket_name=BUCKET_NAME, project_id=PROJECT_ID
+    )
+
+    # [START howto_operator_oracle_to_gcs]
+    upload_oracle_to_gcs = OracleToGCSOperator(
+        task_id='oracle_to_gcs', sql=SQL_QUERY, bucket=BUCKET_NAME, filename=FILENAME, export_format='csv'
+    )
+    # [END howto_operator_oracle_to_gcs]
+
+    delete_bucket = GCSDeleteBucketOperator(
+        task_id="delete_bucket", bucket_name=BUCKET_NAME, trigger_rule=TriggerRule.ALL_DONE
+    )
+
+    (
+        # TEST SETUP
+        create_bucket
+        # TEST BODY
+        >> upload_oracle_to_gcs
+        # TEST TEARDOWN
+        >> delete_bucket
+    )
+
+    from tests.system.utils.watcher import watcher
+
+    # This test needs watcher in order to properly mark success/failure
+    # when "tearDown" task with trigger rule is part of the DAG
+    list(dag.tasks) >> watcher()
+
+from tests.system.utils import get_test_run  # noqa: E402
+
+# Needed to run the example DAG with pytest (see: tests/system/README.md#run_via_pytest)
+test_run = get_test_run(dag)