You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by ka...@apache.org on 2018/08/27 16:26:53 UTC
[33/51] [partial] incubator-airflow-site git commit: 1.10.0
http://git-wip-us.apache.org/repos/asf/incubator-airflow-site/blob/11437c14/_modules/airflow/contrib/operators/dataproc_operator.html
----------------------------------------------------------------------
diff --git a/_modules/airflow/contrib/operators/dataproc_operator.html b/_modules/airflow/contrib/operators/dataproc_operator.html
index d49719e..f26e459 100644
--- a/_modules/airflow/contrib/operators/dataproc_operator.html
+++ b/_modules/airflow/contrib/operators/dataproc_operator.html
@@ -91,7 +91,7 @@
<li class="toctree-l1"><a class="reference internal" href="../../../../start.html">Quick Start</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../installation.html">Installation</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../tutorial.html">Tutorial</a></li>
-<li class="toctree-l1"><a class="reference internal" href="../../../../configuration.html">Configuration</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../howto/index.html">How-to Guides</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../ui.html">UI / Screenshots</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../concepts.html">Concepts</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../profiling.html">Data Profiling</a></li>
@@ -99,8 +99,10 @@
<li class="toctree-l1"><a class="reference internal" href="../../../../scheduler.html">Scheduling & Triggers</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../plugins.html">Plugins</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../security.html">Security</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../timezone.html">Time zones</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../api.html">Experimental Rest API</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../integration.html">Integration</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../lineage.html">Lineage</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../faq.html">FAQ</a></li>
<li class="toctree-l1"><a class="reference internal" href="../../../../code.html">API Reference</a></li>
</ul>
@@ -169,29 +171,42 @@
<h1>Source code for airflow.contrib.operators.dataproc_operator</h1><div class="highlight"><pre>
<span></span><span class="c1"># -*- coding: utf-8 -*-</span>
<span class="c1">#</span>
-<span class="c1"># Licensed under the Apache License, Version 2.0 (the "License");</span>
-<span class="c1"># you may not use this file except in compliance with the License.</span>
-<span class="c1"># You may obtain a copy of the License at</span>
+<span class="c1"># Licensed to the Apache Software Foundation (ASF) under one</span>
+<span class="c1"># or more contributor license agreements. See the NOTICE file</span>
+<span class="c1"># distributed with this work for additional information</span>
+<span class="c1"># regarding copyright ownership. The ASF licenses this file</span>
+<span class="c1"># to you under the Apache License, Version 2.0 (the</span>
+<span class="c1"># "License"); you may not use this file except in compliance</span>
+<span class="c1"># with the License. You may obtain a copy of the License at</span>
<span class="c1">#</span>
-<span class="c1"># http://www.apache.org/licenses/LICENSE-2.0</span>
+<span class="c1"># http://www.apache.org/licenses/LICENSE-2.0</span>
<span class="c1">#</span>
-<span class="c1"># Unless required by applicable law or agreed to in writing, software</span>
-<span class="c1"># distributed under the License is distributed on an "AS IS" BASIS,</span>
-<span class="c1"># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.</span>
-<span class="c1"># See the License for the specific language governing permissions and</span>
-<span class="c1"># limitations under the License.</span>
+<span class="c1"># Unless required by applicable law or agreed to in writing,</span>
+<span class="c1"># software distributed under the License is distributed on an</span>
+<span class="c1"># "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY</span>
+<span class="c1"># KIND, either express or implied. See the License for the</span>
+<span class="c1"># specific language governing permissions and limitations</span>
+<span class="c1"># under the License.</span>
<span class="c1">#</span>
+<span class="kn">import</span> <span class="nn">ntpath</span>
+<span class="kn">import</span> <span class="nn">os</span>
+<span class="kn">import</span> <span class="nn">re</span>
<span class="kn">import</span> <span class="nn">time</span>
+<span class="kn">import</span> <span class="nn">uuid</span>
+<span class="kn">from</span> <span class="nn">datetime</span> <span class="k">import</span> <span class="n">timedelta</span>
<span class="kn">from</span> <span class="nn">airflow.contrib.hooks.gcp_dataproc_hook</span> <span class="k">import</span> <span class="n">DataProcHook</span>
+<span class="kn">from</span> <span class="nn">airflow.contrib.hooks.gcs_hook</span> <span class="k">import</span> <span class="n">GoogleCloudStorageHook</span>
+<span class="kn">from</span> <span class="nn">airflow.exceptions</span> <span class="k">import</span> <span class="n">AirflowException</span>
<span class="kn">from</span> <span class="nn">airflow.models</span> <span class="k">import</span> <span class="n">BaseOperator</span>
<span class="kn">from</span> <span class="nn">airflow.utils.decorators</span> <span class="k">import</span> <span class="n">apply_defaults</span>
<span class="kn">from</span> <span class="nn">airflow.version</span> <span class="k">import</span> <span class="n">version</span>
<span class="kn">from</span> <span class="nn">googleapiclient.errors</span> <span class="k">import</span> <span class="n">HttpError</span>
+<span class="kn">from</span> <span class="nn">airflow.utils</span> <span class="k">import</span> <span class="n">timezone</span>
-<span class="k">class</span> <span class="nc">DataprocClusterCreateOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
+<div class="viewcode-block" id="DataprocClusterCreateOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataprocClusterCreateOperator">[docs]</a><span class="k">class</span> <span class="nc">DataprocClusterCreateOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Create a new cluster on Google Cloud Dataproc. The operator will wait until the</span>
<span class="sd"> creation is successful or an error occurs in the creation process.</span>
@@ -202,9 +217,79 @@
<span class="sd"> for a detailed explanation on the different parameters. Most of the configuration</span>
<span class="sd"> parameters detailed in the link are available as a parameter to this operator.</span>
+
+<span class="sd"> :param cluster_name: The name of the DataProc cluster to create.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param project_id: The ID of the google cloud project in which</span>
+<span class="sd"> to create the cluster</span>
+<span class="sd"> :type project_id: string</span>
+<span class="sd"> :param num_workers: The # of workers to spin up</span>
+<span class="sd"> :type num_workers: int</span>
+<span class="sd"> :param storage_bucket: The storage bucket to use, setting to None lets dataproc</span>
+<span class="sd"> generate a custom one for you</span>
+<span class="sd"> :type storage_bucket: string</span>
+<span class="sd"> :param init_actions_uris: List of GCS uri's containing</span>
+<span class="sd"> dataproc initialization scripts</span>
+<span class="sd"> :type init_actions_uris: list[string]</span>
+<span class="sd"> :param init_action_timeout: Amount of time executable scripts in</span>
+<span class="sd"> init_actions_uris has to complete</span>
+<span class="sd"> :type init_action_timeout: string</span>
+<span class="sd"> :param metadata: dict of key-value google compute engine metadata entries</span>
+<span class="sd"> to add to all instances</span>
+<span class="sd"> :type metadata: dict</span>
+<span class="sd"> :param image_version: the version of software inside the Dataproc cluster</span>
+<span class="sd"> :type image_version: string</span>
+<span class="sd"> :param properties: dict of properties to set on</span>
+<span class="sd"> config files (e.g. spark-defaults.conf), see</span>
+<span class="sd"> https://cloud.google.com/dataproc/docs/reference/rest/v1/ \</span>
+<span class="sd"> projects.regions.clusters#SoftwareConfig</span>
+<span class="sd"> :type properties: dict</span>
+<span class="sd"> :param master_machine_type: Compute engine machine type to use for the master node</span>
+<span class="sd"> :type master_machine_type: string</span>
+<span class="sd"> :param master_disk_size: Disk size for the master node</span>
+<span class="sd"> :type master_disk_size: int</span>
+<span class="sd"> :param worker_machine_type: Compute engine machine type to use for the worker nodes</span>
+<span class="sd"> :type worker_machine_type: string</span>
+<span class="sd"> :param worker_disk_size: Disk size for the worker nodes</span>
+<span class="sd"> :type worker_disk_size: int</span>
+<span class="sd"> :param num_preemptible_workers: The # of preemptible worker nodes to spin up</span>
+<span class="sd"> :type num_preemptible_workers: int</span>
+<span class="sd"> :param labels: dict of labels to add to the cluster</span>
+<span class="sd"> :type labels: dict</span>
+<span class="sd"> :param zone: The zone where the cluster will be located</span>
+<span class="sd"> :type zone: string</span>
+<span class="sd"> :param network_uri: The network uri to be used for machine communication, cannot be</span>
+<span class="sd"> specified with subnetwork_uri</span>
+<span class="sd"> :type network_uri: string</span>
+<span class="sd"> :param subnetwork_uri: The subnetwork uri to be used for machine communication,</span>
+<span class="sd"> cannot be specified with network_uri</span>
+<span class="sd"> :type subnetwork_uri: string</span>
+<span class="sd"> :param tags: The GCE tags to add to all instances</span>
+<span class="sd"> :type tags: list[string]</span>
+<span class="sd"> :param region: leave as 'global', might become relevant in the future</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
+<span class="sd"> :param service_account: The service account of the dataproc instances.</span>
+<span class="sd"> :type service_account: string</span>
+<span class="sd"> :param service_account_scopes: The URIs of service account scopes to be included.</span>
+<span class="sd"> :type service_account_scopes: list[string]</span>
+<span class="sd"> :param idle_delete_ttl: The longest duration that cluster would keep alive while</span>
+<span class="sd"> staying idle. Passing this threshold will cause cluster to be auto-deleted.</span>
+<span class="sd"> A duration in seconds.</span>
+<span class="sd"> :type idle_delete_ttl: int</span>
+<span class="sd"> :param auto_delete_time: The time when cluster will be auto-deleted.</span>
+<span class="sd"> :type auto_delete_time: datetime</span>
+<span class="sd"> :param auto_delete_ttl: The life duration of cluster, the cluster will be</span>
+<span class="sd"> auto-deleted at the end of this duration.</span>
+<span class="sd"> A duration in seconds. (If auto_delete_time is set this parameter will be ignored)</span>
+<span class="sd"> :type auto_delete_ttl: int</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'cluster_name'</span><span class="p">,]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'project_id'</span><span class="p">,</span> <span class="s1">'zone'</span><span class="p">,</span> <span class="s1">'region'</span><span class="p">]</span>
<span class="nd">@apply_defaults</span>
<span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span>
@@ -212,8 +297,12 @@
<span class="n">project_id</span><span class="p">,</span>
<span class="n">num_workers</span><span class="p">,</span>
<span class="n">zone</span><span class="p">,</span>
+ <span class="n">network_uri</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">subnetwork_uri</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">tags</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">storage_bucket</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">init_actions_uris</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">init_action_timeout</span><span class="o">=</span><span class="s2">"10m"</span><span class="p">,</span>
<span class="n">metadata</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">image_version</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">properties</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
@@ -228,64 +317,12 @@
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">service_account</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">service_account_scopes</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">idle_delete_ttl</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">auto_delete_time</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">auto_delete_ttl</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Create a new DataprocClusterCreateOperator.</span>
-
-<span class="sd"> For more info on the creation of a cluster through the API, have a look at:</span>
-
-<span class="sd"> https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters</span>
-
-<span class="sd"> :param cluster_name: The name of the DataProc cluster to create.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param project_id: The ID of the google cloud project in which</span>
-<span class="sd"> to create the cluster</span>
-<span class="sd"> :type project_id: string</span>
-<span class="sd"> :param num_workers: The # of workers to spin up</span>
-<span class="sd"> :type num_workers: int</span>
-<span class="sd"> :param storage_bucket: The storage bucket to use, setting to None lets dataproc</span>
-<span class="sd"> generate a custom one for you</span>
-<span class="sd"> :type storage_bucket: string</span>
-<span class="sd"> :param init_actions_uris: List of GCS uri's containing</span>
-<span class="sd"> dataproc initialization scripts</span>
-<span class="sd"> :type init_actions_uris: list[string]</span>
-<span class="sd"> :param metadata: dict of key-value google compute engine metadata entries</span>
-<span class="sd"> to add to all instances</span>
-<span class="sd"> :type metadata: dict</span>
-<span class="sd"> :param image_version: the version of software inside the Dataproc cluster</span>
-<span class="sd"> :type image_version: string</span>
-<span class="sd"> :param properties: dict of properties to set on</span>
-<span class="sd"> config files (e.g. spark-defaults.conf), see</span>
-<span class="sd"> https://cloud.google.com/dataproc/docs/reference/rest/v1/ \</span>
-<span class="sd"> projects.regions.clusters#SoftwareConfig</span>
-<span class="sd"> :type properties: dict</span>
-<span class="sd"> :param master_machine_type: Compute engine machine type to use for the master node</span>
-<span class="sd"> :type master_machine_type: string</span>
-<span class="sd"> :param master_disk_size: Disk size for the master node</span>
-<span class="sd"> :type int</span>
-<span class="sd"> :param worker_machine_type:Compute engine machine type to use for the worker nodes</span>
-<span class="sd"> :type worker_machine_type: string</span>
-<span class="sd"> :param worker_disk_size: Disk size for the worker nodes</span>
-<span class="sd"> :type worker_disk_size: int</span>
-<span class="sd"> :param num_preemptible_workers: The # of preemptible worker nodes to spin up</span>
-<span class="sd"> :type num_preemptible_workers: int</span>
-<span class="sd"> :param labels: dict of labels to add to the cluster</span>
-<span class="sd"> :type labels: dict</span>
-<span class="sd"> :param zone: The zone where the cluster will be located</span>
-<span class="sd"> :type zone: string</span>
-<span class="sd"> :param region: leave as 'global', might become relevant in the future</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> :param service_account: The service account of the dataproc instances.</span>
-<span class="sd"> :type service_account: string</span>
-<span class="sd"> :param service_account_scopes: The URIs of service account scopes to be included.</span>
-<span class="sd"> :type service_account_scopes: list[string]</span>
-<span class="sd"> """</span>
+
<span class="nb">super</span><span class="p">(</span><span class="n">DataprocClusterCreateOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -295,6 +332,7 @@
<span class="bp">self</span><span class="o">.</span><span class="n">num_preemptible_workers</span> <span class="o">=</span> <span class="n">num_preemptible_workers</span>
<span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span> <span class="o">=</span> <span class="n">storage_bucket</span>
<span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span> <span class="o">=</span> <span class="n">init_actions_uris</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span> <span class="o">=</span> <span class="n">init_action_timeout</span>
<span class="bp">self</span><span class="o">.</span><span class="n">metadata</span> <span class="o">=</span> <span class="n">metadata</span>
<span class="bp">self</span><span class="o">.</span><span class="n">image_version</span> <span class="o">=</span> <span class="n">image_version</span>
<span class="bp">self</span><span class="o">.</span><span class="n">properties</span> <span class="o">=</span> <span class="n">properties</span>
@@ -304,9 +342,15 @@
<span class="bp">self</span><span class="o">.</span><span class="n">worker_disk_size</span> <span class="o">=</span> <span class="n">worker_disk_size</span>
<span class="bp">self</span><span class="o">.</span><span class="n">labels</span> <span class="o">=</span> <span class="n">labels</span>
<span class="bp">self</span><span class="o">.</span><span class="n">zone</span> <span class="o">=</span> <span class="n">zone</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span> <span class="o">=</span> <span class="n">network_uri</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span> <span class="o">=</span> <span class="n">subnetwork_uri</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">tags</span> <span class="o">=</span> <span class="n">tags</span>
<span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
<span class="bp">self</span><span class="o">.</span><span class="n">service_account</span> <span class="o">=</span> <span class="n">service_account</span>
<span class="bp">self</span><span class="o">.</span><span class="n">service_account_scopes</span> <span class="o">=</span> <span class="n">service_account_scopes</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span> <span class="o">=</span> <span class="n">idle_delete_ttl</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span> <span class="o">=</span> <span class="n">auto_delete_time</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span> <span class="o">=</span> <span class="n">auto_delete_ttl</span>
<span class="k">def</span> <span class="nf">_get_cluster_list_for_project</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">service</span><span class="p">):</span>
<span class="n">result</span> <span class="o">=</span> <span class="n">service</span><span class="o">.</span><span class="n">projects</span><span class="p">()</span><span class="o">.</span><span class="n">regions</span><span class="p">()</span><span class="o">.</span><span class="n">clusters</span><span class="p">()</span><span class="o">.</span><span class="n">list</span><span class="p">(</span>
@@ -357,6 +401,19 @@
<span class="k">return</span>
<span class="n">time</span><span class="o">.</span><span class="n">sleep</span><span class="p">(</span><span class="mi">15</span><span class="p">)</span>
+ <span class="k">def</span> <span class="nf">_get_init_action_timeout</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
+ <span class="n">match</span> <span class="o">=</span> <span class="n">re</span><span class="o">.</span><span class="n">match</span><span class="p">(</span><span class="sa">r</span><span class="s2">"^(\d+)(s|m)$"</span><span class="p">,</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span><span class="p">)</span>
+ <span class="k">if</span> <span class="n">match</span><span class="p">:</span>
+ <span class="k">if</span> <span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">2</span><span class="p">)</span> <span class="o">==</span> <span class="s2">"s"</span><span class="p">:</span>
+ <span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span>
+ <span class="k">elif</span> <span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">2</span><span class="p">)</span> <span class="o">==</span> <span class="s2">"m"</span><span class="p">:</span>
+ <span class="n">val</span> <span class="o">=</span> <span class="nb">float</span><span class="p">(</span><span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">1</span><span class="p">))</span>
+ <span class="k">return</span> <span class="s2">"</span><span class="si">{}</span><span class="s2">s"</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="n">timedelta</span><span class="p">(</span><span class="n">minutes</span><span class="o">=</span><span class="n">val</span><span class="p">)</span><span class="o">.</span><span class="n">seconds</span><span class="p">)</span>
+
+ <span class="k">raise</span> <span class="n">AirflowException</span><span class="p">(</span>
+ <span class="s2">"DataprocClusterCreateOperator init_action_timeout"</span>
+ <span class="s2">" should be expressed in minutes or seconds. i.e. 10m, 30s"</span><span class="p">)</span>
+
<span class="k">def</span> <span class="nf">_build_cluster_data</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
<span class="n">zone_uri</span> <span class="o">=</span> \
<span class="s1">'https://www.googleapis.com/compute/v1/projects/</span><span class="si">{}</span><span class="s1">/zones/</span><span class="si">{}</span><span class="s1">'</span><span class="o">.</span><span class="n">format</span><span class="p">(</span>
@@ -392,7 +449,8 @@
<span class="p">}</span>
<span class="p">},</span>
<span class="s1">'secondaryWorkerConfig'</span><span class="p">:</span> <span class="p">{},</span>
- <span class="s1">'softwareConfig'</span><span class="p">:</span> <span class="p">{}</span>
+ <span class="s1">'softwareConfig'</span><span class="p">:</span> <span class="p">{},</span>
+ <span class="s1">'lifecycleConfig'</span><span class="p">:</span> <span class="p">{}</span>
<span class="p">}</span>
<span class="p">}</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">num_preemptible_workers</span> <span class="o">></span> <span class="mi">0</span><span class="p">:</span>
@@ -410,18 +468,37 @@
<span class="c1"># [a-z]([-a-z0-9]*[a-z0-9])? (current airflow version string follows</span>
<span class="c1"># semantic versioning spec: x.y.z).</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'labels'</span><span class="p">]</span><span class="o">.</span><span class="n">update</span><span class="p">({</span><span class="s1">'airflow-version'</span><span class="p">:</span>
- <span class="s1">'v'</span> <span class="o">+</span> <span class="n">version</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">'.'</span><span class="p">,</span> <span class="s1">'-'</span><span class="p">)})</span>
+ <span class="s1">'v'</span> <span class="o">+</span> <span class="n">version</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">'.'</span><span class="p">,</span> <span class="s1">'-'</span><span class="p">)</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">'+'</span><span class="p">,</span><span class="s1">'-'</span><span class="p">)})</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span><span class="p">:</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'configBucket'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">metadata</span><span class="p">:</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'gceClusterConfig'</span><span class="p">][</span><span class="s1">'metadata'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">metadata</span>
+ <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span><span class="p">:</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'gceClusterConfig'</span><span class="p">][</span><span class="s1">'networkUri'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span>
+ <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span><span class="p">:</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'gceClusterConfig'</span><span class="p">][</span><span class="s1">'subnetworkUri'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span>
+ <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">tags</span><span class="p">:</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'gceClusterConfig'</span><span class="p">][</span><span class="s1">'tags'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">tags</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">image_version</span><span class="p">:</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'softwareConfig'</span><span class="p">][</span><span class="s1">'imageVersion'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">image_version</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">properties</span><span class="p">:</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'softwareConfig'</span><span class="p">][</span><span class="s1">'properties'</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">properties</span>
+ <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span><span class="p">:</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'lifecycleConfig'</span><span class="p">][</span><span class="s1">'idleDeleteTtl'</span><span class="p">]</span> <span class="o">=</span> \
+ <span class="s2">"</span><span class="si">{}</span><span class="s2">s"</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span><span class="p">)</span>
+ <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span><span class="p">:</span>
+ <span class="n">utc_auto_delete_time</span> <span class="o">=</span> <span class="n">timezone</span><span class="o">.</span><span class="n">convert_to_utc</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span><span class="p">)</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'lifecycleConfig'</span><span class="p">][</span><span class="s1">'autoDeleteTime'</span><span class="p">]</span> <span class="o">=</span> \
+ <span class="n">utc_auto_delete_time</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="s1">'%Y-%m-</span><span class="si">%d</span><span class="s1">T%H:%M:%S.</span><span class="si">%f</span><span class="s1">Z'</span><span class="p">,</span> <span class="n">formatter</span><span class="o">=</span><span class="s1">'classic'</span><span class="p">)</span>
+ <span class="k">elif</span> <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span><span class="p">:</span>
+ <span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'lifecycleConfig'</span><span class="p">][</span><span class="s1">'autoDeleteTtl'</span><span class="p">]</span> <span class="o">=</span> \
+ <span class="s2">"</span><span class="si">{}</span><span class="s2">s"</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span><span class="p">)</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span><span class="p">:</span>
<span class="n">init_actions_dict</span> <span class="o">=</span> <span class="p">[</span>
- <span class="p">{</span><span class="s1">'executableFile'</span><span class="p">:</span> <span class="n">uri</span><span class="p">}</span> <span class="k">for</span> <span class="n">uri</span> <span class="ow">in</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span>
+ <span class="p">{</span>
+ <span class="s1">'executableFile'</span><span class="p">:</span> <span class="n">uri</span><span class="p">,</span>
+ <span class="s1">'executionTimeout'</span><span class="p">:</span> <span class="bp">self</span><span class="o">.</span><span class="n">_get_init_action_timeout</span><span class="p">()</span>
+ <span class="p">}</span> <span class="k">for</span> <span class="n">uri</span> <span class="ow">in</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span>
<span class="p">]</span>
<span class="n">cluster_data</span><span class="p">[</span><span class="s1">'config'</span><span class="p">][</span><span class="s1">'initializationActions'</span><span class="p">]</span> <span class="o">=</span> <span class="n">init_actions_dict</span>
<span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">service_account</span><span class="p">:</span>
@@ -468,16 +545,30 @@
<span class="k">else</span><span class="p">:</span>
<span class="k">raise</span> <span class="n">e</span>
- <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">)</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">)</span></div>
-<span class="k">class</span> <span class="nc">DataprocClusterDeleteOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
+<div class="viewcode-block" id="DataprocClusterDeleteOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataprocClusterDeleteOperator">[docs]</a><span class="k">class</span> <span class="nc">DataprocClusterDeleteOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Delete a cluster on Google Cloud Dataproc. The operator will wait until the</span>
<span class="sd"> cluster is destroyed.</span>
+
+<span class="sd"> :param cluster_name: The name of the cluster to create.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param project_id: The ID of the google cloud project in which</span>
+<span class="sd"> the cluster runs</span>
+<span class="sd"> :type project_id: string</span>
+<span class="sd"> :param region: leave as 'global', might become relevant in the future</span>
+<span class="sd"> :type region: string</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'cluster_name'</span><span class="p">]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'project_id'</span><span class="p">,</span> <span class="s1">'region'</span><span class="p">]</span>
<span class="nd">@apply_defaults</span>
<span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span>
@@ -488,23 +579,7 @@
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Delete a cluster on Google Cloud Dataproc.</span>
-
-<span class="sd"> :param cluster_name: The name of the cluster to create.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param project_id: The ID of the google cloud project in which</span>
-<span class="sd"> the cluster runs</span>
-<span class="sd"> :type project_id: string</span>
-<span class="sd"> :param region: leave as 'global', might become relevant in the future</span>
-<span class="sd"> :type region: string</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> """</span>
+
<span class="nb">super</span><span class="p">(</span><span class="n">DataprocClusterDeleteOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -541,7 +616,7 @@
<span class="p">)</span><span class="o">.</span><span class="n">execute</span><span class="p">()</span>
<span class="n">operation_name</span> <span class="o">=</span> <span class="n">response</span><span class="p">[</span><span class="s1">'name'</span><span class="p">]</span>
<span class="bp">self</span><span class="o">.</span><span class="n">log</span><span class="o">.</span><span class="n">info</span><span class="p">(</span><span class="s2">"Cluster delete operation name: </span><span class="si">%s</span><span class="s2">"</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span>
- <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span></div>
<div class="viewcode-block" id="DataProcPigOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcPigOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcPigOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
@@ -552,29 +627,60 @@
<span class="sd"> It's a good practice to define dataproc_* parameters in the default_args of the dag</span>
<span class="sd"> like the cluster name and UDFs.</span>
-<span class="sd"> ```</span>
-<span class="sd"> default_args = {</span>
-<span class="sd"> 'cluster_name': 'cluster-1',</span>
-<span class="sd"> 'dataproc_pig_jars': [</span>
-<span class="sd"> 'gs://example/udf/jar/datafu/1.2.0/datafu.jar',</span>
-<span class="sd"> 'gs://example/udf/jar/gpig/1.2/gpig.jar'</span>
-<span class="sd"> ]</span>
-<span class="sd"> }</span>
-<span class="sd"> ```</span>
+<span class="sd"> .. code-block:: python</span>
+
+<span class="sd"> default_args = {</span>
+<span class="sd"> 'cluster_name': 'cluster-1',</span>
+<span class="sd"> 'dataproc_pig_jars': [</span>
+<span class="sd"> 'gs://example/udf/jar/datafu/1.2.0/datafu.jar',</span>
+<span class="sd"> 'gs://example/udf/jar/gpig/1.2/gpig.jar'</span>
+<span class="sd"> ]</span>
+<span class="sd"> }</span>
<span class="sd"> You can pass a pig script as string or file reference. Use variables to pass on</span>
<span class="sd"> variables for the pig script to be resolved on the cluster or use the parameters to</span>
<span class="sd"> be resolved in the script as template parameters.</span>
-<span class="sd"> ```</span>
-<span class="sd"> t1 = DataProcPigOperator(</span>
-<span class="sd"> task_id='dataproc_pig',</span>
-<span class="sd"> query='a_pig_script.pig',</span>
-<span class="sd"> variables={'out': 'gs://example/output/{{ds}}'},</span>
-<span class="sd"> dag=dag)</span>
-<span class="sd"> ```</span>
+<span class="sd"> **Example**: ::</span>
+
+<span class="sd"> t1 = DataProcPigOperator(</span>
+<span class="sd"> task_id='dataproc_pig',</span>
+<span class="sd"> query='a_pig_script.pig',</span>
+<span class="sd"> variables={'out': 'gs://example/output/{{ds}}'},</span>
+<span class="sd"> dag=dag)</span>
+
+<span class="sd"> .. seealso::</span>
+<span class="sd"> For more detail on about job submission have a look at the reference:</span>
+<span class="sd"> https://cloud.google.com/dataproc/reference/rest/v1/projects.regions.jobs</span>
+
+<span class="sd"> :param query: The query or reference to the query file (pg or pig extension).</span>
+<span class="sd"> :type query: string</span>
+<span class="sd"> :param query_uri: The uri of a pig script on Cloud Storage.</span>
+<span class="sd"> :type query_uri: string</span>
+<span class="sd"> :param variables: Map of named parameters for the query.</span>
+<span class="sd"> :type variables: dict</span>
+<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd"> :type job_name: string</span>
+<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param dataproc_pig_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd"> default arguments</span>
+<span class="sd"> :type dataproc_pig_properties: dict</span>
+<span class="sd"> :param dataproc_pig_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
+<span class="sd"> UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd"> :type dataproc_pig_jars: list</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
+<span class="sd"> :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd"> :type region: string</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'dataproc_jars'</span><span class="p">]</span>
<span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">'.pg'</span><span class="p">,</span> <span class="s1">'.pig'</span><span class="p">,)</span>
<span class="n">ui_color</span> <span class="o">=</span> <span class="s1">'#0273d4'</span>
@@ -590,40 +696,10 @@
<span class="n">dataproc_pig_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">'google_cloud_default'</span><span class="p">,</span>
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">region</span><span class="o">=</span><span class="s1">'global'</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Create a new DataProcPigOperator.</span>
-
-<span class="sd"> For more detail on about job submission have a look at the reference:</span>
-
-<span class="sd"> https://cloud.google.com/dataproc/reference/rest/v1/projects.regions.jobs</span>
-<span class="sd"> :param query: The query or reference to the query file (pg or pig extension).</span>
-<span class="sd"> :type query: string</span>
-<span class="sd"> :param query_uri: The uri of a pig script on Cloud Storage.</span>
-<span class="sd"> :type query_uri: string</span>
-<span class="sd"> :param variables: Map of named parameters for the query.</span>
-<span class="sd"> :type variables: dict</span>
-<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd"> :type job_name: string</span>
-<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param dataproc_pig_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd"> default arguments</span>
-<span class="sd"> :type dataproc_pig_properties: dict</span>
-<span class="sd"> :param dataproc_pig_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
-<span class="sd"> UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd"> :type dataproc_pig_jars: list</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> """</span>
<span class="nb">super</span><span class="p">(</span><span class="n">DataProcPigOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -634,6 +710,7 @@
<span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_pig_properties</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_pig_jars</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
<span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
<span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -649,14 +726,41 @@
<span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
<span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
- <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+ <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
<div class="viewcode-block" id="DataProcHiveOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcHiveOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcHiveOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Start a Hive query Job on a Cloud DataProc cluster.</span>
+
+<span class="sd"> :param query: The query or reference to the query file (q extension).</span>
+<span class="sd"> :type query: string</span>
+<span class="sd"> :param query_uri: The uri of a hive script on Cloud Storage.</span>
+<span class="sd"> :type query_uri: string</span>
+<span class="sd"> :param variables: Map of named parameters for the query.</span>
+<span class="sd"> :type variables: dict</span>
+<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd"> :type job_name: string</span>
+<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param dataproc_hive_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd"> default arguments</span>
+<span class="sd"> :type dataproc_hive_properties: dict</span>
+<span class="sd"> :param dataproc_hive_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
+<span class="sd"> UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd"> :type dataproc_hive_jars: list</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
+<span class="sd"> :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd"> :type region: string</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'dataproc_jars'</span><span class="p">]</span>
<span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">'.q'</span><span class="p">,)</span>
<span class="n">ui_color</span> <span class="o">=</span> <span class="s1">'#0273d4'</span>
@@ -672,36 +776,10 @@
<span class="n">dataproc_hive_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">'google_cloud_default'</span><span class="p">,</span>
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">region</span><span class="o">=</span><span class="s1">'global'</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Create a new DataProcHiveOperator.</span>
-
-<span class="sd"> :param query: The query or reference to the query file (q extension).</span>
-<span class="sd"> :type query: string</span>
-<span class="sd"> :param query_uri: The uri of a hive script on Cloud Storage.</span>
-<span class="sd"> :type query_uri: string</span>
-<span class="sd"> :param variables: Map of named parameters for the query.</span>
-<span class="sd"> :type variables: dict</span>
-<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd"> :type job_name: string</span>
-<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param dataproc_hive_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd"> default arguments</span>
-<span class="sd"> :type dataproc_hive_properties: dict</span>
-<span class="sd"> :param dataproc_hive_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
-<span class="sd"> UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd"> :type dataproc_hive_jars: list</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> """</span>
+
<span class="nb">super</span><span class="p">(</span><span class="n">DataProcHiveOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -712,6 +790,7 @@
<span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_hive_properties</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_hive_jars</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
<span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
<span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -728,14 +807,41 @@
<span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
<span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
- <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+ <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
<div class="viewcode-block" id="DataProcSparkSqlOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcSparkSqlOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcSparkSqlOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Start a Spark SQL query Job on a Cloud DataProc cluster.</span>
+
+<span class="sd"> :param query: The query or reference to the query file (q extension).</span>
+<span class="sd"> :type query: string</span>
+<span class="sd"> :param query_uri: The uri of a spark sql script on Cloud Storage.</span>
+<span class="sd"> :type query_uri: string</span>
+<span class="sd"> :param variables: Map of named parameters for the query.</span>
+<span class="sd"> :type variables: dict</span>
+<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd"> :type job_name: string</span>
+<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd"> default arguments</span>
+<span class="sd"> :type dataproc_spark_properties: dict</span>
+<span class="sd"> :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd"> for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd"> :type dataproc_spark_jars: list</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
+<span class="sd"> :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd"> :type region: string</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'query'</span><span class="p">,</span> <span class="s1">'variables'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'dataproc_jars'</span><span class="p">]</span>
<span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">'.q'</span><span class="p">,)</span>
<span class="n">ui_color</span> <span class="o">=</span> <span class="s1">'#0273d4'</span>
@@ -751,36 +857,10 @@
<span class="n">dataproc_spark_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">'google_cloud_default'</span><span class="p">,</span>
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">region</span><span class="o">=</span><span class="s1">'global'</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Create a new DataProcSparkSqlOperator.</span>
-
-<span class="sd"> :param query: The query or reference to the query file (q extension).</span>
-<span class="sd"> :type query: string</span>
-<span class="sd"> :param query_uri: The uri of a spark sql script on Cloud Storage.</span>
-<span class="sd"> :type query_uri: string</span>
-<span class="sd"> :param variables: Map of named parameters for the query.</span>
-<span class="sd"> :type variables: dict</span>
-<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd"> :type job_name: string</span>
-<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd"> default arguments</span>
-<span class="sd"> :type dataproc_spark_properties: dict</span>
-<span class="sd"> :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
-<span class="sd"> for UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd"> :type dataproc_spark_jars: list</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> """</span>
+
<span class="nb">super</span><span class="p">(</span><span class="n">DataProcSparkSqlOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -791,6 +871,7 @@
<span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_spark_properties</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_spark_jars</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
<span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
<span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -807,15 +888,49 @@
<span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
<span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
- <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+ <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
<div class="viewcode-block" id="DataProcSparkOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcSparkOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcSparkOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Start a Spark Job on a Cloud DataProc cluster.</span>
+
+<span class="sd"> :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
+<span class="sd"> the main_class, not both together).</span>
+<span class="sd"> :type main_jar: string</span>
+<span class="sd"> :param main_class: Name of the job class. (use this or the main_jar, not both</span>
+<span class="sd"> together).</span>
+<span class="sd"> :type main_class: string</span>
+<span class="sd"> :param arguments: Arguments for the job.</span>
+<span class="sd"> :type arguments: list</span>
+<span class="sd"> :param archives: List of archived files that will be unpacked in the work</span>
+<span class="sd"> directory. Should be stored in Cloud Storage.</span>
+<span class="sd"> :type archives: list</span>
+<span class="sd"> :param files: List of files to be copied to the working directory</span>
+<span class="sd"> :type files: list</span>
+<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd"> :type job_name: string</span>
+<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd"> default arguments</span>
+<span class="sd"> :type dataproc_spark_properties: dict</span>
+<span class="sd"> :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd"> for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd"> :type dataproc_spark_jars: list</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</span>
+<span class="sd"> :type delegate_to: string</span>
+<span class="sd"> :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd"> :type region: string</span>
<span class="sd"> """</span>
- <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'arguments'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">]</span>
+ <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">'arguments'</span><span class="p">,</span> <span class="s1">'job_name'</span><span class="p">,</span> <span class="s1">'cluster_name'</span><span class="p">,</span> <span class="s1">'dataproc_jars'</span><span class="p">]</span>
<span class="n">ui_color</span> <span class="o">=</span> <span class="s1">'#0273d4'</span>
<span class="nd">@apply_defaults</span>
@@ -832,43 +947,10 @@
<span class="n">dataproc_spark_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
<span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">'google_cloud_default'</span><span class="p">,</span>
<span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+ <span class="n">region</span><span class="o">=</span><span class="s1">'global'</span><span class="p">,</span>
<span class="o">*</span><span class="n">args</span><span class="p">,</span>
<span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
- <span class="sd">"""</span>
-<span class="sd"> Create a new DataProcSparkOperator.</span>
-<span class="sd"> :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
-<span class="sd"> the main_class, not both together).</span>
-<span class="sd"> :type main_jar: string</span>
-<span class="sd"> :param main_class: Name of the job class. (use this or the main_jar, not both</span>
-<span class="sd"> together).</span>
-<span class="sd"> :type main_class: string</span>
-<span class="sd"> :param arguments: Arguments for the job.</span>
-<span class="sd"> :type arguments: list</span>
-<span class="sd"> :param archives: List of archived files that will be unpacked in the work</span>
-<span class="sd"> directory. Should be stored in Cloud Storage.</span>
-<span class="sd"> :type archives: list</span>
-<span class="sd"> :param files: List of files to be copied to the working directory</span>
-<span class="sd"> :type files: list</span>
-<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd"> :type job_name: string</span>
-<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd"> :type cluster_name: string</span>
-<span class="sd"> :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd"> default arguments</span>
-<span class="sd"> :type dataproc_spark_properties: dict</span>
-<span class="sd"> :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
-<span class="sd"> for UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd"> :type dataproc_spark_jars: list</span>
-<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd"> :type gcp_conn_id: string</span>
-<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
-<span class="sd"> delegation enabled.</span>
-<span class="sd"> :type delegate_to: string</span>
-<span class="sd"> """</span>
<span class="nb">super</span><span class="p">(</span><span class="n">DataProcSparkOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
<span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
<span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -881,6 +963,7 @@
<span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_spark_properties</span>
<span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_spark_jars</span>
+ <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
<span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
<span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -895,15 +978,49 @@
<span class="n">job</span><span class="o">.</span><span class="n">add_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">files</span><span class="p">)</span>
<span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
- <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+ <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
<div class="viewcode-block" id="DataProcHadoopOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcHadoopOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcHadoopOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
<span class="sd">"""</span>
<span class="sd"> Start a Hadoop Job on a Cloud DataProc cluster.</span>
+
+<span class="sd"> :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
+<span class="sd"> the main_class, not both together).</span>
+<span class="sd"> :type main_jar: string</span>
+<span class="sd"> :param main_class: Name of the job class. (use this or the main_jar, not both</span>
+<span class="sd"> together).</span>
+<span class="sd"> :type main_class: string</span>
+<span class="sd"> :param arguments: Arguments for the job.</span>
+<span class="sd"> :type arguments: list</span>
+<span class="sd"> :param archives: List of archived files that will be unpacked in the work</span>
+<span class="sd"> directory. Should be stored in Cloud Storage.</span>
+<span class="sd"> :type archives: list</span>
+<span class="sd"> :param files: List of files to be copied to the working directory</span>
+<span class="sd"> :type files: list</span>
+<span class="sd"> :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd"> is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd"> name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd"> :type job_name: string</span>
+<span class="sd"> :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd"> :type cluster_name: string</span>
+<span class="sd"> :param dataproc_hadoop_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd"> default arguments</span>
+<span class="sd"> :type dataproc_hadoop_properties: dict</span>
+<span class="sd"> :param dataproc_hadoop_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd"> for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd"> :type dataproc_hadoop_jars: list</span>
+<span class="sd"> :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd"> :type gcp_conn_id: string</span>
+<span class="sd"> :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd"> For this to work, the service account making the request must have domain-wide</span>
+<span class="sd"> delegation enabled.</s
<TRUNCATED>