You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by ka...@apache.org on 2018/08/27 16:26:53 UTC

[33/51] [partial] incubator-airflow-site git commit: 1.10.0

http://git-wip-us.apache.org/repos/asf/incubator-airflow-site/blob/11437c14/_modules/airflow/contrib/operators/dataproc_operator.html
----------------------------------------------------------------------
diff --git a/_modules/airflow/contrib/operators/dataproc_operator.html b/_modules/airflow/contrib/operators/dataproc_operator.html
index d49719e..f26e459 100644
--- a/_modules/airflow/contrib/operators/dataproc_operator.html
+++ b/_modules/airflow/contrib/operators/dataproc_operator.html
@@ -91,7 +91,7 @@
 <li class="toctree-l1"><a class="reference internal" href="../../../../start.html">Quick Start</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../installation.html">Installation</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../tutorial.html">Tutorial</a></li>
-<li class="toctree-l1"><a class="reference internal" href="../../../../configuration.html">Configuration</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../howto/index.html">How-to Guides</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../ui.html">UI / Screenshots</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../concepts.html">Concepts</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../profiling.html">Data Profiling</a></li>
@@ -99,8 +99,10 @@
 <li class="toctree-l1"><a class="reference internal" href="../../../../scheduler.html">Scheduling &amp; Triggers</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../plugins.html">Plugins</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../security.html">Security</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../timezone.html">Time zones</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../api.html">Experimental Rest API</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../integration.html">Integration</a></li>
+<li class="toctree-l1"><a class="reference internal" href="../../../../lineage.html">Lineage</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../faq.html">FAQ</a></li>
 <li class="toctree-l1"><a class="reference internal" href="../../../../code.html">API Reference</a></li>
 </ul>
@@ -169,29 +171,42 @@
   <h1>Source code for airflow.contrib.operators.dataproc_operator</h1><div class="highlight"><pre>
 <span></span><span class="c1"># -*- coding: utf-8 -*-</span>
 <span class="c1">#</span>
-<span class="c1"># Licensed under the Apache License, Version 2.0 (the &quot;License&quot;);</span>
-<span class="c1"># you may not use this file except in compliance with the License.</span>
-<span class="c1"># You may obtain a copy of the License at</span>
+<span class="c1"># Licensed to the Apache Software Foundation (ASF) under one</span>
+<span class="c1"># or more contributor license agreements.  See the NOTICE file</span>
+<span class="c1"># distributed with this work for additional information</span>
+<span class="c1"># regarding copyright ownership.  The ASF licenses this file</span>
+<span class="c1"># to you under the Apache License, Version 2.0 (the</span>
+<span class="c1"># &quot;License&quot;); you may not use this file except in compliance</span>
+<span class="c1"># with the License.  You may obtain a copy of the License at</span>
 <span class="c1">#</span>
-<span class="c1"># http://www.apache.org/licenses/LICENSE-2.0</span>
+<span class="c1">#   http://www.apache.org/licenses/LICENSE-2.0</span>
 <span class="c1">#</span>
-<span class="c1"># Unless required by applicable law or agreed to in writing, software</span>
-<span class="c1"># distributed under the License is distributed on an &quot;AS IS&quot; BASIS,</span>
-<span class="c1"># WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.</span>
-<span class="c1"># See the License for the specific language governing permissions and</span>
-<span class="c1"># limitations under the License.</span>
+<span class="c1"># Unless required by applicable law or agreed to in writing,</span>
+<span class="c1"># software distributed under the License is distributed on an</span>
+<span class="c1"># &quot;AS IS&quot; BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY</span>
+<span class="c1"># KIND, either express or implied.  See the License for the</span>
+<span class="c1"># specific language governing permissions and limitations</span>
+<span class="c1"># under the License.</span>
 <span class="c1">#</span>
 
+<span class="kn">import</span> <span class="nn">ntpath</span>
+<span class="kn">import</span> <span class="nn">os</span>
+<span class="kn">import</span> <span class="nn">re</span>
 <span class="kn">import</span> <span class="nn">time</span>
+<span class="kn">import</span> <span class="nn">uuid</span>
+<span class="kn">from</span> <span class="nn">datetime</span> <span class="k">import</span> <span class="n">timedelta</span>
 
 <span class="kn">from</span> <span class="nn">airflow.contrib.hooks.gcp_dataproc_hook</span> <span class="k">import</span> <span class="n">DataProcHook</span>
+<span class="kn">from</span> <span class="nn">airflow.contrib.hooks.gcs_hook</span> <span class="k">import</span> <span class="n">GoogleCloudStorageHook</span>
+<span class="kn">from</span> <span class="nn">airflow.exceptions</span> <span class="k">import</span> <span class="n">AirflowException</span>
 <span class="kn">from</span> <span class="nn">airflow.models</span> <span class="k">import</span> <span class="n">BaseOperator</span>
 <span class="kn">from</span> <span class="nn">airflow.utils.decorators</span> <span class="k">import</span> <span class="n">apply_defaults</span>
 <span class="kn">from</span> <span class="nn">airflow.version</span> <span class="k">import</span> <span class="n">version</span>
 <span class="kn">from</span> <span class="nn">googleapiclient.errors</span> <span class="k">import</span> <span class="n">HttpError</span>
+<span class="kn">from</span> <span class="nn">airflow.utils</span> <span class="k">import</span> <span class="n">timezone</span>
 
 
-<span class="k">class</span> <span class="nc">DataprocClusterCreateOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
+<div class="viewcode-block" id="DataprocClusterCreateOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataprocClusterCreateOperator">[docs]</a><span class="k">class</span> <span class="nc">DataprocClusterCreateOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Create a new cluster on Google Cloud Dataproc. The operator will wait until the</span>
 <span class="sd">    creation is successful or an error occurs in the creation process.</span>
@@ -202,9 +217,79 @@
 
 <span class="sd">    for a detailed explanation on the different parameters. Most of the configuration</span>
 <span class="sd">    parameters detailed in the link are available as a parameter to this operator.</span>
+
+<span class="sd">    :param cluster_name: The name of the DataProc cluster to create.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param project_id: The ID of the google cloud project in which</span>
+<span class="sd">        to create the cluster</span>
+<span class="sd">    :type project_id: string</span>
+<span class="sd">    :param num_workers: The # of workers to spin up</span>
+<span class="sd">    :type num_workers: int</span>
+<span class="sd">    :param storage_bucket: The storage bucket to use, setting to None lets dataproc</span>
+<span class="sd">        generate a custom one for you</span>
+<span class="sd">    :type storage_bucket: string</span>
+<span class="sd">    :param init_actions_uris: List of GCS uri&#39;s containing</span>
+<span class="sd">        dataproc initialization scripts</span>
+<span class="sd">    :type init_actions_uris: list[string]</span>
+<span class="sd">    :param init_action_timeout: Amount of time executable scripts in</span>
+<span class="sd">        init_actions_uris has to complete</span>
+<span class="sd">    :type init_action_timeout: string</span>
+<span class="sd">    :param metadata: dict of key-value google compute engine metadata entries</span>
+<span class="sd">        to add to all instances</span>
+<span class="sd">    :type metadata: dict</span>
+<span class="sd">    :param image_version: the version of software inside the Dataproc cluster</span>
+<span class="sd">    :type image_version: string</span>
+<span class="sd">    :param properties: dict of properties to set on</span>
+<span class="sd">        config files (e.g. spark-defaults.conf), see</span>
+<span class="sd">        https://cloud.google.com/dataproc/docs/reference/rest/v1/ \</span>
+<span class="sd">        projects.regions.clusters#SoftwareConfig</span>
+<span class="sd">    :type properties: dict</span>
+<span class="sd">    :param master_machine_type: Compute engine machine type to use for the master node</span>
+<span class="sd">    :type master_machine_type: string</span>
+<span class="sd">    :param master_disk_size: Disk size for the master node</span>
+<span class="sd">    :type master_disk_size: int</span>
+<span class="sd">    :param worker_machine_type: Compute engine machine type to use for the worker nodes</span>
+<span class="sd">    :type worker_machine_type: string</span>
+<span class="sd">    :param worker_disk_size: Disk size for the worker nodes</span>
+<span class="sd">    :type worker_disk_size: int</span>
+<span class="sd">    :param num_preemptible_workers: The # of preemptible worker nodes to spin up</span>
+<span class="sd">    :type num_preemptible_workers: int</span>
+<span class="sd">    :param labels: dict of labels to add to the cluster</span>
+<span class="sd">    :type labels: dict</span>
+<span class="sd">    :param zone: The zone where the cluster will be located</span>
+<span class="sd">    :type zone: string</span>
+<span class="sd">    :param network_uri: The network uri to be used for machine communication, cannot be</span>
+<span class="sd">        specified with subnetwork_uri</span>
+<span class="sd">    :type network_uri: string</span>
+<span class="sd">    :param subnetwork_uri: The subnetwork uri to be used for machine communication,</span>
+<span class="sd">        cannot be specified with network_uri</span>
+<span class="sd">    :type subnetwork_uri: string</span>
+<span class="sd">    :param tags: The GCE tags to add to all instances</span>
+<span class="sd">    :type tags: list[string]</span>
+<span class="sd">    :param region: leave as &#39;global&#39;, might become relevant in the future</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
+<span class="sd">    :param service_account: The service account of the dataproc instances.</span>
+<span class="sd">    :type service_account: string</span>
+<span class="sd">    :param service_account_scopes: The URIs of service account scopes to be included.</span>
+<span class="sd">    :type service_account_scopes: list[string]</span>
+<span class="sd">    :param idle_delete_ttl: The longest duration that cluster would keep alive while</span>
+<span class="sd">        staying idle. Passing this threshold will cause cluster to be auto-deleted.</span>
+<span class="sd">        A duration in seconds.</span>
+<span class="sd">    :type idle_delete_ttl: int</span>
+<span class="sd">    :param auto_delete_time:  The time when cluster will be auto-deleted.</span>
+<span class="sd">    :type auto_delete_time: datetime</span>
+<span class="sd">    :param auto_delete_ttl: The life duration of cluster, the cluster will be</span>
+<span class="sd">        auto-deleted at the end of this duration.</span>
+<span class="sd">        A duration in seconds. (If auto_delete_time is set this parameter will be ignored)</span>
+<span class="sd">    :type auto_delete_ttl: int</span>
 <span class="sd">    &quot;&quot;&quot;</span>
 
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;cluster_name&#39;</span><span class="p">,]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;project_id&#39;</span><span class="p">,</span> <span class="s1">&#39;zone&#39;</span><span class="p">,</span> <span class="s1">&#39;region&#39;</span><span class="p">]</span>
 
     <span class="nd">@apply_defaults</span>
     <span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span>
@@ -212,8 +297,12 @@
                  <span class="n">project_id</span><span class="p">,</span>
                  <span class="n">num_workers</span><span class="p">,</span>
                  <span class="n">zone</span><span class="p">,</span>
+                 <span class="n">network_uri</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">subnetwork_uri</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">tags</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">storage_bucket</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">init_actions_uris</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">init_action_timeout</span><span class="o">=</span><span class="s2">&quot;10m&quot;</span><span class="p">,</span>
                  <span class="n">metadata</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">image_version</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">properties</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
@@ -228,64 +317,12 @@
                  <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">service_account</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="n">service_account_scopes</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">idle_delete_ttl</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">auto_delete_time</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+                 <span class="n">auto_delete_ttl</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="o">*</span><span class="n">args</span><span class="p">,</span>
                  <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Create a new DataprocClusterCreateOperator.</span>
-
-<span class="sd">        For more info on the creation of a cluster through the API, have a look at:</span>
-
-<span class="sd">        https://cloud.google.com/dataproc/docs/reference/rest/v1/projects.regions.clusters</span>
-
-<span class="sd">        :param cluster_name: The name of the DataProc cluster to create.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param project_id: The ID of the google cloud project in which</span>
-<span class="sd">            to create the cluster</span>
-<span class="sd">        :type project_id: string</span>
-<span class="sd">        :param num_workers: The # of workers to spin up</span>
-<span class="sd">        :type num_workers: int</span>
-<span class="sd">        :param storage_bucket: The storage bucket to use, setting to None lets dataproc</span>
-<span class="sd">            generate a custom one for you</span>
-<span class="sd">        :type storage_bucket: string</span>
-<span class="sd">        :param init_actions_uris: List of GCS uri&#39;s containing</span>
-<span class="sd">            dataproc initialization scripts</span>
-<span class="sd">        :type init_actions_uris: list[string]</span>
-<span class="sd">        :param metadata: dict of key-value google compute engine metadata entries</span>
-<span class="sd">            to add to all instances</span>
-<span class="sd">        :type metadata: dict</span>
-<span class="sd">        :param image_version: the version of software inside the Dataproc cluster</span>
-<span class="sd">        :type image_version: string</span>
-<span class="sd">        :param properties: dict of properties to set on</span>
-<span class="sd">            config files (e.g. spark-defaults.conf), see</span>
-<span class="sd">            https://cloud.google.com/dataproc/docs/reference/rest/v1/ \</span>
-<span class="sd">            projects.regions.clusters#SoftwareConfig</span>
-<span class="sd">        :type properties: dict</span>
-<span class="sd">        :param master_machine_type: Compute engine machine type to use for the master node</span>
-<span class="sd">        :type master_machine_type: string</span>
-<span class="sd">        :param master_disk_size: Disk size for the master node</span>
-<span class="sd">        :type int</span>
-<span class="sd">        :param worker_machine_type:Compute engine machine type to use for the worker nodes</span>
-<span class="sd">        :type worker_machine_type: string</span>
-<span class="sd">        :param worker_disk_size: Disk size for the worker nodes</span>
-<span class="sd">        :type worker_disk_size: int</span>
-<span class="sd">        :param num_preemptible_workers: The # of preemptible worker nodes to spin up</span>
-<span class="sd">        :type num_preemptible_workers: int</span>
-<span class="sd">        :param labels: dict of labels to add to the cluster</span>
-<span class="sd">        :type labels: dict</span>
-<span class="sd">        :param zone: The zone where the cluster will be located</span>
-<span class="sd">        :type zone: string</span>
-<span class="sd">        :param region: leave as &#39;global&#39;, might become relevant in the future</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        :param service_account: The service account of the dataproc instances.</span>
-<span class="sd">        :type service_account: string</span>
-<span class="sd">        :param service_account_scopes: The URIs of service account scopes to be included.</span>
-<span class="sd">        :type service_account_scopes: list[string]</span>
-<span class="sd">        &quot;&quot;&quot;</span>
+
         <span class="nb">super</span><span class="p">(</span><span class="n">DataprocClusterCreateOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -295,6 +332,7 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">num_preemptible_workers</span> <span class="o">=</span> <span class="n">num_preemptible_workers</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span> <span class="o">=</span> <span class="n">storage_bucket</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span> <span class="o">=</span> <span class="n">init_actions_uris</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span> <span class="o">=</span> <span class="n">init_action_timeout</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">metadata</span> <span class="o">=</span> <span class="n">metadata</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">image_version</span> <span class="o">=</span> <span class="n">image_version</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">properties</span> <span class="o">=</span> <span class="n">properties</span>
@@ -304,9 +342,15 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">worker_disk_size</span> <span class="o">=</span> <span class="n">worker_disk_size</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">labels</span> <span class="o">=</span> <span class="n">labels</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">zone</span> <span class="o">=</span> <span class="n">zone</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span> <span class="o">=</span> <span class="n">network_uri</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span> <span class="o">=</span> <span class="n">subnetwork_uri</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">tags</span> <span class="o">=</span> <span class="n">tags</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">service_account</span> <span class="o">=</span> <span class="n">service_account</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">service_account_scopes</span> <span class="o">=</span> <span class="n">service_account_scopes</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span> <span class="o">=</span> <span class="n">idle_delete_ttl</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span> <span class="o">=</span> <span class="n">auto_delete_time</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span> <span class="o">=</span> <span class="n">auto_delete_ttl</span>
 
     <span class="k">def</span> <span class="nf">_get_cluster_list_for_project</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">service</span><span class="p">):</span>
         <span class="n">result</span> <span class="o">=</span> <span class="n">service</span><span class="o">.</span><span class="n">projects</span><span class="p">()</span><span class="o">.</span><span class="n">regions</span><span class="p">()</span><span class="o">.</span><span class="n">clusters</span><span class="p">()</span><span class="o">.</span><span class="n">list</span><span class="p">(</span>
@@ -357,6 +401,19 @@
                     <span class="k">return</span>
                 <span class="n">time</span><span class="o">.</span><span class="n">sleep</span><span class="p">(</span><span class="mi">15</span><span class="p">)</span>
 
+    <span class="k">def</span> <span class="nf">_get_init_action_timeout</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
+        <span class="n">match</span> <span class="o">=</span> <span class="n">re</span><span class="o">.</span><span class="n">match</span><span class="p">(</span><span class="sa">r</span><span class="s2">&quot;^(\d+)(s|m)$&quot;</span><span class="p">,</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span><span class="p">)</span>
+        <span class="k">if</span> <span class="n">match</span><span class="p">:</span>
+            <span class="k">if</span> <span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">2</span><span class="p">)</span> <span class="o">==</span> <span class="s2">&quot;s&quot;</span><span class="p">:</span>
+                <span class="k">return</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_action_timeout</span>
+            <span class="k">elif</span> <span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">2</span><span class="p">)</span> <span class="o">==</span> <span class="s2">&quot;m&quot;</span><span class="p">:</span>
+                <span class="n">val</span> <span class="o">=</span> <span class="nb">float</span><span class="p">(</span><span class="n">match</span><span class="o">.</span><span class="n">group</span><span class="p">(</span><span class="mi">1</span><span class="p">))</span>
+                <span class="k">return</span> <span class="s2">&quot;</span><span class="si">{}</span><span class="s2">s&quot;</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="n">timedelta</span><span class="p">(</span><span class="n">minutes</span><span class="o">=</span><span class="n">val</span><span class="p">)</span><span class="o">.</span><span class="n">seconds</span><span class="p">)</span>
+
+        <span class="k">raise</span> <span class="n">AirflowException</span><span class="p">(</span>
+            <span class="s2">&quot;DataprocClusterCreateOperator init_action_timeout&quot;</span>
+            <span class="s2">&quot; should be expressed in minutes or seconds. i.e. 10m, 30s&quot;</span><span class="p">)</span>
+
     <span class="k">def</span> <span class="nf">_build_cluster_data</span><span class="p">(</span><span class="bp">self</span><span class="p">):</span>
         <span class="n">zone_uri</span> <span class="o">=</span> \
             <span class="s1">&#39;https://www.googleapis.com/compute/v1/projects/</span><span class="si">{}</span><span class="s1">/zones/</span><span class="si">{}</span><span class="s1">&#39;</span><span class="o">.</span><span class="n">format</span><span class="p">(</span>
@@ -392,7 +449,8 @@
                     <span class="p">}</span>
                 <span class="p">},</span>
                 <span class="s1">&#39;secondaryWorkerConfig&#39;</span><span class="p">:</span> <span class="p">{},</span>
-                <span class="s1">&#39;softwareConfig&#39;</span><span class="p">:</span> <span class="p">{}</span>
+                <span class="s1">&#39;softwareConfig&#39;</span><span class="p">:</span> <span class="p">{},</span>
+                <span class="s1">&#39;lifecycleConfig&#39;</span><span class="p">:</span> <span class="p">{}</span>
             <span class="p">}</span>
         <span class="p">}</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">num_preemptible_workers</span> <span class="o">&gt;</span> <span class="mi">0</span><span class="p">:</span>
@@ -410,18 +468,37 @@
         <span class="c1"># [a-z]([-a-z0-9]*[a-z0-9])? (current airflow version string follows</span>
         <span class="c1"># semantic versioning spec: x.y.z).</span>
         <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;labels&#39;</span><span class="p">]</span><span class="o">.</span><span class="n">update</span><span class="p">({</span><span class="s1">&#39;airflow-version&#39;</span><span class="p">:</span>
-                                       <span class="s1">&#39;v&#39;</span> <span class="o">+</span> <span class="n">version</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">&#39;.&#39;</span><span class="p">,</span> <span class="s1">&#39;-&#39;</span><span class="p">)})</span>
+                                       <span class="s1">&#39;v&#39;</span> <span class="o">+</span> <span class="n">version</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">&#39;.&#39;</span><span class="p">,</span> <span class="s1">&#39;-&#39;</span><span class="p">)</span><span class="o">.</span><span class="n">replace</span><span class="p">(</span><span class="s1">&#39;+&#39;</span><span class="p">,</span><span class="s1">&#39;-&#39;</span><span class="p">)})</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span><span class="p">:</span>
             <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;configBucket&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">storage_bucket</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">metadata</span><span class="p">:</span>
             <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;gceClusterConfig&#39;</span><span class="p">][</span><span class="s1">&#39;metadata&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">metadata</span>
+        <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span><span class="p">:</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;gceClusterConfig&#39;</span><span class="p">][</span><span class="s1">&#39;networkUri&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">network_uri</span>
+        <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span><span class="p">:</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;gceClusterConfig&#39;</span><span class="p">][</span><span class="s1">&#39;subnetworkUri&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">subnetwork_uri</span>
+        <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">tags</span><span class="p">:</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;gceClusterConfig&#39;</span><span class="p">][</span><span class="s1">&#39;tags&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">tags</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">image_version</span><span class="p">:</span>
             <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;softwareConfig&#39;</span><span class="p">][</span><span class="s1">&#39;imageVersion&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">image_version</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">properties</span><span class="p">:</span>
             <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;softwareConfig&#39;</span><span class="p">][</span><span class="s1">&#39;properties&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="bp">self</span><span class="o">.</span><span class="n">properties</span>
+        <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span><span class="p">:</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;lifecycleConfig&#39;</span><span class="p">][</span><span class="s1">&#39;idleDeleteTtl&#39;</span><span class="p">]</span> <span class="o">=</span> \
+                <span class="s2">&quot;</span><span class="si">{}</span><span class="s2">s&quot;</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">idle_delete_ttl</span><span class="p">)</span>
+        <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span><span class="p">:</span>
+            <span class="n">utc_auto_delete_time</span> <span class="o">=</span> <span class="n">timezone</span><span class="o">.</span><span class="n">convert_to_utc</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_time</span><span class="p">)</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;lifecycleConfig&#39;</span><span class="p">][</span><span class="s1">&#39;autoDeleteTime&#39;</span><span class="p">]</span> <span class="o">=</span> \
+                <span class="n">utc_auto_delete_time</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="s1">&#39;%Y-%m-</span><span class="si">%d</span><span class="s1">T%H:%M:%S.</span><span class="si">%f</span><span class="s1">Z&#39;</span><span class="p">,</span> <span class="n">formatter</span><span class="o">=</span><span class="s1">&#39;classic&#39;</span><span class="p">)</span>
+        <span class="k">elif</span> <span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span><span class="p">:</span>
+            <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;lifecycleConfig&#39;</span><span class="p">][</span><span class="s1">&#39;autoDeleteTtl&#39;</span><span class="p">]</span> <span class="o">=</span> \
+                <span class="s2">&quot;</span><span class="si">{}</span><span class="s2">s&quot;</span><span class="o">.</span><span class="n">format</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">auto_delete_ttl</span><span class="p">)</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span><span class="p">:</span>
             <span class="n">init_actions_dict</span> <span class="o">=</span> <span class="p">[</span>
-                <span class="p">{</span><span class="s1">&#39;executableFile&#39;</span><span class="p">:</span> <span class="n">uri</span><span class="p">}</span> <span class="k">for</span> <span class="n">uri</span> <span class="ow">in</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span>
+                <span class="p">{</span>
+                    <span class="s1">&#39;executableFile&#39;</span><span class="p">:</span> <span class="n">uri</span><span class="p">,</span>
+                    <span class="s1">&#39;executionTimeout&#39;</span><span class="p">:</span> <span class="bp">self</span><span class="o">.</span><span class="n">_get_init_action_timeout</span><span class="p">()</span>
+                <span class="p">}</span> <span class="k">for</span> <span class="n">uri</span> <span class="ow">in</span> <span class="bp">self</span><span class="o">.</span><span class="n">init_actions_uris</span>
             <span class="p">]</span>
             <span class="n">cluster_data</span><span class="p">[</span><span class="s1">&#39;config&#39;</span><span class="p">][</span><span class="s1">&#39;initializationActions&#39;</span><span class="p">]</span> <span class="o">=</span> <span class="n">init_actions_dict</span>
         <span class="k">if</span> <span class="bp">self</span><span class="o">.</span><span class="n">service_account</span><span class="p">:</span>
@@ -468,16 +545,30 @@
             <span class="k">else</span><span class="p">:</span>
                 <span class="k">raise</span> <span class="n">e</span>
 
-        <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">)</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">)</span></div>
 
 
-<span class="k">class</span> <span class="nc">DataprocClusterDeleteOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
+<div class="viewcode-block" id="DataprocClusterDeleteOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataprocClusterDeleteOperator">[docs]</a><span class="k">class</span> <span class="nc">DataprocClusterDeleteOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Delete a cluster on Google Cloud Dataproc. The operator will wait until the</span>
 <span class="sd">    cluster is destroyed.</span>
+
+<span class="sd">    :param cluster_name: The name of the cluster to create.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param project_id: The ID of the google cloud project in which</span>
+<span class="sd">        the cluster runs</span>
+<span class="sd">    :type project_id: string</span>
+<span class="sd">    :param region: leave as &#39;global&#39;, might become relevant in the future</span>
+<span class="sd">    :type region: string</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
 <span class="sd">    &quot;&quot;&quot;</span>
 
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;cluster_name&#39;</span><span class="p">]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;project_id&#39;</span><span class="p">,</span> <span class="s1">&#39;region&#39;</span><span class="p">]</span>
 
     <span class="nd">@apply_defaults</span>
     <span class="k">def</span> <span class="nf">__init__</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span>
@@ -488,23 +579,7 @@
                  <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
                  <span class="o">*</span><span class="n">args</span><span class="p">,</span>
                  <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Delete a cluster on Google Cloud Dataproc.</span>
-
-<span class="sd">        :param cluster_name: The name of the cluster to create.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param project_id: The ID of the google cloud project in which</span>
-<span class="sd">            the cluster runs</span>
-<span class="sd">        :type project_id: string</span>
-<span class="sd">        :param region: leave as &#39;global&#39;, might become relevant in the future</span>
-<span class="sd">        :type region: string</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        &quot;&quot;&quot;</span>
+
         <span class="nb">super</span><span class="p">(</span><span class="n">DataprocClusterDeleteOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -541,7 +616,7 @@
         <span class="p">)</span><span class="o">.</span><span class="n">execute</span><span class="p">()</span>
         <span class="n">operation_name</span> <span class="o">=</span> <span class="n">response</span><span class="p">[</span><span class="s1">&#39;name&#39;</span><span class="p">]</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">log</span><span class="o">.</span><span class="n">info</span><span class="p">(</span><span class="s2">&quot;Cluster delete operation name: </span><span class="si">%s</span><span class="s2">&quot;</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span>
-        <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">_wait_for_done</span><span class="p">(</span><span class="n">service</span><span class="p">,</span> <span class="n">operation_name</span><span class="p">)</span></div>
 
 
 <div class="viewcode-block" id="DataProcPigOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcPigOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcPigOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
@@ -552,29 +627,60 @@
 <span class="sd">    It&#39;s a good practice to define dataproc_* parameters in the default_args of the dag</span>
 <span class="sd">    like the cluster name and UDFs.</span>
 
-<span class="sd">    ```</span>
-<span class="sd">    default_args = {</span>
-<span class="sd">        &#39;cluster_name&#39;: &#39;cluster-1&#39;,</span>
-<span class="sd">        &#39;dataproc_pig_jars&#39;: [</span>
-<span class="sd">            &#39;gs://example/udf/jar/datafu/1.2.0/datafu.jar&#39;,</span>
-<span class="sd">            &#39;gs://example/udf/jar/gpig/1.2/gpig.jar&#39;</span>
-<span class="sd">        ]</span>
-<span class="sd">    }</span>
-<span class="sd">    ```</span>
+<span class="sd">    .. code-block:: python</span>
+
+<span class="sd">        default_args = {</span>
+<span class="sd">            &#39;cluster_name&#39;: &#39;cluster-1&#39;,</span>
+<span class="sd">            &#39;dataproc_pig_jars&#39;: [</span>
+<span class="sd">                &#39;gs://example/udf/jar/datafu/1.2.0/datafu.jar&#39;,</span>
+<span class="sd">                &#39;gs://example/udf/jar/gpig/1.2/gpig.jar&#39;</span>
+<span class="sd">            ]</span>
+<span class="sd">        }</span>
 
 <span class="sd">    You can pass a pig script as string or file reference. Use variables to pass on</span>
 <span class="sd">    variables for the pig script to be resolved on the cluster or use the parameters to</span>
 <span class="sd">    be resolved in the script as template parameters.</span>
 
-<span class="sd">    ```</span>
-<span class="sd">    t1 = DataProcPigOperator(</span>
-<span class="sd">        task_id=&#39;dataproc_pig&#39;,</span>
-<span class="sd">        query=&#39;a_pig_script.pig&#39;,</span>
-<span class="sd">        variables={&#39;out&#39;: &#39;gs://example/output/{{ds}}&#39;},</span>
-<span class="sd">    dag=dag)</span>
-<span class="sd">    ```</span>
+<span class="sd">    **Example**: ::</span>
+
+<span class="sd">        t1 = DataProcPigOperator(</span>
+<span class="sd">                task_id=&#39;dataproc_pig&#39;,</span>
+<span class="sd">                query=&#39;a_pig_script.pig&#39;,</span>
+<span class="sd">                variables={&#39;out&#39;: &#39;gs://example/output/{{ds}}&#39;},</span>
+<span class="sd">                dag=dag)</span>
+
+<span class="sd">    .. seealso::</span>
+<span class="sd">        For more detail on about job submission have a look at the reference:</span>
+<span class="sd">        https://cloud.google.com/dataproc/reference/rest/v1/projects.regions.jobs</span>
+
+<span class="sd">    :param query: The query or reference to the query file (pg or pig extension).</span>
+<span class="sd">    :type query: string</span>
+<span class="sd">    :param query_uri: The uri of a pig script on Cloud Storage.</span>
+<span class="sd">    :type query_uri: string</span>
+<span class="sd">    :param variables: Map of named parameters for the query.</span>
+<span class="sd">    :type variables: dict</span>
+<span class="sd">    :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd">        is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd">        name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd">    :type job_name: string</span>
+<span class="sd">    :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param dataproc_pig_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd">        default arguments</span>
+<span class="sd">    :type dataproc_pig_properties: dict</span>
+<span class="sd">    :param dataproc_pig_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
+<span class="sd">        UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd">    :type dataproc_pig_jars: list</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
+<span class="sd">    :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd">    :type region: string</span>
 <span class="sd">    &quot;&quot;&quot;</span>
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;dataproc_jars&#39;</span><span class="p">]</span>
     <span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">&#39;.pg&#39;</span><span class="p">,</span> <span class="s1">&#39;.pig&#39;</span><span class="p">,)</span>
     <span class="n">ui_color</span> <span class="o">=</span> <span class="s1">&#39;#0273d4&#39;</span>
 
@@ -590,40 +696,10 @@
             <span class="n">dataproc_pig_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
             <span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">&#39;google_cloud_default&#39;</span><span class="p">,</span>
             <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+            <span class="n">region</span><span class="o">=</span><span class="s1">&#39;global&#39;</span><span class="p">,</span>
             <span class="o">*</span><span class="n">args</span><span class="p">,</span>
             <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Create a new DataProcPigOperator.</span>
-
-<span class="sd">        For more detail on about job submission have a look at the reference:</span>
-
-<span class="sd">        https://cloud.google.com/dataproc/reference/rest/v1/projects.regions.jobs</span>
 
-<span class="sd">        :param query: The query or reference to the query file (pg or pig extension).</span>
-<span class="sd">        :type query: string</span>
-<span class="sd">        :param query_uri: The uri of a pig script on Cloud Storage.</span>
-<span class="sd">        :type query_uri: string</span>
-<span class="sd">        :param variables: Map of named parameters for the query.</span>
-<span class="sd">        :type variables: dict</span>
-<span class="sd">        :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd">            is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd">            name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd">        :type job_name: string</span>
-<span class="sd">        :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param dataproc_pig_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd">            default arguments</span>
-<span class="sd">        :type dataproc_pig_properties: dict</span>
-<span class="sd">        :param dataproc_pig_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
-<span class="sd">            UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd">        :type dataproc_pig_jars: list</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        &quot;&quot;&quot;</span>
         <span class="nb">super</span><span class="p">(</span><span class="n">DataProcPigOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -634,6 +710,7 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_pig_properties</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_pig_jars</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
 
     <span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
         <span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -649,14 +726,41 @@
         <span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
         <span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
 
-        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
 
 
 <div class="viewcode-block" id="DataProcHiveOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcHiveOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcHiveOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Start a Hive query Job on a Cloud DataProc cluster.</span>
+
+<span class="sd">    :param query: The query or reference to the query file (q extension).</span>
+<span class="sd">    :type query: string</span>
+<span class="sd">    :param query_uri: The uri of a hive script on Cloud Storage.</span>
+<span class="sd">    :type query_uri: string</span>
+<span class="sd">    :param variables: Map of named parameters for the query.</span>
+<span class="sd">    :type variables: dict</span>
+<span class="sd">    :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd">        is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd">        name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd">    :type job_name: string</span>
+<span class="sd">    :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param dataproc_hive_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd">        default arguments</span>
+<span class="sd">    :type dataproc_hive_properties: dict</span>
+<span class="sd">    :param dataproc_hive_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
+<span class="sd">        UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd">    :type dataproc_hive_jars: list</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
+<span class="sd">    :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd">    :type region: string</span>
 <span class="sd">    &quot;&quot;&quot;</span>
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;dataproc_jars&#39;</span><span class="p">]</span>
     <span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">&#39;.q&#39;</span><span class="p">,)</span>
     <span class="n">ui_color</span> <span class="o">=</span> <span class="s1">&#39;#0273d4&#39;</span>
 
@@ -672,36 +776,10 @@
             <span class="n">dataproc_hive_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
             <span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">&#39;google_cloud_default&#39;</span><span class="p">,</span>
             <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+            <span class="n">region</span><span class="o">=</span><span class="s1">&#39;global&#39;</span><span class="p">,</span>
             <span class="o">*</span><span class="n">args</span><span class="p">,</span>
             <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Create a new DataProcHiveOperator.</span>
-
-<span class="sd">        :param query: The query or reference to the query file (q extension).</span>
-<span class="sd">        :type query: string</span>
-<span class="sd">        :param query_uri: The uri of a hive script on Cloud Storage.</span>
-<span class="sd">        :type query_uri: string</span>
-<span class="sd">        :param variables: Map of named parameters for the query.</span>
-<span class="sd">        :type variables: dict</span>
-<span class="sd">        :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd">            is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd">            name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd">        :type job_name: string</span>
-<span class="sd">        :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param dataproc_hive_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd">            default arguments</span>
-<span class="sd">        :type dataproc_hive_properties: dict</span>
-<span class="sd">        :param dataproc_hive_jars: URIs to jars provisioned in Cloud Storage (example: for</span>
-<span class="sd">            UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd">        :type dataproc_hive_jars: list</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        &quot;&quot;&quot;</span>
+
         <span class="nb">super</span><span class="p">(</span><span class="n">DataProcHiveOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -712,6 +790,7 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_hive_properties</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_hive_jars</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
 
     <span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
         <span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -728,14 +807,41 @@
         <span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
         <span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
 
-        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
 
 
 <div class="viewcode-block" id="DataProcSparkSqlOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcSparkSqlOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcSparkSqlOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Start a Spark SQL query Job on a Cloud DataProc cluster.</span>
+
+<span class="sd">    :param query: The query or reference to the query file (q extension).</span>
+<span class="sd">    :type query: string</span>
+<span class="sd">    :param query_uri: The uri of a spark sql script on Cloud Storage.</span>
+<span class="sd">    :type query_uri: string</span>
+<span class="sd">    :param variables: Map of named parameters for the query.</span>
+<span class="sd">    :type variables: dict</span>
+<span class="sd">    :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd">        is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd">        name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd">    :type job_name: string</span>
+<span class="sd">    :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd">        default arguments</span>
+<span class="sd">    :type dataproc_spark_properties: dict</span>
+<span class="sd">    :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd">        for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd">    :type dataproc_spark_jars: list</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
+<span class="sd">    :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd">    :type region: string</span>
 <span class="sd">    &quot;&quot;&quot;</span>
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;query&#39;</span><span class="p">,</span> <span class="s1">&#39;variables&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;dataproc_jars&#39;</span><span class="p">]</span>
     <span class="n">template_ext</span> <span class="o">=</span> <span class="p">(</span><span class="s1">&#39;.q&#39;</span><span class="p">,)</span>
     <span class="n">ui_color</span> <span class="o">=</span> <span class="s1">&#39;#0273d4&#39;</span>
 
@@ -751,36 +857,10 @@
             <span class="n">dataproc_spark_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
             <span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">&#39;google_cloud_default&#39;</span><span class="p">,</span>
             <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+            <span class="n">region</span><span class="o">=</span><span class="s1">&#39;global&#39;</span><span class="p">,</span>
             <span class="o">*</span><span class="n">args</span><span class="p">,</span>
             <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Create a new DataProcSparkSqlOperator.</span>
-
-<span class="sd">        :param query: The query or reference to the query file (q extension).</span>
-<span class="sd">        :type query: string</span>
-<span class="sd">        :param query_uri: The uri of a spark sql script on Cloud Storage.</span>
-<span class="sd">        :type query_uri: string</span>
-<span class="sd">        :param variables: Map of named parameters for the query.</span>
-<span class="sd">        :type variables: dict</span>
-<span class="sd">        :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd">            is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd">            name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd">        :type job_name: string</span>
-<span class="sd">        :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd">            default arguments</span>
-<span class="sd">        :type dataproc_spark_properties: dict</span>
-<span class="sd">        :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
-<span class="sd">            for UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd">        :type dataproc_spark_jars: list</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        &quot;&quot;&quot;</span>
+
         <span class="nb">super</span><span class="p">(</span><span class="n">DataProcSparkSqlOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -791,6 +871,7 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_spark_properties</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_spark_jars</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
 
     <span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
         <span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -807,15 +888,49 @@
         <span class="n">job</span><span class="o">.</span><span class="n">add_jar_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span><span class="p">)</span>
         <span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
 
-        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
 
 
 <div class="viewcode-block" id="DataProcSparkOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcSparkOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcSparkOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Start a Spark Job on a Cloud DataProc cluster.</span>
+
+<span class="sd">    :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
+<span class="sd">            the main_class, not both together).</span>
+<span class="sd">    :type main_jar: string</span>
+<span class="sd">    :param main_class: Name of the job class. (use this or the main_jar, not both</span>
+<span class="sd">        together).</span>
+<span class="sd">    :type main_class: string</span>
+<span class="sd">    :param arguments: Arguments for the job.</span>
+<span class="sd">    :type arguments: list</span>
+<span class="sd">    :param archives: List of archived files that will be unpacked in the work</span>
+<span class="sd">        directory. Should be stored in Cloud Storage.</span>
+<span class="sd">    :type archives: list</span>
+<span class="sd">    :param files: List of files to be copied to the working directory</span>
+<span class="sd">    :type files: list</span>
+<span class="sd">    :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd">        is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd">        name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd">    :type job_name: string</span>
+<span class="sd">    :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd">        default arguments</span>
+<span class="sd">    :type dataproc_spark_properties: dict</span>
+<span class="sd">    :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd">        for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd">    :type dataproc_spark_jars: list</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</span>
+<span class="sd">    :type delegate_to: string</span>
+<span class="sd">    :param region: The specified region where the dataproc cluster is created.</span>
+<span class="sd">    :type region: string</span>
 <span class="sd">    &quot;&quot;&quot;</span>
 
-    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;arguments&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">]</span>
+    <span class="n">template_fields</span> <span class="o">=</span> <span class="p">[</span><span class="s1">&#39;arguments&#39;</span><span class="p">,</span> <span class="s1">&#39;job_name&#39;</span><span class="p">,</span> <span class="s1">&#39;cluster_name&#39;</span><span class="p">,</span> <span class="s1">&#39;dataproc_jars&#39;</span><span class="p">]</span>
     <span class="n">ui_color</span> <span class="o">=</span> <span class="s1">&#39;#0273d4&#39;</span>
 
     <span class="nd">@apply_defaults</span>
@@ -832,43 +947,10 @@
             <span class="n">dataproc_spark_jars</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
             <span class="n">gcp_conn_id</span><span class="o">=</span><span class="s1">&#39;google_cloud_default&#39;</span><span class="p">,</span>
             <span class="n">delegate_to</span><span class="o">=</span><span class="kc">None</span><span class="p">,</span>
+            <span class="n">region</span><span class="o">=</span><span class="s1">&#39;global&#39;</span><span class="p">,</span>
             <span class="o">*</span><span class="n">args</span><span class="p">,</span>
             <span class="o">**</span><span class="n">kwargs</span><span class="p">):</span>
-        <span class="sd">&quot;&quot;&quot;</span>
-<span class="sd">        Create a new DataProcSparkOperator.</span>
 
-<span class="sd">        :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
-<span class="sd">            the main_class, not both together).</span>
-<span class="sd">        :type main_jar: string</span>
-<span class="sd">        :param main_class: Name of the job class. (use this or the main_jar, not both</span>
-<span class="sd">            together).</span>
-<span class="sd">        :type main_class: string</span>
-<span class="sd">        :param arguments: Arguments for the job.</span>
-<span class="sd">        :type arguments: list</span>
-<span class="sd">        :param archives: List of archived files that will be unpacked in the work</span>
-<span class="sd">            directory. Should be stored in Cloud Storage.</span>
-<span class="sd">        :type archives: list</span>
-<span class="sd">        :param files: List of files to be copied to the working directory</span>
-<span class="sd">        :type files: list</span>
-<span class="sd">        :param job_name: The job name used in the DataProc cluster. This name by default</span>
-<span class="sd">            is the task_id appended with the execution data, but can be templated. The</span>
-<span class="sd">            name will always be appended with a random number to avoid name clashes.</span>
-<span class="sd">        :type job_name: string</span>
-<span class="sd">        :param cluster_name: The name of the DataProc cluster.</span>
-<span class="sd">        :type cluster_name: string</span>
-<span class="sd">        :param dataproc_spark_properties: Map for the Pig properties. Ideal to put in</span>
-<span class="sd">            default arguments</span>
-<span class="sd">        :type dataproc_spark_properties: dict</span>
-<span class="sd">        :param dataproc_spark_jars: URIs to jars provisioned in Cloud Storage (example:</span>
-<span class="sd">            for UDFs and libs) and are ideal to put in default arguments.</span>
-<span class="sd">        :type dataproc_spark_jars: list</span>
-<span class="sd">        :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
-<span class="sd">        :type gcp_conn_id: string</span>
-<span class="sd">        :param delegate_to: The account to impersonate, if any.</span>
-<span class="sd">            For this to work, the service account making the request must have domain-wide</span>
-<span class="sd">            delegation enabled.</span>
-<span class="sd">        :type delegate_to: string</span>
-<span class="sd">        &quot;&quot;&quot;</span>
         <span class="nb">super</span><span class="p">(</span><span class="n">DataProcSparkOperator</span><span class="p">,</span> <span class="bp">self</span><span class="p">)</span><span class="o">.</span><span class="fm">__init__</span><span class="p">(</span><span class="o">*</span><span class="n">args</span><span class="p">,</span> <span class="o">**</span><span class="n">kwargs</span><span class="p">)</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span> <span class="o">=</span> <span class="n">gcp_conn_id</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">delegate_to</span> <span class="o">=</span> <span class="n">delegate_to</span>
@@ -881,6 +963,7 @@
         <span class="bp">self</span><span class="o">.</span><span class="n">cluster_name</span> <span class="o">=</span> <span class="n">cluster_name</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_properties</span> <span class="o">=</span> <span class="n">dataproc_spark_properties</span>
         <span class="bp">self</span><span class="o">.</span><span class="n">dataproc_jars</span> <span class="o">=</span> <span class="n">dataproc_spark_jars</span>
+        <span class="bp">self</span><span class="o">.</span><span class="n">region</span> <span class="o">=</span> <span class="n">region</span>
 
     <span class="k">def</span> <span class="nf">execute</span><span class="p">(</span><span class="bp">self</span><span class="p">,</span> <span class="n">context</span><span class="p">):</span>
         <span class="n">hook</span> <span class="o">=</span> <span class="n">DataProcHook</span><span class="p">(</span><span class="n">gcp_conn_id</span><span class="o">=</span><span class="bp">self</span><span class="o">.</span><span class="n">gcp_conn_id</span><span class="p">,</span>
@@ -895,15 +978,49 @@
         <span class="n">job</span><span class="o">.</span><span class="n">add_file_uris</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">files</span><span class="p">)</span>
         <span class="n">job</span><span class="o">.</span><span class="n">set_job_name</span><span class="p">(</span><span class="bp">self</span><span class="o">.</span><span class="n">job_name</span><span class="p">)</span>
 
-        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">())</span></div>
+        <span class="n">hook</span><span class="o">.</span><span class="n">submit</span><span class="p">(</span><span class="n">hook</span><span class="o">.</span><span class="n">project_id</span><span class="p">,</span> <span class="n">job</span><span class="o">.</span><span class="n">build</span><span class="p">(),</span> <span class="bp">self</span><span class="o">.</span><span class="n">region</span><span class="p">)</span></div>
 
 
 <div class="viewcode-block" id="DataProcHadoopOperator"><a class="viewcode-back" href="../../../../integration.html#airflow.contrib.operators.dataproc_operator.DataProcHadoopOperator">[docs]</a><span class="k">class</span> <span class="nc">DataProcHadoopOperator</span><span class="p">(</span><span class="n">BaseOperator</span><span class="p">):</span>
     <span class="sd">&quot;&quot;&quot;</span>
 <span class="sd">    Start a Hadoop Job on a Cloud DataProc cluster.</span>
+
+<span class="sd">    :param main_jar: URI of the job jar provisioned on Cloud Storage. (use this or</span>
+<span class="sd">            the main_class, not both together).</span>
+<span class="sd">    :type main_jar: string</span>
+<span class="sd">    :param main_class: Name of the job class. (use this or the main_jar, not both</span>
+<span class="sd">        together).</span>
+<span class="sd">    :type main_class: string</span>
+<span class="sd">    :param arguments: Arguments for the job.</span>
+<span class="sd">    :type arguments: list</span>
+<span class="sd">    :param archives: List of archived files that will be unpacked in the work</span>
+<span class="sd">        directory. Should be stored in Cloud Storage.</span>
+<span class="sd">    :type archives: list</span>
+<span class="sd">    :param files: List of files to be copied to the working directory</span>
+<span class="sd">    :type files: list</span>
+<span class="sd">    :param job_name: The job name used in the DataProc cluster. This name by default</span>
+<span class="sd">        is the task_id appended with the execution data, but can be templated. The</span>
+<span class="sd">        name will always be appended with a random number to avoid name clashes.</span>
+<span class="sd">    :type job_name: string</span>
+<span class="sd">    :param cluster_name: The name of the DataProc cluster.</span>
+<span class="sd">    :type cluster_name: string</span>
+<span class="sd">    :param dataproc_hadoop_properties: Map for the Pig properties. Ideal to put in</span>
+<span class="sd">        default arguments</span>
+<span class="sd">    :type dataproc_hadoop_properties: dict</span>
+<span class="sd">    :param dataproc_hadoop_jars: URIs to jars provisioned in Cloud Storage (example:</span>
+<span class="sd">        for UDFs and libs) and are ideal to put in default arguments.</span>
+<span class="sd">    :type dataproc_hadoop_jars: list</span>
+<span class="sd">    :param gcp_conn_id: The connection ID to use connecting to Google Cloud Platform.</span>
+<span class="sd">    :type gcp_conn_id: string</span>
+<span class="sd">    :param delegate_to: The account to impersonate, if any.</span>
+<span class="sd">        For this to work, the service account making the request must have domain-wide</span>
+<span class="sd">        delegation enabled.</s

<TRUNCATED>