You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "Luca Falsina (JIRA)" <ji...@apache.org> on 2017/06/27 14:01:00 UTC

[jira] [Created] (AIRFLOW-1350) Add "query_uri" parameter for Google DataProc operator for both SparkSQL and Hive jobs

Luca Falsina created AIRFLOW-1350:
-------------------------------------

             Summary: Add "query_uri" parameter for Google DataProc operator for both SparkSQL and Hive jobs
                 Key: AIRFLOW-1350
                 URL: https://issues.apache.org/jira/browse/AIRFLOW-1350
             Project: Apache Airflow
          Issue Type: Improvement
          Components: gcp, operators
            Reporter: Luca Falsina
            Assignee: Luca Falsina


Google DataProc APIs allow to provide as an input for some types of job (Pig, SparkSQL, Hive) a "query_file" parameter pointing to a file stored in Google Cloud Storage.

The current implementation of the DataProc operator allows this functionality only for Pig jobs. The code modification should be trivial as I plan to just reuse the same implementation in SparkSQL and Hive operator.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)