You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@airflow.apache.org by "ASF subversion and git services (JIRA)" <ji...@apache.org> on 2017/10/18 18:31:00 UTC

[jira] [Commented] (AIRFLOW-1718) Increase num_retries polling value on Dataproc hook

    [ https://issues.apache.org/jira/browse/AIRFLOW-1718?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16209792#comment-16209792 ] 

ASF subversion and git services commented on AIRFLOW-1718:
----------------------------------------------------------

Commit 6078e753aac35aa4f5971a719d9f736c35396770 in incubator-airflow's branch refs/heads/master from [~cjqian]
[ https://git-wip-us.apache.org/repos/asf?p=incubator-airflow.git;h=6078e75 ]

[AIRFLOW-1718] Set num_retries on Dataproc job request execution

Closes #2696 from cjqian/1718


> Increase num_retries polling value on Dataproc hook
> ---------------------------------------------------
>
>                 Key: AIRFLOW-1718
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-1718
>             Project: Apache Airflow
>          Issue Type: Bug
>            Reporter: Crystal Qian
>            Assignee: Crystal Qian
>            Priority: Minor
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> Currently, `num_retries = 0` when execute() is called (https://google.github.io/google-api-python-client/docs/epy/googleapiclient.http.HttpRequest-class.html#execute), which causes intermittent 500 errors (https://stackoverflow.com/questions/46522261/deadline-exceeded-when-airflow-runs-spark-jobs). We should increase this to allow retries for internal Dataproc queries to other services in the short-term; also seeing if the `num_retries` count can be increased at the _google-api-python-client_ level in the long-term.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)