You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by "Daniel Halperin (JIRA)" <ji...@apache.org> on 2016/07/18 17:10:20 UTC

[jira] [Updated] (BEAM-383) BigQueryIO: update sink to shard into multiple write jobs

     [ https://issues.apache.org/jira/browse/BEAM-383?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]

Daniel Halperin updated BEAM-383:
---------------------------------
    Assignee: Ian Zhou

> BigQueryIO: update sink to shard into multiple write jobs
> ---------------------------------------------------------
>
>                 Key: BEAM-383
>                 URL: https://issues.apache.org/jira/browse/BEAM-383
>             Project: Beam
>          Issue Type: Bug
>          Components: sdk-java-gcp
>            Reporter: Daniel Halperin
>            Assignee: Ian Zhou
>
> BigQuery has global limits on both the # files that can be written in a single job and the total bytes in those files. We should be able to modify BigQueryIO.Write to chunk into multiple smaller jobs that meet these limits, write to temp tables, and atomically copy into the destination table.
> This functionality will let us safely stay within BQ's load job limits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)