You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@beam.apache.org by pa...@apache.org on 2021/03/30 18:50:31 UTC

[beam] 01/01: Documenting temp_dataset parameter in ReadFromBigQuery

This is an automated email from the ASF dual-hosted git repository.

pabloem pushed a commit to branch pabloem-patch-2
in repository https://gitbox.apache.org/repos/asf/beam.git

commit e1571c36828b0ec75b6eeb517513b616cfb3890c
Author: Pablo <pa...@users.noreply.github.com>
AuthorDate: Tue Mar 30 11:49:29 2021 -0700

    Documenting temp_dataset parameter in ReadFromBigQuery
---
 sdks/python/apache_beam/io/gcp/bigquery.py | 4 ++++
 1 file changed, 4 insertions(+)

diff --git a/sdks/python/apache_beam/io/gcp/bigquery.py b/sdks/python/apache_beam/io/gcp/bigquery.py
index 3097d02..ae2d59e 100644
--- a/sdks/python/apache_beam/io/gcp/bigquery.py
+++ b/sdks/python/apache_beam/io/gcp/bigquery.py
@@ -1911,6 +1911,10 @@ class ReadFromBigQuery(PTransform):
       To learn more about type conversions between BigQuery and Avro, see:
       https://cloud.google.com/bigquery/docs/loading-data-cloud-storage-avro\
               #avro_conversions
+    temp_dataset (``google.cloud.bigquery.dataset.DatasetReference``):
+        The dataset in which to create temporary tables when performing file
+        loads. By default, a new dataset is created in the execution project for
+        temporary tables.
    """
 
   COUNTER = 0