You are viewing a plain text version of this content. The canonical link for it is here.
Posted to commits@arrow.apache.org by jo...@apache.org on 2022/05/12 11:25:34 UTC

[arrow] branch master updated: ARROW-16526: [Python] test_partitioned_dataset fails when building with PARQUET but without DATASET

This is an automated email from the ASF dual-hosted git repository.

jorisvandenbossche pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/arrow.git


The following commit(s) were added to refs/heads/master by this push:
     new 7a955f07b3 ARROW-16526: [Python] test_partitioned_dataset fails when building with PARQUET but without DATASET
7a955f07b3 is described below

commit 7a955f07b3472a36d9174eb71883f8f9c33083ae
Author: Weston Pace <we...@gmail.com>
AuthorDate: Thu May 12 13:25:26 2022 +0200

    ARROW-16526: [Python] test_partitioned_dataset fails when building with PARQUET but without DATASET
    
    One of the legacy parquet dataset tests was not properly passing use_legacy_dataset and this caused the test to attempt to use the new datasets module even if it wasn't enabled
    
    Closes #13116 from westonpace/bugfix/MINOR--missing-dataset-mark
    
    Authored-by: Weston Pace <we...@gmail.com>
    Signed-off-by: Joris Van den Bossche <jo...@gmail.com>
---
 python/pyarrow/tests/parquet/test_dataset.py | 3 ++-
 1 file changed, 2 insertions(+), 1 deletion(-)

diff --git a/python/pyarrow/tests/parquet/test_dataset.py b/python/pyarrow/tests/parquet/test_dataset.py
index 7b6845bbc2..2c660a3f6e 100644
--- a/python/pyarrow/tests/parquet/test_dataset.py
+++ b/python/pyarrow/tests/parquet/test_dataset.py
@@ -1542,7 +1542,8 @@ def test_partitioned_dataset(tempdir, use_legacy_dataset):
     })
     table = pa.Table.from_pandas(df)
     pq.write_to_dataset(table, root_path=str(path),
-                        partition_cols=['one', 'two'])
+                        partition_cols=['one', 'two'],
+                        use_legacy_dataset=use_legacy_dataset)
     table = pq.ParquetDataset(
         path, use_legacy_dataset=use_legacy_dataset).read()
     pq.write_table(table, path / "output.parquet")