You are viewing a plain text version of this content. The canonical link for it is here.
Posted to jira@arrow.apache.org by "Weston Pace (Jira)" <ji...@apache.org> on 2022/09/26 16:25:00 UTC
[jira] [Resolved] (ARROW-17614) [CI][Python] test test_write_dataset_max_rows_per_file is producing several nightly build failures
[ https://issues.apache.org/jira/browse/ARROW-17614?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Weston Pace resolved ARROW-17614.
---------------------------------
Resolution: Fixed
Issue resolved by pull request 14199
[https://github.com/apache/arrow/pull/14199]
> [CI][Python] test test_write_dataset_max_rows_per_file is producing several nightly build failures
> --------------------------------------------------------------------------------------------------
>
> Key: ARROW-17614
> URL: https://issues.apache.org/jira/browse/ARROW-17614
> Project: Apache Arrow
> Issue Type: Bug
> Components: Continuous Integration, Python
> Reporter: Raúl Cumplido
> Priority: Blocker
> Labels: Nightly, pull-request-available
> Fix For: 10.0.0
>
> Time Spent: 2h
> Remaining Estimate: 0h
>
> The following failure has been seen on multiple nightly builds:
> {code:java}
> _____________________ test_write_dataset_max_rows_per_file _____________________tempdir = PosixPath('/tmp/pytest-of-root/pytest-0/test_write_dataset_max_rows_pe0') @pytest.mark.parquet
> def test_write_dataset_max_rows_per_file(tempdir):
> directory = tempdir / 'ds'
> max_rows_per_file = 10
> max_rows_per_group = 10
> num_of_columns = 2
> num_of_records = 35
>
> record_batch = _generate_data_and_columns(num_of_columns,
> num_of_records)
>
> ds.write_dataset(record_batch, directory, format="parquet",
> max_rows_per_file=max_rows_per_file,
> > max_rows_per_group=max_rows_per_group)usr/local/lib/python3.7/site-packages/pyarrow/tests/test_dataset.py:3921:
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> usr/local/lib/python3.7/site-packages/pyarrow/dataset.py:992: in write_dataset
> min_rows_per_group, max_rows_per_group, create_dir
> pyarrow/_dataset.pyx:2811: in pyarrow._dataset._filesystemdataset_write
> ???
> _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ > ???
> E FileNotFoundError: [Errno 2] Failed to open local file '/tmp/pytest-of-root/pytest-0/test_write_dataset_max_rows_pe0/ds/part-1.parquet'. Detail: [errno 2] No such file or directory {code}
> Example of failed builds:
> [verify-rc-source-python-macos-conda-amd64|https://github.com/ursacomputing/crossbow/runs/8176702861?check_suite_focus=true]
> [wheel-manylinux2014-cp37-amd64|https://github.com/ursacomputing/crossbow/runs/8175319639?check_suite_focus=true]
> It seems flaky as there were some nightly jobs executed on a previous day without new commits that were successful.
--
This message was sent by Atlassian Jira
(v8.20.10#820010)