You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "Raúl Cumplido (Jira)" <ji...@apache.org> on 2022/05/12 15:56:00 UTC

[jira] [Created] (ARROW-16548) [Python] Fix test_parquet_invalid_version test when no parquet or dataset are defined

Raúl Cumplido created ARROW-16548:
-------------------------------------

             Summary: [Python] Fix test_parquet_invalid_version test when no parquet or dataset are defined
                 Key: ARROW-16548
                 URL: https://issues.apache.org/jira/browse/ARROW-16548
             Project: Apache Arrow
          Issue Type: Bug
          Components: Python
    Affects Versions: 8.0.0
            Reporter: Raúl Cumplido
            Assignee: Raúl Cumplido
             Fix For: 9.0.0


Working on the minimal_builds I have found that when building arrow and pyarrow without DATASET nor PARQUET the following error happens:
{code:java}
    def test_parquet_invalid_version(tempdir):
        table = pa.table({'a': [1, 2, 3]})
        with pytest.raises(ValueError, match="Unsupported Parquet format version"):
>           _write_table(table, tempdir / 'test_version.parquet', version="2.2")
E           NameError: name '_write_table' is not defined

pyarrow/tests/parquet/test_basic.py:52: NameError{code}
This is the first test on the parquet/test_basic.py module.

I have done some investigation and if I change the test order (move that test on the file), the test that fails is a different one, always the first test on the module.
If I import (but not use)  from pyarrow.tests.parquet.common import pytestmark no test failures happen.

There seems to be a bug on how the pytestmark is injected. I am creating a ticket because it seems it might require more than a MINOR fix.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)