You are viewing a plain text version of this content. The canonical link for it is here.
Posted to jira@arrow.apache.org by "Nicolas Elie (Jira)" <ji...@apache.org> on 2020/06/24 09:57:00 UTC
[jira] [Commented] (ARROW-7939) [Python] crashes when reading
parquet file compressed with snappy
[ https://issues.apache.org/jira/browse/ARROW-7939?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17143714#comment-17143714 ]
Nicolas Elie commented on ARROW-7939:
-------------------------------------
Hello,
I just faced the same issue with pyarrow 0.17.1 installed from conda-forge on Windows 7 64 bit with Python 3.8.3.
The example code given in description crash with array_nok and compression='snappy' but not with other compression algorithms nor with array_ok. Works fine with pyarrow 0.15.1
I also tried with a clean environment as suggested in first comment (pyarrow 0.16) and the problem is the same.
I tried to open the generated parquet files (that can't be read by pyarrow, compression='snappy') with [fastparquet |https://github.com/dask/fastparquet] or [ParquetViewer|https://github.com/mukunku/ParquetViewer]: They both can't read the parquet file...
What additionnal information could I give you?
> [Python] crashes when reading parquet file compressed with snappy
> -----------------------------------------------------------------
>
> Key: ARROW-7939
> URL: https://issues.apache.org/jira/browse/ARROW-7939
> Project: Apache Arrow
> Issue Type: Bug
> Components: Python
> Affects Versions: 0.16.0
> Environment: Windows 7
> python 3.6.9
> pyarrow 0.16 from conda-forge
> Reporter: Marc Bernot
> Assignee: Wes McKinney
> Priority: Major
> Fix For: 0.17.0
>
>
> When I installed pyarrow 0.16, some parquet files created with pyarrow 0.15.1 would make python crash. I drilled down to the simplest example I could find.
> It happens that some parquet files created with pyarrow 0.16 cannot either be read back. The example below works fine with arrays_ok but python crashes with arrays_nok (and as soon as they are at least three different values apparently).
> Besides, it works fine with 'none', 'gzip' and 'brotli' compression. The problem seems to happen only with snappy.
> {code:python}
> import pyarrow.parquet as pq
> import pyarrow as pa
> arrays_ok = [[0,1]]
> arrays_ok = [[0,1,1]]
> arrays_nok = [[0,1,2]]
> table = pa.Table.from_arrays(arrays_nok,names=['a'])
> pq.write_table(table,'foo.parquet',compression='snappy')
> pq.read_table('foo.parquet')
> {code}
--
This message was sent by Atlassian Jira
(v8.3.4#803005)