You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "Krisztian Szucs (Jira)" <ji...@apache.org> on 2019/09/17 14:59:00 UTC

[jira] [Comment Edited] (ARROW-5072) [Python] write_table fails silently on S3 errors

    [ https://issues.apache.org/jira/browse/ARROW-5072?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16931542#comment-16931542 ] 

Krisztian Szucs edited comment on ARROW-5072 at 9/17/19 2:58 PM:
-----------------------------------------------------------------

The root cause originates from s3fs. {{S3File}} uses buffering, and no S3 request happens until it is closed or flushed:

{code:python}
out = S3File(S3FileSystem(), 's3://some-bogus-bucket/df.parquet', mode='wb')
out.write(b'bbb')  # returns with 3
del out  # raises exception 
{code}

If we would call {{out.close()}} then it would be flushed, but the file itself is opened outside of arrow, so the caller should be responsible for closing it. We could fall {{flush()}} on file-like objects, although it works a bit differently with {{S3File}}:

{code:python}
out.flush()  # doesn't raise
out.flush(force=True)  # raises, but force is S3File - fsspec specific
{code}

When {{S3FileSystem}} is used, then the file object is opened and thus closed by arrow, so the exception propagates from {{pq.write_table()}}:

{code:python}
pq.write_table(table, s3_filepath, filesystem=S3FileSystem())
{code}

I'd consider this as an s3fs issue, because {{S3File().write()}} work on non-existing file (actually I get a non-authorised error, because I don't have S3 credentials set up), and the error is raised on object destruction, from {{S3File.__del__()}}. 

From arrow perspective the provided behaviour is the expected, although we can have a note for this s3fs case in the documentation.

cc [~wesmckinn] [~pitrou]




was (Author: kszucs):
The root cause originates from s3fs. {{S3File}} uses buffering, and no S3 request happens until it is closed or flushed:

{code:python}
out = S3File(S3FileSystem(), 's3://some-bogus-bucket/df.parquet', mode='wb')
out.write(b'bbb')  # returns with 3
del out  # raises exception 
{code}

If we would call {out.close()} then it would be flushed, but the file itself is opened outside of arrow, so the caller should be responsible for closing it. We could fall {flush()} on file-like objects, although it works a bit differently with {S3File}:

{code:python}
out.flush()  # doesn't raise
out.flush(force=True)  # raises, but force is S3File - fsspec specific
{code}

When {S3FileSystem} is used, then the file object is opened and thus closed by arrow, so the exception propagates from {pq.write_table()}:

{code:python}
pq.write_table(table, s3_filepath, filesystem=S3FileSystem())
{code}

I'd consider this as an s3fs issue, because {S3File().write()} work on non-existing file (actually I get a non-authorised error, because I don't have S3 credentials set up), and the error is raised on object destruction, from {S3File.__del__()}. 

From arrow perspective the provided behaviour is the expected, although we can have a note for this s3fs case in the documentation.

cc [~wesmckinn] [~pitrou]



> [Python] write_table fails silently on S3 errors
> ------------------------------------------------
>
>                 Key: ARROW-5072
>                 URL: https://issues.apache.org/jira/browse/ARROW-5072
>             Project: Apache Arrow
>          Issue Type: Bug
>          Components: Python
>    Affects Versions: 0.12.1
>         Environment: Python 3.6.8
>            Reporter: Paul George
>            Priority: Minor
>              Labels: filesystem, parquet
>             Fix For: 0.15.0
>
>
> {{pyarrow==0.12.1}}
> *pyarrow.parquet.write_table* called with where=S3File(...) fails silently when encountering errors while writing to S3 (in the example below, boto3 is raising a NoSuchBucket exception). However, instead of using S3File(), calling write_table with where=_<filepath>_ and with filesystem=S3FileSystem() does *not* fail silently and raises, as is expected.
> h4. Code/Repro
>  
> {code:java}
> import pandas as pd
> import pyarrow as pa
> import pyarrow.parquet as pq
> from s3fs import S3File, S3FileSystem
> df = pd.DataFrame({'col0': []})
> s3_filepath = 's3://some-bogus-bucket/df.parquet'
> print('>> test 1')
> try:
> # use S3File --> fails silently
> pq.write_table(pa.Table.from_pandas(df.copy()),
> S3File(S3FileSystem(), s3_filepath, mode='wb'))
> except Exception:
> print('>>>> Exception raised!')
> else:
> print('>>>> Exception **NOT** raised!')
> print('>> test 2')
> try:
> # use filepath and S3FileSystem --> raises Exception, as expected
> pq.write_table(pa.Table.from_pandas(df.copy()),
> s3_filepath,
> filesystem=S3FileSystem())
> except Exception:
> print('>>>> Exception raised!')
> else:
> print('>>>> Exception **NOT** raised!'){code}
>  
> h4.  
> h4. Output
> {code:java}
> >> test 1
> Exception ignored in: <bound method S3File.__del__ of <S3File some-bogus-bucket/df.parquet>>
> Traceback (most recent call last):
> File "<redacted>/lib/python3.6/site-packages/s3fs/core.py", line 1476, in __del__
> self.close()
> File "<redacted>/lib/python3.6/site-packages/s3fs/core.py", line 1454, in close
> raise_from(IOError('Write failed: %s' % self.path), e)
> File "<string>", line 3, in raise_from
> OSError: Write failed: some-bogus-bucket/df.parquet
> >>>> Exception **NOT** raised!
> >> test 2
> >>>> Exception raised!
> Exception ignored in: <bound method S3File.__del__ of <S3File some-bogus-bucket/df.parquet>>
> Traceback (most recent call last):
> File "<redacted>/lib/python3.6/site-packages/s3fs/core.py", line 1476, in __del__
> self.close()
> File "<redacted>/lib/python3.6/site-packages/s3fs/core.py", line 1454, in close
> raise_from(IOError('Write failed: %s' % self.path), e)
> File "<string>", line 3, in raise_from
> OSError: Write failed: some-bogus-bucket/df.parquet
> {code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)