You are viewing a plain text version of this content. The canonical link for it is here.
Posted to issues@arrow.apache.org by "Sep Dehpour (Jira)" <ji...@apache.org> on 2020/07/15 01:28:00 UTC

[jira] [Created] (ARROW-9474) Column type inference in read_csv vs. open_csv. CSV conversion error to null.

Sep Dehpour created ARROW-9474:
----------------------------------

             Summary: Column type inference in read_csv vs. open_csv. CSV conversion error to null.
                 Key: ARROW-9474
                 URL: https://issues.apache.org/jira/browse/ARROW-9474
             Project: Apache Arrow
          Issue Type: Bug
            Reporter: Sep Dehpour


The open_csv stream does not adjust the inferred column type based on the new data seen in new blocks.

For example if a csv has null values in the first few blocks of open_csv reader, the column is inferred as Null type. As PyArrow iterates over blocks and sees non null values in that column,  it crashes.

Example Error:
{code:java}
pyarrow.lib.ArrowInvalid: In CSV column #44: CSV conversion error to null: invalid value '-176400' {code}
 

This problem is resolved if a read_option with a huge block size is passed to the open_csv. But that negates the whole point of having a stream vs. read_csv.

 

System info:

PyArrow 0.17.1, Mac OS Catalina, Python 3.7.4



--
This message was sent by Atlassian Jira
(v8.3.4#803005)